VibeBuilders.ai Logo
VibeBuilders.ai

Considering

Explore resources related to considering to help implement AI solutions for your business.

I am considering starting a SaaS business that automates the creation of long-form SEO-optimized blog posts. Is this something you would find useful, as a business owner?
reddit
LLM Vibe Score0
Human Vibe Score1
What_The_HexThis week

I am considering starting a SaaS business that automates the creation of long-form SEO-optimized blog posts. Is this something you would find useful, as a business owner?

Trying to gauge the general interest level, from other entrepreneurs/business owners. The idea is, a tool that automates the process of creating long-form SEO optimized blog posts to promote your business -- perhaps creating entire batches of such posts, all from just one button click. Like if you could just describe your business, click a button, and BAM, it just outputs like an entire month's worth of absolutely fire SEO-optimized long-form blog posts? That would be super fucking convenient. Yes you can use ChatGPT for this, but the character limits make it so it can only output very short posts. Otherwise it requires first asking for an outline, then getting the different sections piecemeal and pasting it all together yourself. Still super time-consuming to do it that way. A GPT-based solution could probably automate the process I've hit upon in my own SEO blog-posting workflow -- where I output not just finished long-form blog posts, but also convert them into SEO-optimized HTML code so you can just paste it into your blog post website and have all the header tags etc set up for optimal SEO/keyword ranking purposes. Biggest counter-argument I make against this is, there are undoubtedly lots of companies already offering this. Doesn't mean I can't make money doing it. I just don't like entering super crowded marketplaces. Other main argument I have is, if I used my OpenAI account for this, there's the risk of some malicious/idiot user firing prompts that violate the OpenAI ToS and get me banned. I COULD have them input their own OpenAI API tokens, but that just adds adoption/usage barriers that would make it way harder to market/acquire initial customers. I guess I could sanitize the user inputs as a pre-processing step to block any obscene prompts or anything like that, but still, it's a risk. Let me know your thoughts on this idea. ASSUMING it worked effectively -- and made it very easy for you to just describe your business offerings / value propositions / target market(s), then get genuinely useful long-form SEO-optimized blog posts, is this something you'd be willing to pay for? If so, what dollar amount, to you, would seem reasonable? It would probably just be hosted on a website. Then you'd just copy the outputted final result for use as needed on your website. That would be the simplest way to do it. Technically it could function as like, a plugin for specific websites that maybe auto-posts them for you too -- it would be simpler, on my end, to start out doing this on a standalone website. (Might also make it easier to allow users to try it out, on first visit.) One last point -- MAYBE it would have an optional intermediate step, where it would first output the planned outline for the blog post, allowing you to pop in, quickly modify that, add your own thoughts / valuable ideas (to help make the blog post more unique, truly useful for readers, more your own) -- THEN you could finalize it and hit submit. Again, that's the workflow I've hit upon in my own semi-automated blog-posting workflow, and it's led to some pretty useful long-form content that isn't just, boring garbage, but contains lots of genuinely useful ideas that I would include in my own uniquely-created blog posts on the subject. But instead of me taking the time to write it, I just kinda toss in a few quickly typed out ideas to expand upon, and ChatGPT does the rest. Imagine that kind of optional / customizable workflow, but the rest of it is fully automated. OR you could just get the fully automated blog posts with no revisions on your part. Thanks!

Anyone finding that they just don't NEED to add more Employees anymore? (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score0.6
wilschroterThis week

Anyone finding that they just don't NEED to add more Employees anymore? (I will not promote)

A friend of mine who was looking for work asked me if we were hiring and I responded "You know, it's weird but all of our growth goals don't seem to map back to hiring people anymore." This isn't about the economy or growth goals. It's a really fascinating shift in focus and costs for startups. My gut reaction is that I HATE the idea of not creating more jobs. In my career I've hired thousands of people, and I've always prided myself on job creation. We just sold a company that employed 200 people last year, and I'm proud of the work we were able to create. What's interesting is that I simply don't feel like we NEED to like we used to. As we're looking at all of our growth goals, for the first time I'm not assigning FTEs to them. Nearly everything we're doing is actually reducing the need for more humans, not adding them - and we're not even trying to reduce the need. Obviously the timing of AI has had a major impact. Product - Our team is shipping more code than ever before, and even our designers who have never touched code are shipping final code. If we doubled the size of the team, it would make no difference (this is a big deal considering the historical cost here). Marketing - So many aspects of our marketing are getting automated and streamlined, to the point where even a single FTE can create a massive amount of reach across channels. Support - Our Success team is able to effectively respond to tickets in a fraction of the time, which essentially doubles their capacity without adding any more staff. Management - With less staff we need less managers, which are a big expense, but it also means reporting and decisions are more streamlined, which is a positive. But it also means those positions simply don't get created like they used to. I think this is a big deal for the younger startups because it translates into needing less capital (or none!) which provides for more ownership and agency. Clearly we still need some folks to build out the core team, but that's very different than a massive staffing line item. Anyone else here finding the same trend? Opposite? I don't have a strong opinion either way, but I'd love to hear how other Founders are processing this.

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

Content aggregation that acts as a middleman for content discovery via third-party marketplace & revenue sharing (i will not promote but I'm looking for fellow researchers)
reddit
LLM Vibe Score0
Human Vibe Score1
colbyn-wadmanThis week

Content aggregation that acts as a middleman for content discovery via third-party marketplace & revenue sharing (i will not promote but I'm looking for fellow researchers)

High level I’m considering a content aggregation business model, but one that acts as an open marketplace where third party devs and where world class data scientists compete to build the best recommenders for different use cases. (E.g. the incentives can be ad revenue sharing or subscription based for niche professional markets.) The idea is to facilitate more bottom up innovation from third party data scientists. The platform itself just acts as the middleman. (Also something that strips out original ads and makes it easy to skip paid sponsorship sections would be great.)  I’ve seen startups building web crawlers and content aggregation systems for other AI startups. My proposal is better in the sense that third party devs are instead responsible for implementing whatever questionable hacks are necessarily to scrape platforms that don’t necessarily want to be scraped.  Personally, I’m more concerned about getting the right information than ever before, to this end I can’t rely on platform specific recommenders. The solution is more bottom up innovation in content promotion. More generally, if you’re also concerned about consuming game changing information that’s too easily missed: we need a platform that incentivizes bottom up innovation of content promotion. What we need is a platform that functions like a marketplace where third party devs and where world class data scientists compete to build the best recommenders for different use cases. Here’s some elevator pitches I’m considering:  Did you know that the magic behind YouTube is its recommendation engine? Now, imagine an open platform where independent engines compete to deliver the most personalized content feed—from news to local events—directly to you. Interested in rethinking how we find content? “In today’s fragmented digital landscape, a single platform no longer holds sway over content discovery. The Network Effect is dead: audiences are more mobile than ever; and big tech killed it. In such a fragmented landscape we’re building a bottom-up, decentralized marketplace for recommendation engines—a solution that taps into diverse revenue streams through subscriptions, ad revenue, and affiliate partnerships. Invest in the future of personalized content aggregation.” “Are you a developer passionate about algorithms and content discovery? Our open marketplace lets you build and monetize your own recommendation engine, competing to deliver the most engaging, personalized feeds. Join a revolution where your innovation can directly shape how the world finds content.” “Are you tired of being told what to watch or read by one mysterious algorithm? Imagine taking control—choosing from a marketplace of smart recommendation engines that curate content just for you. It’s a revolution in content discovery where you hold the power.” (As a Utahn this one is interesting because even mormons are talking about the dangers of “doom scrolling” though it’s seldom discussed in society at large.) As far as simple hooks I’m considering:  One platform to rule them all and in the darkness bind them.  Choose how you discover—content recommenders that work for you.  The area where recommender engines battle to win your feed. Request I would love to start prototyping this idea and see what else I can uncover from such preliminary research. But I want to get a couple other likeminded individuals onboard.  I'm the best when it comes to iOS/macOS development, but there's tons of backend work that needs to be done which I wouldn’t have the time for if i'm focused on the native clients. Who am I 'ideally' looking for?  I’ve heard of weird stats to the effect that if you scale up a population to billions of people, the number of life overlaps starts skyrocketing. Not just physical lookalikes, but people with eerily similar life paths, personalities, habits, and even thoughts — without ever knowing each other. Where are my clones? Such is whom I’m looking for in an ideal world.  Take a hunch  People nowadays have no concept of going out on a limb, taking a ‘hunch’, and backing their instincts. Everything has to be calculated, proven, and guaranteed before they make a move. In contrast consider the success of the Chinese DeepSeek project: According to Asianometry’s YouTube video on DeepSeek, their “memory-saving multi-head latent architecture” (whatever that means, just quoting the name) came about from a researchers ‘hunch’, which the company bet big on and the result was drastically improved performance on low end hardware…  Here in the west the idea of betting on a hunch is inconceivable. We have no balls to chase long term insights. My own instincts when it comes to software is such because I’ve wasted too much of my life on small scale projects. All I’m trying to do is attempt a more scaled up experiment based on some hunches with me and a few other likeminded individuals.  Just as the early oil prospectors didn’t have precise maps—just intuition and test drills. They had to drill, analyze the pressure, and adjust. The best oil fields weren’t found by foresight alone, but by adaptive exploration. The startup space itself is liken to the first prospectors who got the gold nuggets lying in the riverbed. In such an environment moving first has its advantages but nowadays I wish I could have all those shitty ‘engineers’ sent to their maker.  Today the reality is such that you’ve got to dig deep—where vast stores of wealth can be found—or go home, and those who dig into the depths cannot use mere forethought, for what lies beneath cannot be seen by the mind’s eye.  I will not promote but I'm looking for fellow research oriented minds.

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

Technical Co-Founder Seeking Commercial/Marketing Partner for Micro SaaS Projects
reddit
LLM Vibe Score0
Human Vibe Score1
Weekly-Offer-4172This week

Technical Co-Founder Seeking Commercial/Marketing Partner for Micro SaaS Projects

Hi everyone, I’m looking for a commercial or marketing co-founder to join me in developing some Micro SaaS (MSaaS) apps. Here’s a bit about where I’m coming from and what I’m hoping to find: About Me: I’m a full-stack developer with over 15 years of experience, including some work in AI. I’m currently working part-time, which gives me the time to focus on developing MVPs quickly. I’m passionate about creating SaaS solutions and would love to find someone who can help bring these ideas to life. Based in french alps. What I’m Looking For: Role: Non-Technical Co-Founder (Commercial/Marketing) Location: Remote Equity: 50% co-founder stake What I’m Hoping You’ll Bring: Experience: Background in business development, marketing, or similar fields. Vision: An eye for potential in new SaaS ideas and a drive to help make them successful. Commitment: Enthusiasm for building and growing a business together. What’s In It For You: Revenue Potential: Share in the financial rewards of successful products with a 50% equity stake, giving you a direct share of the profits. Fast ROI: Benefit from rapid MVP development, which allows for quicker validation and faster revenue generation. Dynamic Approach: We move quickly—if an app doesn’t gain traction in a few weeks, we pivot to the next idea, keeping our efforts focused on what works. Financial Growth: As we iterate and scale, there are opportunities for significant financial upside based on the success of our products. Shared Success: Be an integral part of a partnership where both of us share equally in the risks and rewards, creating a strong incentive for mutual success. What’s In It For You: Partnership: Equal share in the business (50/50). Opportunity: Work on interesting MSaaS projects with room for creativity. Flexibility: A remote role that fits around your schedule. If you’re interested or would like to learn more, please reach out. I’d be thrilled to discuss how we might work together. Thank you for considering this!

We built a tool to help you find relevant grants. Would you pay for it?
reddit
LLM Vibe Score0
Human Vibe Score1
CliznitchThis week

We built a tool to help you find relevant grants. Would you pay for it?

Hi everyone, About a year ago, I asked you guys whether it would make sense to develop a tool to help entrepreneurs find relevant grants. Many of you provided incredibly valuable feedback, which we used to refine the concept. With this concept, we went through Techstars and finally launched a beta version of our grant scan tool last week! Along the way, we realized something interesting: when you ask a grant advisor which grants might be a great fit for you, they almost always recommend the ones they know well. This makes sense since most work on a success fee basis, and referring you to lesser-known grants (which take more time to write and have lower success rates) isn’t worth it for them. Plus, memorizing the details of 20,000+ grants is, understandably, pretty tough. Our platform uses AI to scan and analyze thousands of grants. It identifies the best matches, estimates your chances of success, and calculates how much time you might need for the application and reporting phases. We can then match you with a grant advisor with relevant expertise—whether to write the application for you or provide feedback on your draft. We’re considering launching both a free and a paid version. The free version would provide basic insights, while the paid version would include more comprehensive results, expert comments (such as explaining why certain grants are a good fit), and updates when new relevant grants become available. Both versions will allow you to connect with relevant experts. Would you pay for the paid version? And if so, which features should it include? Also, any general feedback is much appreciated! Thanks!

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

10y of product development, 2 bankruptcies, and 1 Exit — what next? [Extended Story]
reddit
LLM Vibe Score0
Human Vibe Score1
Slight-Explanation29This week

10y of product development, 2 bankruptcies, and 1 Exit — what next? [Extended Story]

10 years of obsessive pursuit from the bottom to impressive product-market fit and exit. Bootstrapping tech products as Software Developer and 3x Startup Founder (2 bankruptcies and 1 exit). Hi everyone, your motivation has inspired me to delve deeper into my story. So, as promised to some of you, I've expanded on it a bit more, along with my brief reflections. There are many founders, product creators, and proactive individuals, I’ve read many of your crazy stories and lessons so I decided to share mine and the lessons I learned from the bottom to impressive product-market fit and exit. I've spent almost the past 10 years building tech products as a Corporate Team Leader, Senior Software Developer, Online Course Creator, Programming Tutor, Head of Development/CTO, and 3x Startup Founder (2 bankruptcies, and 1 exit). And what next? good question... A brief summary of my journey: Chapter 1: Software Developer / Team Leader / Senior Software Developer I’ve always wanted to create products that win over users’ hearts, carry value, and influence users. Ever since my school days, I’ve loved the tech part of building digital products. At the beginning of school, I started hosting servers for games, blogs and internet forums, and other things that did not require much programming knowledge. My classmates and later even over 100 people played on servers that I hosted on my home PC. Later, as the only person in school, I passed the final exam in computer science. During my computer science studies, I started my first job as a software developer. It was crazy, I was spending 200–300 hours a month in the office attending also to daily classes. Yes, I didn’t have a life, but it truly was the fulfillment of my dreams. I was able to earn good money doing what I love, and I devoted fully myself to it. My key to effectively studying IT and growing my knowledge at rocket speed was learning day by day reading guides, building products to the portfolio, watching youtube channels and attending conferences, and even watching them online, even if I didn’t understand everything at the beginning. In one year we’ve been to every possible event within 400km. We were building healthcare products that were actually used in hospitals and medical facilities. It was a beautiful adventure and tons of knowledge I took from this place. That time I built my first product teams, hired many great people, and over the years became a senior developer and team leader. Even I convinced my study mates to apply to this company and we studied together and worked as well. Finally, there were 4 of us, when I left a friend of mine took over my position and still works there. If you’re reading this, I’m sending you a flood of love and appreciation. I joined as the 8th person, and after around 4 years, when I left hungry for change, there were already over 30 of us, now around 100. It was a good time, greetings to everyone. I finished my Master’s and Engineering degrees in Computer Science, and it was time for changes. Chapter 2: 1st time as a Co-founder — Marketplace In the meantime, there was also my first startup (a marketplace) with four of my friends. We all worked on the product, each of us spent thousands of hours, after hours, entire weekends… and I think finally over a year of work. As you might guess, we lacked the most important things: sales, marketing, and product-market fit. We thought users think like us. We all also worked commercially, so the work went very smoothly, but we didn’t know what we should do next with it… Finally, we didn’t have any customers, but you know what, I don’t regret it, a lot of learning things which I used many times later. The first attempts at validating the idea with the market and business activities. In the end, the product was Airbnb-sized. Landing pages, listings, user panels, customer panels, admin site, notifications, caches, queues, load balancing, and much more. We wanted to publish the fully ready product to the market. It was a marketplace, so if you can guess, we had to attract both sides to be valuable. “Marketplace” — You can imagine something like Uber, if you don’t have passengers it was difficult to convince taxi drivers, if you don’t have a large number of taxi drivers you cannot attract passengers. After a year of development, we were overloaded, and without business, marketing, sales knowledge, and budget. Chapter 3: Corp Team Lead / Programming Tutor / Programming Architecture Workshop Leader Working in a corporation, a totally different environment, an international fintech, another learning experience, large products, and workmates who were waiting for 5 pm to finish — it wasn’t for me. Very slow product development, huge hierarchy, being an ant at the bottom, and low impact on the final product. At that time I understood that being a software developer is not anything special and I compared my work to factory worker. Sorry for that. High rates have been pumped only by high demand. Friends of mine from another industry do more difficult things and have a bigger responsibility for lower rates. That’s how the market works. This lower responsibility time allowed for building the first online course after hours, my own course platform, individual teaching newbies programming, and my first huge success — my first B2C customers, and B2B clients for workshops. I pivoted to full focus on sales, marketing, funnels, advertisements, demand, understanding the market, etc. It was 10x easier than startups but allowed me to learn and validate my conceptions and ideas on an easier market and showed me that it’s much easier to locate their problem/need/want and create a service/product that responds to it than to convince people of your innovative ideas. It’s just supply and demand, such a simple and basic statement, in reality, is very deep and difficult to understand without personal experience. If you’re inexperienced and you think you understand, you don’t. To this day, I love to analyze this catchword in relation to various industries / services / products and rediscover it again and again... While writing this sentence, I’m wondering if I’m not obsessed. Chapter 4: Next try — 2nd time as a founder — Edtech Drawing upon my experiences in selling services, offering trainings, and teaching programming, I wanted to broaden my horizons, delve into various fields of knowledge, involve more teachers, and so on. We started with simple services in different fields of knowledge, mainly relying on teaching in the local area (without online lessons). As I had already gathered some knowledge and experience in marketing and sales, things were going well and were moving in the right direction. The number of teachers in various fields was growing, as was the number of students. I don’t remember the exact statistics anymore, but it was another significant achievement that brought me a lot of satisfaction and new experiences. As you know, I’m a technology lover and couldn’t bear to look at manual processes — I wanted to automate everything: lessons, payments, invoices, customer service, etc. That’s when I hired our first developers (if you’re reading this, I’m sending you a flood of love — we spent a lot of time together and I remember it as a very fruitful and great year) and we began the process of tool and automation development. After a year we had really extended tools for students, teachers, franchise owners, etc. We had really big goals, we wanted to climb higher and higher. Maybe I wouldn’t even fully call it Startup, as the client was paying for the lessons, not for the software. But it gave us positive income, bootstrap financing, and tool development for services provided. Scaling this model was not as costless as SaaS because customer satisfaction was mainly on the side of the teacher, not the quality of the product (software). Finally, we grew to nearly 10 people and dozens of teachers, with zero external funding, and almost $50k monthly revenue. We worked very hard, day and night, and by November 2019, we were packed with clients to the brim. And as you know, that’s when the pandemic hit. It turned everything upside down by 180 degrees. Probably no one was ready for it. With a drastic drop in revenues, society started to save. Tired from the previous months, we had to work even harder. We had to reduce the team, change the model, and save what we had built. We stopped the tool’s development and sales, and with the developers, we started supporting other product teams to not fire them in difficult times. The tool worked passively for the next two years, reducing incomes month by month. With a smaller team providing programming services, we had full stability and earned more than relying only on educational services. At the peak of the pandemic, I promised myself that it was the last digital product I built… Never say never… Chapter 5: Time for fintech — Senior Software Developer / Team Lead / Head of Development I worked for small startups and companies. Building products from scratch, having a significant impact on the product, and complete fulfillment. Thousands of hours and sacrifices. This article mainly talks about startups that I built, so I don’t want to list all the companies, products, and applications that I supported as a technology consultant. These were mainly start-ups with a couple of people up to around 100 people on board. Some of the products were just a rescue mission, others were building an entire tech team. I was fully involved in all of them with the hope that we would work together for a long time, but I wasn’t the only one who made mistakes when looking for a product-market fit. One thing I fully understood: You can’t spend 8–15 hours a day writing code, managing a tech team, and still be able to help build an audience. In marketing and sales, you need to be rested and very creative to bring results and achieve further results and goals. If you have too many responsibilities related to technology, it becomes ineffective. I noticed that when I have more free time, more time to think, and more time to bounce the ball against the wall, I come up with really working marketing/sales strategies and solutions. It’s impossible when you are focused on code all day. You must know that this chapter of my life was long and has continued until now. Chapter 6: 3rd time as a founder — sold Never say never… right?\\ It was a time when the crypto market was really high and it was really trending topic. You know that I love technology right? So I cannot miss the blockchain world. I had experience in blockchain topics by learning on my own and from startups where I worked before. I was involved in crypto communities and I noticed a “starving crowd”. People who did things manually and earned money(crypto) on it.I found potential for building a small product that solves a technological problem. I said a few years before that I don’t want to start from scratch. I decided to share my observations and possibilities with my good friend. He said, “If you gonna built it, I’m in”. I couldn’t stop thinking about it. I had thought and planned every aspect of marketing and sales. And you know what. On this huge mindmap “product” was only one block. 90% of the mindmap was focused on marketing and sales. Now, writing this article, I understood what path I went from my first startup to this one. In the first (described earlier) 90% was the product, but in the last one 90% was sales and marketing. Many years later, I did this approach automatically. What has changed in my head over the years and so many mistakes? At that time, the company for which I provided services was acquired. The next day I got a thank you for my hard work and all my accounts were blocked. Life… I was shocked. We were simply replaced by their trusted technology managers. They wanted to get full control. They acted a bit unkindly, but I knew that they had all my knowledge about the product in the documentation, because I’m used to drawing everything so that in the moment of my weakness (illness, whatever) the team could handle it. That’s what solid leaders do, right? After a time, I know that these are normal procedures in financial companies, the point is that under the influence of emotions, do not do anything inappropriate. I quickly forgot about it, that I was brutally fired. All that mattered was to bring my plan to life. And it has been started, 15–20 hours a day every day. You have to believe me, getting back into the game was incredibly satisfying for me. I didn’t even know that I would be so excited. Then we also noticed that someone was starting to think about the same product as me. So the race began a game against time and the market. I assume that if you have reached this point, you are interested in product-market fit, marketing, and sales, so let me explain my assumptions to you: Product: A very very small tool that allowed you to automate proper tracking and creation of on-chain transactions. Literally, the whole app for the user was located on only three subpages. Starving Crowd: We tapped into an underserved market. The crypto market primarily operates via communities on platforms like Discord, Reddit, Twitter, Telegram, and so on. Therefore, our main strategy was directly communicating with users and demonstrating our tool. This was essentially “free marketing” (excluding the time we invested), as we did not need to invest in ads, promotional materials, or convince people about the efficacy of our tool. The community could directly observe on-chain transactions executed by our algorithms, which were processed at an exceptionally fast rate. This was something they couldn’t accomplish manually, so whenever someone conducted transactions using our algorithm, it was immediately noticeable and stirred a curiosity within the community (how did they do that!). Tests: I conducted the initial tests of the application on myself — we had already invested significantly in developing the product, but I preferred risking my own resources over that of the users. I provided the tool access to my wallet, containing 0.3ETH, and went to sleep. Upon waking up, I discovered that the transactions were successful and my wallet had grown to 0.99ETH. My excitement knew no bounds, it felt like a windfall. But, of course, there was a fair chance I could have lost it too. It worked. As we progressed, some users achieved higher results, but it largely hinged on the parameters set by them. As you can surmise, the strategy was simple — buy low, sell high. There was considerable risk involved. Churn: For those versed in marketing, the significance of repeat visitors cannot be overstated. Access to our tool was granted only after email verification and a special technique that I’d prefer to keep confidential. And this was all provided for free. While we had zero followers on social media, we saw an explosion in our email subscriber base and amassed a substantial number of users and advocates. Revenue Generation: Our product quickly gained popularity as we were effectively helping users earn — an undeniable value proposition. Now, it was time to capitalize on our efforts. We introduced a subscription model charging $300 per week or $1,000 per month — seemingly high rates, but the demand was so intense that it wasn’t an issue. Being a subscriber meant you were prioritized in the queue, ensuring you were among the first to reap benefits — thus adding more “value”. Marketing: The quality of our product and its ability to continually engage users contributed to it achieving what can best be described as viral. It was both a source of pride and astonishment to witness users sharing charts and analyses derived from our tool in forum discussions. They weren’t actively promoting our product but rather using screenshots from our application to illustrate certain aspects of the crypto world. By that stage, we had already assembled a team to assist with marketing, and programming, and to provide round-the-clock helpdesk support. Unforgettable Time: Despite the hype, my focus remained steadfast on monitoring our servers, their capacity, and speed. Considering we had only been on the market for a few weeks, we were yet to implement alerts, server scaling, etc. Our active user base spanned from Japan to the West Coast of the United States. Primarily, our application was used daily during the evenings, but considering the variety of time zones, the only time I could afford to sleep was during the evening hours in Far Eastern Europe, where we had the least users. However, someone always needed to be on guard, and as such, my phone was constantly by my side. After all, we couldn’t afford to let our users down. We found ourselves working 20 hours a day, catering to thousands of users, enduring physical fatigue, engaging in talks with VCs, and participating in conferences. Sudden Downturn: Our pinnacle was abruptly interrupted by the war in Ukraine (next macroeconomic shot straight in the face, lucky guy), a precipitous drop in cryptocurrency value, and swiftly emerging competition. By this time, there were 5–8 comparable tools had infiltrated the market. It was a challenging period as we continually stumbled upon new rivals. They immediately embarked on swift fundraising endeavors — a strategy we overlooked, which in retrospect was a mistake. Although our product was superior, the competitors’ rapid advancement and our insufficient funds for expeditious scaling posed significant challenges. Nonetheless, we made a good decision. We sold the product (exit) to competitors. The revenue from “exit” compensated for all the losses, leaving us with enough rest. We were a small team without substantial budgets for rapid development, and the risk of forming new teams without money to survive for more than 1–2 months was irresponsible. You have to believe me that this decision consumed us sleepless nights. Finally, we sold it. They turned off our app but took algorithms and users. Whether you believe it or not, after several months of toiling day and night, experiencing burnout, growing weary of the topic, and gaining an extra 15 kg in weight, we finally found our freedom… The exit wasn’t incredibly profitable, but we knew they had outdone us. The exit covered all our expenses and granted us a well-deserved rest for the subsequent quarter. It was an insane ride. Despite the uncertainty, stress, struggles, and sleepless nights, the story and experience will remain etched in my memory for the rest of my life. Swift Takeaways: Comprehending User Needs: Do you fully understand the product-market fit? Is your offering just an accessory or does it truly satisfy the user’s needs? The Power of Viral Marketing: Take inspiration from giants like Snapchat, ChatGPT, and Clubhouse. While your product might not attain the same scale (but remember, never say never…), the closer your concept is to theirs, the easier your journey will be. If your user is motivated to text a friend saying, “Hey, check out how cool this is” (like sharing ChatGPT), then you’re on the best track. Really. Even if it doesn’t seem immediately evident, there could be a way to incorporate this into your product. Keep looking until you find it. Niche targeting — the more specific and tailored your product is to a certain audience, the easier your journey will be People love buying from people — establishing a personal brand and associating yourself with the product can make things easier. Value: Seek to understand why users engage with your product and keep returning. The more specific and critical the issue you’re aiming to solve, the easier your path will be. Consider your offerings in terms of products and services and focus on sales and marketing, regardless of personal sentiments. These are just a few points, I plan to elaborate on all of them in a separate article. Many products undergo years of development in search of market fit, refining the user experience, and more. And guess what? There’s absolutely nothing wrong with that. Each product and market follows its own rules. Many startups have extensive histories before they finally make their mark (for instance, OpenAI). This entire journey spanned maybe 6–8 months. I grasped and capitalized on the opportunity, but we understood from the start that establishing a startup carried a significant risk, and our crypto product was 10 times riskier. Was it worth it? Given my passion for product development — absolutely. Was it profitable? — No, considering the hours spent — we lose. Did it provide a stable, problem-free life — nope. Did this entire adventure offer a wealth of happiness, joy, and unforgettable experiences — definitely yes. One thing is certain — we’ve amassed substantial experience and it’s not over yet :) So, what lies ahead? Chapter 7: Reverting to the contractor, developing a product for a crypto StartupReturning to the past, we continue our journey… I had invested substantial time and passion into the tech rescue mission product. I came on board as the technical Team Leader of a startup that had garnered over $20M in seed round funding, affiliated with the realm of cryptocurrencies. The investors were individuals with extensive backgrounds in the crypto world. My role was primarily technical, and there was an abundance of work to tackle. I was fully immersed, and genuinely devoted to the role. I was striving for excellence, knowing that if we secured another round of financing, the startup would accelerate rapidly. As for the product and marketing, I was more of an observer. After all, there were marketing professionals with decades of experience on board. These were individuals recruited from large crypto-related firms. I had faith in them, kept an eye on their actions, and focused on my own responsibilities. However, the reality was far from satisfactory. On the last day, the principal investor for the Series A round withdrew. The board made the tough decision to shut down. It was a period of intense observation and gaining experience in product management. This was a very brief summary of the last 10 years. And what next? (Last) Chapter 8: To be announced — Product Owner / Product Consultant / Strategist / CTO After spending countless hours and days deliberating my next steps, one thing is clear: My aspiration is to continue traversing the path of software product development, with the hopeful anticipation that one day, I might ride the crest of the next big wave and ascend to the prestigious status of a unicorn company. I find myself drawn to the process of building products, exploring product-market fit, strategizing, engaging in software development, seeking out new opportunities, networking, attending conferences, and continuously challenging myself by understanding the market and its competitive landscape. Product Owner / Product Consultant / CTO / COO: I’m not entirely sure how to categorize this role, as I anticipate that it will largely depend on the product to which I will commit myself fully. My idea is to find one startup/company that wants to build a product / or already has a product, want to speed up, or simply doesn’t know what’s next. Alternatively, I could be a part of an established company with a rich business history, which intends to invest in digitization and technological advancements. The goal would be to enrich their customer experience by offering complementary digital products Rather than initiating a new venture from ground zero with the same team, I am receptive to new challenges. I am confident that my past experiences will prove highly beneficial for the founders of promising, burgeoning startups that already possess a product, or are in the initial phases of development. ‘Consultant’ — I reckon we interpret this term differently. My aim is to be completely absorbed in a single product, crafting funnels, niches, strategies, and all that is necessary to repeatedly achieve the ‘product-market fit’ and significant revenue. To me, ‘consultant’ resonates more akin to freelancing than being an employee. My current goal is to kickstart as a consultant and aide, dealing with facilitating startups in their journey from point A to B. Here are two theoretical scenarios to illustrate my approach: Scenario 1: (Starting from point A) You have a product but struggle with marketing, adoption, software, strategy, sales, fundraising, or something else. I conduct an analysis and develop a strategy to reach point B. I take on the “dirty work” and implement necessary changes, including potential pivots or shifts (going all-in) to guide the product to point B. The goal is to reach point B, which could involve achieving a higher valuation, expanding the user base, increasing sales, or generating monthly revenue, among other metrics. Scenario 2: (Starting from point A) You have a plan or idea but face challenges with marketing, adoption, strategy, software, sales, fundraising, or something else. I analyze the situation and devise a strategy to reach point B. I tackle the necessary tasks, build the team, and overcome obstacles to propel the product to point B. I have come across the view that finding the elusive product-market fit is the job of the founder, and it’s hard for me to disagree. However, I believe that my support and experiences can help save money, many failures, and most importantly, time. I have spent a great deal of time learning from my mistakes, enduring failure after failure, and even had no one to ask for support or opinion, which is why I offer my help. Saving even a couple of years, realistically speaking, seems like a value I’m eager to provide… I invite you to share your thoughts and insights on these scenarios :) Closing Remarks: I appreciate your time and effort in reaching this point. This has been my journey, and I wouldn’t change it for the world. I had an extraordinary adventure, and now I’m ready for the next exciting battle with the market and new software products. While my entire narrative is centered around startups, especially the ones I personally built, I’m planning to share more insights drawn from all of my experiences, not just those as a co-founder. If you’re currently developing your product or even just considering the idea, I urge you to reach out to me. Perhaps together, we can create something monumental :) Thank you for your time and insights. I eagerly look forward to engaging in discussions and hearing your viewpoints. Please remember to like and subscribe. Nothing motivates to write more than positive feedback :) Matt.

Voice AI Isn’t Just for Big Brands – Here’s How Startups Can Use It (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
Altruistic_Bid_3044This week

Voice AI Isn’t Just for Big Brands – Here’s How Startups Can Use It (I will not promote)

When you think about Voice AI, it’s easy to picture massive companies like Amazon or Google pouring millions into complex systems. But it isn’t just for the big guys anymore. Startups can use it too, and it doesn’t have to cost a fortune. Why Startups Should Care About Voice AI Voice AI used to be expensive and complicated, but that’s changed a lot. Today, even small startups can use it to save time, cut costs, and make customers happier—all without needing a massive budget. If you think that repetitive tasks are eating up your team’s time, or if customers are getting frustrated by slow responses, Voice AI can help. And it’s not just for call centers or tech giants. Startups can benefit from it just as much, if not more. 3 Practical Ways Startups Can Use Voice AI Automated Scheduling and Appointment Setting Whether it’s booking meetings, setting reminders, or rescheduling, Voice AI can handle it all. This is especially useful for service-based startups, like healthcare clinics, legal firms, or consulting agencies. Answering Frequently Asked Questions (FAQs) Every startup gets repetitive questions—“What are your hours?” “What’s your refund policy?” Instead of answering the same things over and over, Voice AI can automate it. Order Tracking and Status Updates For e-commerce startups, Voice AI can provide real-time order updates without involving a human. Customers get quick answers, and your team can focus on more important tasks. Simple Workflow: How It Works Customer Initiates Call Customer calls the business for scheduling, FAQs, or order updates. Voice AI Answers AI responds with a natural, human-like voice. AI Handles the Request Schedules appointments, answers FAQs, or provides order updates. Integration and Confirmation Syncs with calendars or order management systems. Confirms booking or provides tracking info. Call Ends Customer gets what they need without waiting. Team stays focused on higher-priority tasks. If the fear is that Voice AI will sound robotic or annoy customers, it’s worth reconsidering. Today’s tech is way more natural and human-like than it used to be. You can use free trial of platforms like Retell AI or Play AI or Bland AI (I will not promote) Would it make sense for your startup to try Voice AI?

Looking for a Marketing Partner for an Innovative AI Mobile App [i will not promote]
reddit
LLM Vibe Score0
Human Vibe Score1
Altruistic-Flan-8222This week

Looking for a Marketing Partner for an Innovative AI Mobile App [i will not promote]

Hello everyone! I'm a software engineer and AI developer working on something great in the mobile AI space. If you have been following the trends on TikTok and similar platforms, you have probably noticed the explosion of AI apps (like Rizz AI and similar) that follow the simple "scan → solve" concept. These apps have been massively successful because they solve specific problems with minimal user friction. Here's what makes my project different: I have identified an unique market where there is currently zero competition for this app idea that I'm creating and the potential user base is massive - we are talking about 200M+ potential users in the US alone (60% of the US population could use this app). Even capturing just 0.05% of this market could generate significant revenue, considering similar apps typically charge $4-6 per user. What I'm looking for: A marketing partner (preferably US-based or someone familiar with the US market/audience) who can help grow this app. Initially, it requires about 30–60 minutes per day for content creation and posting. No experience is required. If you don't have marketing experience, don't worry. In today's marketing, passion is often more important than skills (and a bit of luck, haha). What I'm offering: For now, it's a revenue share partnership. I have invested my savings into the development of the app and the necessary equipment and I'm offering a revenue share until we generate enough profit for paid positions. Once we gain traction, the goal is to transition this into a part-time or full-time role. If you have zero creativity skills, I can provide you with my automated content generation tool to assist with marketing. It is basically a script that generates the type of content that gets the most views on other AI apps promoted on social media platforms. This is also a long-term partnership, if we achieve some results but not good enough with one app, we can try a new niche or just continue on this one. About the project: The app is almost complete and will likely launch in mid-February. It is a self-funded venture, meaning all profits will be reinvested into growth, including ads, revenue sharing and potentially useful tools to improve marketing. Also, the app is unique, I made a deep research and there is no similar app in this niche and it is very easy to promote. Overall, it follows a simple and effective business model with a clear monetization strategy. If you're interested in being part of something with genuine growth potential and want to learn more, DM me. We can discuss details on Reddit, Discord, LinkedIn, anything you like. The app launches in mid-February so I'm looking to bring someone on board soon to help out. Note: I will share specific details about the niche and app functionality in private messages to protect the idea before launch.

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
GummyBear8659This week

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)

Hey everyone! Long-time lurker, first-time founder here. I’m reaching out to get feedback on a recent startup experience—what went wrong, what I could have done better, and how I should approach future opportunities. The Background There were three founders in this venture: • Founder A (CEO, 50%) – The product/growth guy who identified the problem space. • Founder B (Me, CTO, 37.5%) – A software engineer with a software dev shop and multiple clients. I wanted to diversify into building my own products but am not inherently a “product person.” • Founder C (COO, 12.5%) – Brought into the mix by Founder A, with the goal of leveraging his network for traction once the product was built. The idea was to create Product X, a solution targeting the SMB space while competitors were moving upmarket. It wasn’t revolutionary—more of a strategic market play. The Initial Plan & My Role • Founder A would define and prioritize product specs, guiding what needed to be built. • I (Founder B) didn’t have time to code myself, so I allocated engineers from my dev shop (which I personally paid for). My stake was adjusted from 32.5% to 37.5% to reflect this contribution. • Founder C was more of an observer early on, planning to help with traction once we had a product ready. We agreed on a 1-year cliff and a 4-year vesting schedule for equity. Where Things Started to Go Wrong • Lack of a Clear Product Roadmap – Founder A was very focused on getting something built fast, but we never signed off on a structured roadmap or milestones. I underestimated the complexity of what was actually needed for customer conversations. • Engineering Expectations vs. Reality – The team (one part-time lead + two full-time juniors from my dev shop) faced early feedback that development was too slow. In response, I ramped up the lead to full-time and added a part-time PM. But Founder A continued pushing for speed, despite real hurdles (OAuth integrations, etc.). • Shifting MVP Goalposts – Midway, Founder A concluded that an MVP wouldn’t cut it—we needed a more complete product to be competitive. This meant more engineering, more delays, and more of my own money spent on development. The Breaking Point Near the 1-year vesting mark, we had an opportunity: a paying client willing to fund an app. I didn’t have devs on the bench, so I asked Founder A to hold off our project briefly while I hired more engineers to avoid stalling either effort. This was the final straw. Founder A (with Founder C somewhat aligned) decided the arrangement wasn’t working—citing past disagreements and the “slowness” issue. The decision was made to end the partnership. Now, Founder A, as majority holder, is requesting a full handover of the code, Founder C is indifferent, and all engineering costs I covered are essentially lost. Key Takeaways (So Far) Crystal-Clear Agreements Upfront – A formalized product roadmap and timeline should’ve been locked in from day one. Business Needs > Engineering Standards – I wanted to build something solid and scalable, but in an early-stage startup, speed to market is king. This was before AI tools became mainstream, so our approach wasn’t as optimized. Don’t Overextend Without Protection – I personally financed all engineering, but without clear safeguards, that investment became a sunk cost. Expenses Must Be Distributed – I was solely covering engineering salaries, which created an imbalance in financial risk. Future partnerships should ensure costs are shared proportionally, rather than one person shouldering the burden. Where I Need Advice Looking back, I want to improve as an engineer, CEO, and co-founder. • What should I have done differently in structuring this partnership? • How do you balance engineering quality with the startup need for speed? • As a dev shop owner, how can I better navigate equity deals where I’m also bringing in engineering resources? I really appreciate everyone who went through this long post and provide any insights from founders, engineers, or anyone who has been in a similar situation. Thanks for reading! ===================================================================== For readers who might be thinking what set this type of expectation? Because I had a dev shop and I thought my co-founders will be understanding of my business circumstance and I was a bit trigger to build a product with a C-exec team, I gave the impression of "unlimited" engineering which I later realized down the line that it was not feasible for me. Something I learned that I have to be more careful with and set expectations accordingly from the very beginning. And from the feedback of the commenters here, I am much more aware what I should offer and how to set expectations, esp. in the early stages of execution. So thank you all! 🙏🏾 EDIT: I would like to thank everyone who contributed to this thread. You not only helped me but future founders who are considering to get into the startup scene!

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
GummyBear8659This week

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)

Hey everyone! Long-time lurker, first-time founder here. I’m reaching out to get feedback on a recent startup experience—what went wrong, what I could have done better, and how I should approach future opportunities. The Background There were three founders in this venture: • Founder A (CEO, 50%) – The product/growth guy who identified the problem space. • Founder B (Me, CTO, 37.5%) – A software engineer with a software dev shop and multiple clients. I wanted to diversify into building my own products but am not inherently a “product person.” • Founder C (COO, 12.5%) – Brought into the mix by Founder A, with the goal of leveraging his network for traction once the product was built. The idea was to create Product X, a solution targeting the SMB space while competitors were moving upmarket. It wasn’t revolutionary—more of a strategic market play. The Initial Plan & My Role • Founder A would define and prioritize product specs, guiding what needed to be built. • I (Founder B) didn’t have time to code myself, so I allocated engineers from my dev shop (which I personally paid for). My stake was adjusted from 32.5% to 37.5% to reflect this contribution. • Founder C was more of an observer early on, planning to help with traction once we had a product ready. We agreed on a 1-year cliff and a 4-year vesting schedule for equity. Where Things Started to Go Wrong • Lack of a Clear Product Roadmap – Founder A was very focused on getting something built fast, but we never signed off on a structured roadmap or milestones. I underestimated the complexity of what was actually needed for customer conversations. • Engineering Expectations vs. Reality – The team (one part-time lead + two full-time juniors from my dev shop) faced early feedback that development was too slow. In response, I ramped up the lead to full-time and added a part-time PM. But Founder A continued pushing for speed, despite real hurdles (OAuth integrations, etc.). • Shifting MVP Goalposts – Midway, Founder A concluded that an MVP wouldn’t cut it—we needed a more complete product to be competitive. This meant more engineering, more delays, and more of my own money spent on development. The Breaking Point Near the 1-year vesting mark, we had an opportunity: a paying client willing to fund an app. I didn’t have devs on the bench, so I asked Founder A to hold off our project briefly while I hired more engineers to avoid stalling either effort. This was the final straw. Founder A (with Founder C somewhat aligned) decided the arrangement wasn’t working—citing past disagreements and the “slowness” issue. The decision was made to end the partnership. Now, Founder A, as majority holder, is requesting a full handover of the code, Founder C is indifferent, and all engineering costs I covered are essentially lost. Key Takeaways (So Far) Crystal-Clear Agreements Upfront – A formalized product roadmap and timeline should’ve been locked in from day one. Business Needs > Engineering Standards – I wanted to build something solid and scalable, but in an early-stage startup, speed to market is king. This was before AI tools became mainstream, so our approach wasn’t as optimized. Don’t Overextend Without Protection – I personally financed all engineering, but without clear safeguards, that investment became a sunk cost. Expenses Must Be Distributed – I was solely covering engineering salaries, which created an imbalance in financial risk. Future partnerships should ensure costs are shared proportionally, rather than one person shouldering the burden. Where I Need Advice Looking back, I want to improve as an engineer, CEO, and co-founder. • What should I have done differently in structuring this partnership? • How do you balance engineering quality with the startup need for speed? • As a dev shop owner, how can I better navigate equity deals where I’m also bringing in engineering resources? I really appreciate everyone who went through this long post and provide any insights from founders, engineers, or anyone who has been in a similar situation. Thanks for reading! ===================================================================== For readers who might be thinking what set this type of expectation? Because I had a dev shop and I thought my co-founders will be understanding of my business circumstance and I was a bit trigger to build a product with a C-exec team, I gave the impression of "unlimited" engineering which I later realized down the line that it was not feasible for me. Something I learned that I have to be more careful with and set expectations accordingly from the very beginning. And from the feedback of the commenters here, I am much more aware what I should offer and how to set expectations, esp. in the early stages of execution. So thank you all! 🙏🏾 EDIT: I would like to thank everyone who contributed to this thread. You not only helped me but future founders who are considering to get into the startup scene!

I spent 6 months on building a web product, and got zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a web product, and got zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I have stuff to post on Reddit very rarely, but I share how my project is going on, random stuff, and memes on X. Just in case few might want to keep in touch 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

Seeking Feedback: Would a No-Code AI Solution Benefit Your Business?
reddit
LLM Vibe Score0
Human Vibe Score0
chrisparkerofficialThis week

Seeking Feedback: Would a No-Code AI Solution Benefit Your Business?

Hi, fellow small business owners. I'm currently working on an AI startup, with the goal of providing small businesses a seamless and intuitive way to integrate AI into operations without the need for any coding or tech expertise. We're designing an auto machine learning application that's user-friendly and tailored to the unique needs of small businesses. Before we scale, I would really appreciate any insights and feedback. Here are a few questions that would be helpful to get answers to: Pain Points: Are there specific tasks or processes in your operations that you think could be automated or enhanced using AI? This could be anything from customer service chatbots, inventory management, sales forecasting, or anything else you might think of. Features: What features would you want in a no-code AI solution? Perhaps easy integration with existing software? Drag-and-drop model training? Pre-built models for common tasks? Training & Support: How important would training and support be for you in implementing and using an AI solution? Would you prefer video tutorials, live-chat support, or hands-on workshops? Pricing: Would you be willing to invest in such a tool? If so, what would be a reasonable price point for you? We're considering a tiered model based on usage, with a potential starting point of $X/month. Does that sound feasible? Trial Period: Would a free trial period be beneficial for you? How long would you need to assess the tool's impact on your business? Data Concerns: How comfortable are you with sharing data with an AI application? What privacy and security measures would make you feel at ease? Your feedback is really useful. We're building this solution with you in mind, and your insights will guide the next steps. In appreciation for your time and input, we're offering a special discount for early adopters from this community once we launch. Just drop a comment below, and I'll make sure to get in touch when we are ready. Many Thanks, Chris Parker

The Advantages of a Custom CRM Solution
reddit
LLM Vibe Score0
Human Vibe Score-1
NeerajKumarChaurasiaThis week

The Advantages of a Custom CRM Solution

The growth in the global CRM market continues to accelerate. According to techspective, the global CRM market is now worth \~ $40B USD and is expected to surpass $80B USD by 2025. Despite this phenomenal growth, the CRM market is still dominated by off-the-shelf solutions that are “cookie-cutter” in design and that provide little to no options for customization. These non-customized CRM solutions can significantly inhibit an enterprise’s ability to maximize the advantages of CRM adoption and to realize a robust ROI. As a result, companies are increasingly opting for digital CRM solutions that are customized to meet the unique needs of the enterprise. What is driving the increased demand for custom CRM solutions? What are some of the inherent advantages of a custom CRM solution when compared to a typical off-the-shelf product? Off-the-Shelf CRM Solutions – the Limitations Static CRM solutions are inflexible and self-limiting. Enterprises saddled with these cookie-cutter solutions increasingly report a consistent listing of issues that limit business growth.  These include…. A lack of real-time visibility into shifting customer trends and demands Delayed reaction to coordination of internal resources to meet changing business conditions Lost business opportunities due to lack of flexible, and real-time, opportunity lifecycle management Reporting and dashboarding capabilities that are slow, static, and disconnected Poor quote-to-cash performance that degrades financial performance A CRM investment that delivers poor ROI and that cannot grow with the enterprise All of the above can combine to limit the enterprise’s ability to fully capitalize on its hard-won business opportunities and, over time, limit its ability to create new opportunities. A Customized CRM – What is it? What is a “customized CRM”? Simply put, it is a holistic CRM solution that has been specifically tailored for the individual enterprise. The provider of a truly customized CRM solution will deliver a solution that has been designed to meet the specific—and unique—demands and objectives of the enterprise. A tailored CRM solution will address the enterprise’s sales and operational requirements as well as its customer experience objectives. Unlike standard off-the-shelf CRM providers, a provider of enterprise-grade custom CRM solutions will employ a comprehensive project discovery and requirements gathering process. This is an integral process that provides the foundation for the development of a custom solution that will provide the enterprise with long term flexibility and scalability. A customized digital CRM solution can provide distinct competitive advantages; including: Dynamic, Flexible, Powerful, Real-Time Management and Engagement A customized, technology\-fueled, CRM solution provides the enterprise with the means with which to dynamically engage with customers in ways that build customer loyalty, generate market growth, and drive strong financial performance. Distinct advantages include: Real-time sales opportunity tracking. Helps eliminate lost opportunities due to slow or inadequate reaction. Customized, AI and IoT-fueled, data analytics. A customized CRM solution can be designed to deliver real-time insights. Allows the enterprise to anticipate, and then satisfy, the needs of the customer. Customizable Dashboards and Reports. Widget-based, customized, dashboards and reports that provide real-time data and actionable insights. Sales process automation. Intelligent Workflow-based automation and control of critical sales processes. Increases overall operational efficiency. Outstanding ROI. A custom CRM solution typically delivers superior ROI when compared to off-the-shelf CRM products. Enterprises today spend considerable time, money, and effort in the development of customer relationships. For many enterprises, the continued use of CRM solutions that are rigid and outdated can prove to be impediments to business growth. When considering investment in a new CRM solution any enterprise will be well served by full consideration of a CRM solution that can be fully customized to meet its long-range requirements.

Restarting My Agency / Compared To Full Time Corporate
reddit
LLM Vibe Score0
Human Vibe Score1
nomadpaddyThis week

Restarting My Agency / Compared To Full Time Corporate

I’m currently thinking about going back to consulting / agency work compared to my current tech job I have. Over a year ago I signed this tech client and they wanted more and more from me which ended up becoming a full time role. At the time, the challenge excited me as it was working on a very large project on a global scale, competing with some of the biggest brands in the world. I was making good money before working in my agency and consulting with lots of different brands on their paid media, websites and e-commerce. I have a healthy package where I’m at at the moment but want more. Working with different clients always created curiosity, no day was the same and that what I loved about it. So now I’m considering going to back to starting the business and working with clients again. My question is: What do businesses ACTUALLY want? Everyone wants great roas and an amazing site but what are core things people are looking for in a growth partner / agency? I’m thinking of relaunching with three pillars in mind: Digital (Paid Media, Lead Gen, Web Dev) AI implementation as a lot of businesses don’t know how to leverage AI completely for cost saving and efficiencies. Content (Video, SEO, Content Writing) for modern day ranking I’m currently rebuilding my pitch deck and thought I would ask the question here before I go back to my network and start opening up conversations again. Would love to hear people’s thoughts in addition to anyone that’s done the same?

I spent 6 months on building a web product, and got zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a web product, and got zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I have stuff to post on Reddit very rarely, but I share how my project is going on, random stuff, and memes on X. Just in case few might want to keep in touch 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

The "AI Agent" Hype is out of control and businesses suffer
reddit
LLM Vibe Score0
Human Vibe Score0.429
ImpossibleBell4759This week

The "AI Agent" Hype is out of control and businesses suffer

Ah, the sweet smell of AI hype in the morning. Nothing quite like it to get the blood pumping and the venture capital flowing. Let's cut through the BS... The "AI Agent" craze is the tech industry's latest attempt to separate businesses from their hard-earned cash. It's like watching a bunch of sheep rushing towards a cliff, except the cliff is made of overpriced software and empty promises. The tech giants are having a field day with this nonsense. Microsoft, Google, Salesforce - they're all pushing AI agents like they're the second coming. The sad truth is, businesses are suffering from a severe case of FOMO (Fear of Missing Out). They're so terrified of being left behind in the AI race that they're willing to throw good money after bad. Here's a radical idea: how about focusing on actual business problems instead of chasing the latest tech fad? I know, I know, it's not as sexy as having an AI Agent, but it might actually, you know, work. In the end, the only ones truly benefiting from this AI agent hype are the vendors selling the snake oil and the consultants charging exorbitant fees to implement it. Everyone else is just along for the ride, hoping they don't crash and burn too spectacularly. So, to all the businesses out there considering jumping on the AI Agent bandwagon... take a step back, take a deep breath, and ask yourself if you really need an overpriced chatbot with delusions of grandeur. Chances are, you don't. The AI agent hype is like a bad reality TV show—overproduced, lacking substance, and leaving businesses with nothing but regret. Companies are throwing money at AI solutions, expecting miracles, only to find they've bought into overpriced fantasies. The AI agent hype is nothing more than a high-tech emperor with no clothes. It's time for businesses to wake up, smell the silicon, and start making decisions based on reality rather than sci-fi fantasies.  I think AI Agents are the future, but as of right now AI Agents aren't autonomous or agentic. From what I've seen as of now is glorified Chatbots, ChatGPT wrappers and basic automations, and nothing actually autonomous. So far it's all just hype, but we'll see how it improves businesses and the bottom line! How do you think AI Agents will help small businesses now or in the future?

I spent 6 months on building a web product, and got zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a web product, and got zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I have stuff to post on Reddit very rarely, but I share how my project is going on, random stuff, and memes on X. Just in case few might want to keep in touch 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

The Advantages of a Custom CRM Solution
reddit
LLM Vibe Score0
Human Vibe Score-1
NeerajKumarChaurasiaThis week

The Advantages of a Custom CRM Solution

The growth in the global CRM market continues to accelerate. According to techspective, the global CRM market is now worth \~ $40B USD and is expected to surpass $80B USD by 2025. Despite this phenomenal growth, the CRM market is still dominated by off-the-shelf solutions that are “cookie-cutter” in design and that provide little to no options for customization. These non-customized CRM solutions can significantly inhibit an enterprise’s ability to maximize the advantages of CRM adoption and to realize a robust ROI. As a result, companies are increasingly opting for digital CRM solutions that are customized to meet the unique needs of the enterprise. What is driving the increased demand for custom CRM solutions? What are some of the inherent advantages of a custom CRM solution when compared to a typical off-the-shelf product? Off-the-Shelf CRM Solutions – the Limitations Static CRM solutions are inflexible and self-limiting. Enterprises saddled with these cookie-cutter solutions increasingly report a consistent listing of issues that limit business growth.  These include…. A lack of real-time visibility into shifting customer trends and demands Delayed reaction to coordination of internal resources to meet changing business conditions Lost business opportunities due to lack of flexible, and real-time, opportunity lifecycle management Reporting and dashboarding capabilities that are slow, static, and disconnected Poor quote-to-cash performance that degrades financial performance A CRM investment that delivers poor ROI and that cannot grow with the enterprise All of the above can combine to limit the enterprise’s ability to fully capitalize on its hard-won business opportunities and, over time, limit its ability to create new opportunities. A Customized CRM – What is it? What is a “customized CRM”? Simply put, it is a holistic CRM solution that has been specifically tailored for the individual enterprise. The provider of a truly customized CRM solution will deliver a solution that has been designed to meet the specific—and unique—demands and objectives of the enterprise. A tailored CRM solution will address the enterprise’s sales and operational requirements as well as its customer experience objectives. Unlike standard off-the-shelf CRM providers, a provider of enterprise-grade custom CRM solutions will employ a comprehensive project discovery and requirements gathering process. This is an integral process that provides the foundation for the development of a custom solution that will provide the enterprise with long term flexibility and scalability. A customized digital CRM solution can provide distinct competitive advantages; including: Dynamic, Flexible, Powerful, Real-Time Management and Engagement A customized, technology\-fueled, CRM solution provides the enterprise with the means with which to dynamically engage with customers in ways that build customer loyalty, generate market growth, and drive strong financial performance. Distinct advantages include: Real-time sales opportunity tracking. Helps eliminate lost opportunities due to slow or inadequate reaction. Customized, AI and IoT-fueled, data analytics. A customized CRM solution can be designed to deliver real-time insights. Allows the enterprise to anticipate, and then satisfy, the needs of the customer. Customizable Dashboards and Reports. Widget-based, customized, dashboards and reports that provide real-time data and actionable insights. Sales process automation. Intelligent Workflow-based automation and control of critical sales processes. Increases overall operational efficiency. Outstanding ROI. A custom CRM solution typically delivers superior ROI when compared to off-the-shelf CRM products. Enterprises today spend considerable time, money, and effort in the development of customer relationships. For many enterprises, the continued use of CRM solutions that are rigid and outdated can prove to be impediments to business growth. When considering investment in a new CRM solution any enterprise will be well served by full consideration of a CRM solution that can be fully customized to meet its long-range requirements.

The case for micro PE [x-post from r/micro_pe]
reddit
LLM Vibe Score0
Human Vibe Score0
newy66This week

The case for micro PE [x-post from r/micro_pe]

Any SMB owners considering a sale? What have your challenges been so far? \-- The high-flying venture capital party is quieting down. The pullback in the public tech valuations and high-profile failures have made venture capitalists more cautious, doing fewer deals, no doubt stemming from antsy LPs. But at the same time, real tech has been built that improves business efficiency. AI to cut costs, target customers, improve products. SaaS products to automate everything from billing to marketing. New platforms that open up new modes of customer acquisition. Some of the hyped venture-backed companies from the past decade, while not quite achieving world domination, demonstrated models that provided real value to customers. The on-demand universe - rides, rooms, meals, home services, pets, leisure, showed that customers value convenience and experience. On another front, there's a silver tsunami on the horizon as aging business owners start to cash out. Nearly 60% of private companies are run by the 55+ crowd. Trillions in assets will change hands in the next 15 years as they retire. The tech layoffs have flooded the labor market with brainpower. No shortage of sharp operators looking for their next act. Put it together and you have the ingredients for a new investment approach: micro private equity. Modest valuations, reasonable return expectations, solid companies with positive cash flow or a clear path to profitability. Maybe with debt financing or an acquisition of an existing business at the outset. More targeted, grounded bets are emerging as an alternative to the high-risk venture model. r/micro_pe

40% Of SMBs Still Can't Pay Their Rent, Extending High Delinquency From September Into October
reddit
LLM Vibe Score0
Human Vibe Score1
Aegidius25This week

40% Of SMBs Still Can't Pay Their Rent, Extending High Delinquency From September Into October

https://www.alignable.com/forum/q4s-off-to-a-rough-start-40-of-smbs-still-cant-pay-their-rent October 31, 2023: While the federal government reported a surge in economic growth for the U.S. last week, that news doesn't hold true for many small business owners. In fact, in October polling by Alignable, only 12% said their companies are experiencing significant growth this month. Beyond that, Alignable’s October Rent Report, released today, shows that a whopping 40% of SMBs couldn't even pay their October rent in full and on time. This marks the second consecutive month of a 40% rent delinquency rate -- extending 2023's record high from September through October. These findings are based on responses from 4,246 randomly selected small business owners surveyed from 10/1/23 to 10/30/23, as well as input from 44,000+ other respondents over the past year. As the chart below shows, October's SMB rent delinquency rate is 10 percentage points higher than it was in January, reflecting cumulative economic struggles: increased rents, high interest rates, still-stifling inflation, rising labor costs, and revenues that have declined since this time last year. Rent delinquency rates among small businesses during 2023 based on Alignable surveys So, Why's Rent Delinquency At 40% For A 2nd Month? Here’s the current list of problems contributing to two months' worth of the highest delinquency rate 2023 has seen so far: Consumer Spending Declines On Main Street: Quarterly, we ask about customer spending habits at retailers. This month, 45% of independent Mom and Pop Shops said spending has been down over the last 30 days. Some said it was due to more people spending money online with big retailers like Amazon. This figure is quite high, especially considering that back in July, only 24% reported a drop in consumer spending -- 21 percentage points less severe than it is now. Revenue Troubles: 42% are making half or less of the income they generated monthly prior to COVID. For businesses that are less than three years old, this situation is even worse: 53% of this group reports making half or less of what they generated this time last year. High Interest Rates: Over half of all SMB owners polled said the past 19 months of high interest rates have hurt their margins, reduced revenues, and put their expansion plans on hold, as they don't want to apply for loans. Increased Rent Prices: 50% say they’re being charged more for rent now than they were six months ago, with 15% saying rent has increased by 20% or more. At present, only 37% of pre-COVID businesses have recovered financially from the pandemic era, leaving 63% still striving to make up for time they lost due to COVID, inflationary pressures, and high interest rates. There's a slight silver lining here, though, as the 37% figure is three percentage points higher than it was in September. But, with that said, a recovery rate of 37% after more than three and a half years is still very low and speaks volumes about the ongoing list of troubles small business owners face looking into the rest of 2023. Tech, Manufacturing, Gyms, Beauty & Retail Struggle Examining the rent delinquency landscape in terms of sectors, there's quite a negative shift occurring among some industries in October. Let's look at the charts below to see what's really happening. Sectors most affected by rent delinquency include tech and retail Details on sectors affected by rent delinquency in October This is alarming for a few reasons: The countless technology layoffs at larger companies over the past year appear to be affecting the small companies now, too, who are often dependent on the larger ones as clients. Right now, 54% of science/technology small companies couldn't pay their October rent, up 10 percentage points from September and 16 percentage points since August. There are also some comments in the surveys of technology roles being reduced or replaced by ChatGPT and other AI, which can write software programs. Gyms have been struggling now for a while and now 50% of them can't afford the rent, up 8 percentage points from September. The biggest shift between October and September occurred among manufacturers, partially due to ongoing fluctuation in the price of gas and other inflationary issues. For quite some time, manufacturers were improving a lot in terms of their rent delinquency rates, but in October, they jumped 25 percentage points, doubling their rate, which is now 50%. This is also a record high for manufacturers in 2023. We hope this is just a blip, but we'll see in November. Also due, in part, to fluctuating gas prices and costs of vehicles, 45% of transportation companies couldn't pay October rent in full and on time. That's up 6 percentage points from last month. Sadly, 47% of salon owners couldn't cover October rent, after showing a lot of stability over the past few months. But that stability ended this month, as salons' rent delinquency rates jumped nine percentage points. Though rates have dropped three percentage points in October, a high percentage of retailers are still having trouble paying the rent. Last month, it was 47%. This month, it's better, but is still over 40%, landing at 44%. This is worrisome, especially since Q4 is a "make it or break it" time for many Main Street merchants. Looking more closely at the industries, there was some good news, in that a few others experienced lower delinquency rates in October, including restaurants, which dipped to 40% from 44% in September. Travel/lodging dropped seven percentage points to 38% (from 45% last month), as did education, which is also at 38%, down from 43%. When looking at rent delinquency from the vantage point of the states that are most affected, many surges can be seen between October and September, while a few states saw some dramatic, encouraging declines, too. Rent Troubles Increase For IL, VA, TX, MA, FL, & CO Looking at the states' charts, you can see how tumultuous the rent story has become this fall. Let's first talk about those with significant jumps in their delinquency rates. Here's the rundown: Illinois leads the list once again. After having a better month in September, its delinquency rate has soared, once more, landing at 54% for October (up from 46% last month). In fact, the 54% figure is the highest rate IL-based SMBs have seen in 2023. Virginia was in great shape last month, with a delinquency rate of just 19%. But Virginia-based small business owners have had a very rough month, at least in terms of rent. Now, 50% of them who took our poll say they couldn't cover rent (an increase of 31 percentage points). Texas is third on the list, with an 11-percentage-point lift from 38% in September to 49% in October. MA is next up at 48%, which marks the largest jump on the chart -- 32 percentage points from a low of just 16% in September. Small businesses in Florida have also experienced two challenging months in terms of rent delinquency. Right now, 45% of SMBs there couldn't afford to pay, up nine percentage points from September and 15 percentage points from August. Colorado's businesses regressed in October, hitting a new record high of 40%. That rent delinquency rate jumped 13 percentage points from September to October. While we just covered states with some very high delinquency rates, there were also several more positive swings that have occurred in October. Though encouraging, we'll have to see how long those delinquency rates continue. Here are the most remarkable: New York -- After reaching a record rate of 55% last month, New York's small business owners now report a more stable number: just 29%. That's down 26 percentage points. New Jersey -- New York's neighbor has an even more impressive story in October: only 20% of New Jersey's SMBs couldn't pay rent this month, a record low over at least the past 14 months, down 34 percentage points from a record high of 54%. Michigan -- Similarly, Michigan's small business owners boast a rate of just 20%, down from 45% in September.

Need Advice on Implementing Reranking Models for an AI-Based Document-Specific Copilot feature
reddit
LLM Vibe Score0
Human Vibe Score1
Swimming_Teach_7579This week

Need Advice on Implementing Reranking Models for an AI-Based Document-Specific Copilot feature

Hey everyone! I'm currently working on an AI-based grant writing system that includes two main features: Main AI: Uses LLMs to generate grant-specific suggestions based on user-uploaded documents. Copilot Feature: Allows document-specific Q&A by utilizing a query format like /{filename} {query} to fetch information from the specified document. Currently, we use FAISS for vector storage and retrieval, with metadata managed through .pkl files. This setup works for similarity-based retrieval of relevant content. However, I’m considering introducing a reranking model to further enhance retrieval accuracy, especially for our Copilot feature. Challenges with Current Setup: Document-Specific Retrieval: We're storing document-specific embeddings and metadata in .pkl files, and retrieval works by first querying FAISS. Objective: Improve the precision of the results retrieved by Copilot when the user requests data from a specific document (e.g., /example.pdf summarize content). Questions for the Community: Is using a reranking model (e.g., BERT-based reranker, MiniLM) a good idea to add another layer of precision for document retrieval, especially when handling specific document requests? If I implement a reranking model, do I still need the structured .pkl files, or can I rely solely on the embeddings and reranking for retrieval? How can I effectively integrate a reranking model into my current FAISS + Langchain setup? I’d love to hear your thoughts, and if you have experience in using reranking models with FAISS or similar, any advice would be highly appreciated. Thank you!

Study Plan for Learning Data Science Over the Next 12 Months [D]
reddit
LLM Vibe Score0
Human Vibe Score1
daniel-dataThis week

Study Plan for Learning Data Science Over the Next 12 Months [D]

In this thread, I address a study plan for 2021. In case you're interested, I wrote a whole article about this topic: Study Plan for Learning Data Science Over the Next 12 Months Let me know your thoughts on this. &#x200B; https://preview.redd.it/emg20nzhet661.png?width=1170&format=png&auto=webp&s=cf09e4dc5e82ba2fd7b57c706ba2873be57fe8de We are ending 2020 and it is time to make plans for next year, and one of the most important plans and questions we must ask is what do we want to study?, what do we want to enhance?, what changes do we want to make?, and what is the direction we are going to take (or continue) in our professional careers?. Many of you will be starting on the road to becoming a data scientist, in fact you may be evaluating it, since you have heard a lot about it, but you have some doubts, for example about the amount of job offers that may exist in this area, doubts about the technology itself, and about the path you should follow, considering the wide range of options to learn. I’m a believer that we should learn from various sources, from various mentors, and from various formats. By sources I mean the various virtual platforms and face-to-face options that exist to study. By mentors I mean that it is always a good idea to learn from different points of view and learning from different teachers/mentors, and by formats I mean the choices between books, videos, classes, and other formats where the information is contained. When we extract information from all these sources we reinforce the knowledge learned, but we always need a guide, and this post aims to give you some practical insights and strategies in this regard. To decide on sources, mentors and formats it is up to you to choose. It depends on your preferences and ease of learning: for example, some people are better at learning from books, while others prefer to learn from videos. Some prefer to study on platforms that are practical (following online code), and others prefer traditional platforms: like those at universities (Master’s Degree, PHDs or MOOCs). Others prefer to pay for quality content, while others prefer to look only for free material. That’s why I won’t give a specific recommendation in this post, but I’ll give you the whole picture: a study plan. To start you should consider the time you’ll spend studying and the depth of learning you want to achieve, because if you find yourself without a job you could be available full time to study, which is a huge advantage. On the other hand, if you are working, you’ll have less time and you’ll have to discipline yourself to be able to have the time available in the evenings, mornings or weekends. Ultimately, the important thing is to meet the goal of learning and perhaps dedicating your career to this exciting area! We will divide the year into quarters as follows First Quarter: Learning the Basics Second Quarter: Upgrading the Level: Intermediate Knowledge Third Quarter: A Real World Project — A Full-stack Project Fourth Quarter: Seeking Opportunities While Maintaining Practice First Quarter: Learning the Basics &#x200B; https://preview.redd.it/u7t9bthket661.png?width=998&format=png&auto=webp&s=4ad29cb43618e7acf793259243aa5a60a8535f0a If you want to be more rigorous you can have start and end dates for this period of study of the bases. It could be something like: From January 1 to March 30, 2021 as deadline. During this period you will study the following: A programming language that you can apply to data science: Python or R. We recommend Python due to the simple fact that approximately 80% of data science job offers ask for knowledge in Python. That same percentage is maintained with respect to the real projects you will find implemented in production. And we add the fact that Python is multipurpose, so you won’t “waste” your time if at some point you decide to focus on web development, for example, or desktop development. This would be the first topic to study in the first months of the year. Familiarize yourself with statistics and mathematics. There is a big debate in the data science community about whether we need this foundation or not. I will write a post later on about this, but the reality is that you DO need it, but ONLY the basics (at least in the beginning). And I want to clarify this point before continuing. We could say that data science is divided in two big fields: Research on one side and putting Machine Learning algorithms into production on the other side. If you later decide to focus on Research then you are going to need mathematics and statistics in depth (very in depth). If you are going to go for the practical part, the libraries will help you deal with most of it, under the hood. It should be noted that most job offers are in the practical part. For both cases, and in this first stage you will only need the basics of: Statistics (with Python and NumPy) Descriptive statistics Inferential Statistics Hypothesis testing Probability Mathematics (with Python and NumPy) Linear Algebra (For example: SVD) Multivariate Calculus Calculus (For example: gradient descent) Note: We recommend that you study Python first before seeing statistics and mathematics, because the challenge is to implement these statistical and mathematical bases with Python. Don’t look for theoretical tutorials that show only slides or statistical and/or mathematical examples in Excel/Matlab/Octave/SAS and other different to Python or R, it gets very boring and impractical! You should choose a course, program or book that teaches these concepts in a practical way and using Python. Remember that Python is what we finally use, so you need to choose well. This advice is key so you don’t give up on this part, as it will be the most dense and difficult. If you have these basics in the first three months, you will be ready to make a leap in your learning for the next three months. Second Quarter: Upgrading the Level: Intermediate Knowledge &#x200B; https://preview.redd.it/y1y55vynet661.png?width=669&format=png&auto=webp&s=bd3e12bb112943025c39a8975faf4d64514df275 If you want to be more rigorous you can have start and end dates for this period of study at the intermediate level. It could be something like: From April 1 to June 30, 2021 as deadline. Now that you have a good foundation in programming, statistics and mathematics, it is time to move forward and learn about the great advantages that Python has for applying data analysis. For this stage you will be focused on: Data science Python stack Python has the following libraries that you should study, know and practice at this stage Pandas: for working with tabular data and make in-depth analysis Matplotlib and Seaborn: for data visualization Pandas is the in-facto library for data analysis, it is one of the most important (if not the most important) and powerful tools you should know and master during your career as a data scientist. Pandas will make it much easier for you to manipulate, cleanse and organize your data. Feature Engineering Many times people don’t go deep into Feature Engineering, but if you want to have Machine Learning models that make good predictions and improve your scores, spending some time on this subject is invaluable! Feature engineering is the process of using domain knowledge to extract features from raw data using data mining techniques. These features can be used to improve the performance of machine learning algorithms. Feature engineering can be considered as applied machine learning itself. To achieve the goal of good feature engineering you must know the different techniques that exist, so it is a good idea to at least study the main ones. Basic Models of Machine Learning At the end of this stage you will start with the study of Machine Learning. This is perhaps the most awaited moment! This is where you start to learn about the different algorithms you can use, which particular problems you can solve and how you can apply them in real life. The Python library we recommend you to start experimenting with ML is: scikit-learn. However it is a good idea that you can find tutorials where they explain the implementation of the algorithms (at least the simplest ones) from scratch with Python, since the library could be a “Black Box” and you might not understand what is happening under the hood. If you learn how to implement them with Python, you can have a more solid foundation. If you implement the algorithms with Python (without a library), you will put into practice everything seen in the statistics, mathematics and Pandas part. These are some recommendations of the algorithms that you should at least know in this initial stage Supervised learning Simple Linear Regression Multiple Linear Regression K-nearest neighbors (KNN) Logistic Regression Decision Trees Random Forest Unsupervised Learning K-Means PCA Bonus: if you have the time and you are within the time ranges, you can study these others Gradient Boosting Algorithms GBM XGBoost LightGBM CatBoost Note: do not spend more than the 3 months stipulated for this stage. Because you will be falling behind and not complying with the study plan. We all have shortcomings at this stage, it is normal, go ahead and then you can resume some concepts that did not understand in detail. The important thing is to have the basic knowledge and move forward! If at least you succeed to study the mentioned algorithms of supervised and unsupervised learning, you will have a very clear idea of what you will be able to do in the future. So don’t worry about covering everything, remember that it is a process, and ideally you should have some clearly established times so that you don’t get frustrated and feel you are advancing. So far, here comes your “theoretical” study of the basics of data science. Now we’ll continue with the practical part! Third Quarter: A Real World Project — A Full-stack Project &#x200B; https://preview.redd.it/vrn783vqet661.png?width=678&format=png&auto=webp&s=664061b3d33b34979b74b10b9f8a3d0f7b8b99ee If you want to be more rigorous you can have start and end dates for this period of study at the intermediate level. It could be something like: From July 1 to September 30, 2021 as deadline. Now that you have a good foundation in programming, statistics, mathematics, data analysis and machine learning algorithms, it is time to move forward and put into practice all this knowledge. Many of these suggestions may sound out of the box, but believe me they will make a big difference in your career as a data scientist. The first thing is to create your web presence: Create a Github (or GitLab) account, and learn Git*. Being able to manage different versions of your code is important, you should have version control over them, not to mention that having an active Github account is very valuable in demonstrating your true skills. On Github, you can also set up your Jupyter Notebooks and make them public, so you can show off your skills as well. This is mine for example: https://github.com/danielmoralesp Learn the basics of web programming*. The advantage is that you already have Python as a skill, so you can learn Flask to create a simple web page. Or you can use a template engine like Github Pages, Ghost or Wordpress itself and create your online portfolio. Buy a domain with your name*. Something like myname.com, myname.co, myname.dev, etc. This is invaluable so you can have your CV online and update it with your projects. There you can make a big difference, showing your projects, your Jupyter Notebooks and showing that you have the practical skills to execute projects in this area. There are many front-end templates for you to purchase for free or for payment, and give it a more personalized and pleasant look. Don’t use free sub-domains of Wordpress, Github or Wix, it looks very unprofessional, make your own. Here is mine for example: https://www.danielmorales.dev/ Choose a project you are passionate about and create a Machine Learning model around it. The final goal of this third quarter is to create ONE project, that you are passionate about, and that is UNIQUE among others. It turns out that there are many typical projects in the community, such as predicting the Titanic Survivors, or predicting the price of Houses in Boston. Those kinds of projects are good for learning, but not for showing off as your UNIQUE projects. If you are passionate about sports, try predicting the soccer results of your local league. If you are passionate about finance, try predicting your country’s stock market prices. If you are passionate about marketing, try to find someone who has an e-commerce and implement a product recommendation algorithm and upload it to production. If you are passionate about business: make a predictor of the best business ideas for 2021 :) As you can see, you are limited by your passions and your imagination. In fact, those are the two keys for you to do this project: Passion and Imagination. However don’t expect to make money from it, you are in a learning stage, you need that algorithm to be deployed in production, make an API in Flask with it, and explain in your website how you did it and how people can access it. This is the moment to shine, and at the same time it’s the moment of the greatest learning. You will most likely face obstacles, if your algorithm gives 60% of Accuracy after a huge optimization effort, it doesn’t matter, finish the whole process, deploy it to production, try to get a friend or family member to use it, and that will be the goal achieved for this stage: Make a Full-stack Machine Learning project. By full-stack I mean that you did all the following steps: You got the data from somewhere (scrapping, open data or API) You did a data analysis You cleaned and transformed the data You created Machine Learning Models You deployed the best model to production for other people to use. This does not mean that this whole process is what you will always do in your daily job, but it does mean that you will know every part of the pipeline that is needed for a data science project for a company. You will have a unique perspective! Fourth Quarter: Seeking Opportunities While Maintaining Practice &#x200B; https://preview.redd.it/qd0osystet661.png?width=1056&format=png&auto=webp&s=2da456b15985b2793041256f5e45bca99a23b51a If you want to be more rigorous you can have start and end dates for this period of study at the final level. It could be something like: From October 1 to December 31, 2021 as deadline. Now you have theoretical and practical knowledge. You have implemented a model in production. The next step depends on you and your personality. Let’s say you are an entrepreneur, and you have the vision to create something new from something you discovered or saw an opportunity to do business with this discipline, so it’s time to start planning how to do it. If that’s the case, obviously this post won’t cover that process, but you should know what the steps might be (or start figuring them out). But if you are one of those who want to get a job as a data scientist, here is my advice. Getting a job as a data scientist “You’re not going to get a job as fast as you think, if you keep thinking the same way”.Author It turns out that all people who start out as data scientists imagine themselves working for the big companies in their country or region. Or even remote. It turns out that if you aspire to work for a large company like data scientist you will be frustrated by the years of experience they ask for (3 or more years) and the skills they request. Large companies don’t hire Juniors (or very few do), precisely because they are already large companies. They have the financial muscle to demand experience and skills and can pay a commensurate salary (although this is not always the case). The point is that if you focus there you’re going to get frustrated! Here we must return to the following advise: “You need creativity to get a job in data science”. Like everything else in life we have to start at different steps, in this case, from the beginning. Here are the scenarios If you are working in a company and in a non-engineering role you must demonstrate your new skills to the company you are working for*. If you are working in the customer service area, you should apply it to your work, and do for example, detailed analysis of your calls, conversion rates, store data and make predictions about it! If you can have data from your colleagues, you could try to predict their sales! This may sound funny, but it’s about how creatively you can apply data science to your current work and how to show your bosses how valuable it is and EVANGELIZE them about the benefits of implementation. You’ll be noticed and they could certainly create a new data related department or job. And you already have the knowledge and experience. The key word here is Evangelize. Many companies and entrepreneurs are just beginning to see the power of this discipline, and it is your task to nurture that reality. If you are working in an area related to engineering, but that is not data science*. Here the same applies as the previous example, but you have some advantages, and that is that you could access the company’s data, and you could use it for the benefit of the company, making analyses and/or predictions about it, and again EVANGELIZING your bosses your new skills and the benefits of data science. If you are unemployed (or do not want, or do not feel comfortable following the two examples above)*, you can start looking outside, and what I recommend is that you look for technology companies and / or startups where they are just forming the first teams and are paying some salary, or even have options shares of the company. Obviously here the salaries will not be exorbitant, and the working hours could be longer, but remember that you are in the learning and practice stage (just in the first step), so you can not demand too much, you must land your expectations and fit that reality, and stop pretending to be paid $ 10,000 a month at this stage. But, depending of your country $1.000 USD could be something very interesting to start this new career. Remember, you are a Junior at this stage. The conclusion is: don’t waste your time looking at and/or applying to offers from big companies, because you will get frustrated. Be creative, and look for opportunities in smaller or newly created companies. Learning never stops While you are in that process of looking for a job or an opportunity, which could take half of your time (50% looking for opportunities, 50% staying in practice), you have to keep learning, you should advance to concepts such as Deep Learning, Data Engineer or other topics that you feel were left loose from the past stages or focus on the topics that you are passionate about within this group of disciplines in data science. At the same time you can choose a second project, and spend some time running it from end-to-end, and thus increase your portfolio and your experience. If this is the case, try to find a completely different project: if the first one was done with Machine Learning, let this second one be done with Deep learning. If the first one was deployed to a web page, that this second one is deployed to a mobile platform. Remember, creativity is the key! Conclusion We are at an ideal time to plan for 2021, and if this is the path you want to take, start looking for the platforms and media you want to study on. Get to work and don’t miss this opportunity to become a data scientist in 2021! Note: we are building a private community in Slack of data scientist, if you want to join us write to the email: support@datasource.ai I hope you enjoyed this reading! you can follow me on twitter or linkedin Thank you for reading!

MMML | Deploy HuggingFace training model rapidly based on MetaSpore
reddit
LLM Vibe Score0
Human Vibe Score1
qazmkoppThis week

MMML | Deploy HuggingFace training model rapidly based on MetaSpore

A few days ago, HuggingFace announced a $100 million Series C funding round, which was big news in open source machine learning and could be a sign of where the industry is headed. Two days before the HuggingFace funding announcement, open-source machine learning platform MetaSpore released a demo based on the HuggingFace Rapid deployment pre-training model. As deep learning technology makes innovative breakthroughs in computer vision, natural language processing, speech understanding, and other fields, more and more unstructured data are perceived, understood, and processed by machines. These advances are mainly due to the powerful learning ability of deep learning. Through pre-training of deep models on massive data, the models can capture the internal data patterns, thus helping many downstream tasks. With the industry and academia investing more and more energy in the research of pre-training technology, the distribution warehouses of pre-training models such as HuggingFace and Timm have emerged one after another. The open-source community release pre-training significant model dividends at an unprecedented speed. In recent years, the data form of machine modeling and understanding has gradually evolved from single-mode to multi-mode, and the semantic gap between different modes is being eliminated, making it possible to retrieve data across modes. Take CLIP, OpenAI’s open-source work, as an example, to pre-train the twin towers of images and texts on a dataset of 400 million pictures and texts and connect the semantics between pictures and texts. Many researchers in the academic world have been solving multimodal problems such as image generation and retrieval based on this technology. Although the frontier technology through the semantic gap between modal data, there is still a heavy and complicated model tuning, offline data processing, high performance online reasoning architecture design, heterogeneous computing, and online algorithm be born multiple processes and challenges, hindering the frontier multimodal retrieval technologies fall to the ground and pratt &whitney. DMetaSoul aims at the above technical pain points, abstracting and uniting many links such as model training optimization, online reasoning, and algorithm experiment, forming a set of solutions that can quickly apply offline pre-training model to online. This paper will introduce how to use the HuggingFace community pre-training model to conduct online reasoning and algorithm experiments based on MetaSpore technology ecology so that the benefits of the pre-training model can be fully released to the specific business or industry and small and medium-sized enterprises. And we will give the text search text and text search graph two multimodal retrieval demonstration examples for your reference. Multimodal semantic retrieval The sample architecture of multimodal retrieval is as follows: Our multimodal retrieval system supports both text search and text search application scenarios, including offline processing, model reasoning, online services, and other core modules: https://preview.redd.it/mdyyv1qmdz291.png?width=1834&format=png&auto=webp&s=e9e10710794c78c64cc05adb75db385aa53aba40 Offline processing, including offline data processing processes for different application scenarios of text search and text search, including model tuning, model export, data index database construction, data push, etc. Model inference. After the offline model training, we deployed our NLP and CV large models based on the MetaSpore Serving framework. MetaSpore Serving helps us conveniently perform online inference, elastic scheduling, load balancing, and resource scheduling in heterogeneous environments. Online services. Based on MetaSpore’s online algorithm application framework, MetaSpore has a complete set of reusable online search services, including Front-end retrieval UI, multimodal data preprocessing, vector recall and sorting algorithm, AB experimental framework, etc. MetaSpore also supports text search by text and image scene search by text and can be migrated to other application scenarios at a low cost. The HuggingFace open source community has provided several excellent baseline models for similar multimodal retrieval problems, which are often the starting point for actual optimization in the industry. MetaSpore also uses the pre-training model of the HuggingFace community in its online services of searching words by words and images by words. Searching words by words is based on the semantic similarity model of the question and answer field optimized by MetaSpore, and searching images by words is based on the community pre-training model. These community open source pre-training models are exported to the general ONNX format and loaded into MetaSpore Serving for online reasoning. The following sections will provide a detailed description of the model export and online retrieval algorithm services. The reasoning part of the model is standardized SAAS services with low coupling with the business. Interested readers can refer to my previous post: The design concept of MetaSpore, a new generation of the one-stop machine learning platform. 1.1 Offline Processing Offline processing mainly involves the export and loading of online models and index building and pushing of the document library. You can follow the step-by-step instructions below to complete the offline processing of text search and image search and see how the offline pre-training model achieves reasoning at MetaSpore. 1.1.1 Search text by text Traditional text retrieval systems are based on literal matching algorithms such as BM25. Due to users’ diverse query words, a semantic gap between query words and documents is often encountered. For example, users misspell “iPhone” as “Phone,” and search terms are incredibly long, such as “1 \~ 3 months old baby autumn small size bag pants”. Traditional text retrieval systems will use spelling correction, synonym expansion, search terms rewriting, and other means to alleviate the semantic gap but fundamentally fail to solve this problem. Only when the retrieval system fully understands users’ query terms and documents can it meet users’ retrieval demands at the semantic level. With the continuous progress of pre-training and representational learning technology, some commercial search engines continue to integrate semantic vector retrieval methods based on symbolic learning into the retrieval ecology. Semantic retrieval model This paper introduces a set of semantic vector retrieval applications. MetaSpore built a set of semantic retrieval systems based on encyclopedia question and answer data. MetaSpore adopted the Sentence-Bert model as the semantic vector representation model, which fine-tunes the twin tower BERT in supervised or unsupervised ways to make the model more suitable for retrieval tasks. The model structure is as follows: The query-Doc symmetric two-tower model is used in text search and question and answer retrieval. The vector representation of online Query and offline DOC share the same vector representation model, so it is necessary to ensure the consistency of the offline DOC library building model and online Query inference model. The case uses MetaSpore’s text representation model Sbert-Chinese-QMC-domain-V1, optimized in the open-source semantically similar data set. This model will express the question and answer data as a vector in offline database construction. The user query will be expressed as a vector by this model in online retrieval, ensuring that query-doc in the same semantic space, users’ semantic retrieval demands can be guaranteed by vector similarity metric calculation. Since the text presentation model does vector encoding for Query online, we need to export the model for use by the online service. Go to the q&A data library code directory and export the model concerning the documentation. In the script, Pytorch Tracing is used to export the model. The models are exported to the “./export “directory. The exported models are mainly ONNX models used for wired reasoning, Tokenizer, and related configuration files. The exported models are loaded into MetaSpore Serving by the online Serving system described below for model reasoning. Since the exported model will be copied to the cloud storage, you need to configure related variables in env.sh. \Build library based on text search \ The retrieval database is built on the million-level encyclopedia question and answer data set. According to the description document, you need to download the data and complete the database construction. The question and answer data will be coded as a vector by the offline model, and then the database construction data will be pushed to the service component. The whole process of database construction is described as follows: Preprocessing, converting the original data into a more general JSonline format for database construction; Build index, use the same model as online “sbert-Chinese-qmc-domain-v1” to index documents (one document object per line); Push inverted (vector) and forward (document field) data to each component server. The following is an example of the database data format. After offline database construction is completed, various data are pushed to corresponding service components, such as Milvus storing vector representation of documents and MongoDB storing summary information of documents. Online retrieval algorithm services will use these service components to obtain relevant data. 1.1.2 Search by text Text and images are easy for humans to relate semantically but difficult for machines. First of all, from the perspective of data form, the text is the discrete ID type of one-dimensional data based on words and words. At the same time, images are continuous two-dimensional or three-dimensional data. Secondly, the text is a subjective creation of human beings, and its expressive ability is vibrant, including various turning points, metaphors, and other expressions, while images are machine representations of the objective world. In short, bridging the semantic gap between text and image data is much more complex than searching text by text. The traditional text search image retrieval technology generally relies on the external text description data of the image or the nearest neighbor retrieval technology and carries out the retrieval through the image associated text, which in essence degrades the problem to text search. However, it will also face many issues, such as obtaining the associated text of pictures and whether the accuracy of text search by text is high enough. The depth model has gradually evolved from single-mode to multi-mode in recent years. Taking the open-source project of OpenAI, CLIP, as an example, train the model through the massive image and text data of the Internet and map the text and image data into the same semantic space, making it possible to implement the text and image search technology based on semantic vector. CLIP graphic model The text search pictures introduced in this paper are implemented based on semantic vector retrieval, and the CLIP pre-training model is used as the two-tower retrieval architecture. Because the CLIP model has trained the semantic alignment of the twin towers’ text and image side models on the massive graphic and text data, it is particularly suitable for the text search graph scene. Due to the different image and text data forms, the Query-Doc asymmetric twin towers model is used for text search image retrieval. The image-side model of the twin towers is used for offline database construction, and the text-side model is used for the online return. In the final online retrieval, the database data of the image side model will be searched after the text side model encodes Query, and the CLIP pre-training model guarantees the semantic correlation between images and texts. The model can draw the graphic pairs closer in vector space by pre-training on a large amount of visual data. Here we need to export the text-side model for online MetaSpore Serving inference. Since the retrieval scene is based on Chinese, the CLIP model supporting Chinese understanding is selected. The exported content includes the ONNX model used for online reasoning and Tokenizer, similar to the text search. MetaSpore Serving can load model reasoning through the exported content. Build library on Image search You need to download the Unsplash Lite library data and complete the construction according to the instructions. The whole process of database construction is described as follows: Preprocessing, specify the image directory, and then generate a more general JSOnline file for library construction; Build index, use OpenAI/Clip-Vit-BASE-Patch32 pre-training model to index the gallery, and output one document object for each line of index data; Push inverted (vector) and forward (document field) data to each component server. Like text search, after offline database construction, relevant data will be pushed to service components, called by online retrieval algorithm services to obtain relevant data. 1.2 Online Services The overall online service architecture diagram is as follows: &#x200B; https://preview.redd.it/nz8zrbbpdz291.png?width=1280&format=png&auto=webp&s=28dae7e031621bc8819519667ed03d8d085d8ace Multi-mode search online service system supports application scenarios such as text search and text search. The whole online service consists of the following parts: Query preprocessing service: encapsulate preprocessing logic (including text/image, etc.) of pre-training model, and provide services through gRPC interface; Retrieval algorithm service: the whole algorithm processing link includes AB experiment tangent flow configuration, MetaSpore Serving call, vector recall, sorting, document summary, etc.; User entry service: provides a Web UI interface for users to debug and track down problems in the retrieval service. From a user request perspective, these services form invocation dependencies from back to front, so to build up a multimodal sample, you need to run each service from front to back first. Before doing this, remember to export the offline model, put it online and build the library first. This article will introduce the various parts of the online service system and make the whole service system step by step according to the following guidance. See the ReadME at the end of this article for more details. 1.2.1 Query preprocessing service Deep learning models tend to be based on tensors, but NLP/CV models often have a preprocessing part that translates raw text and images into tensors that deep learning models can accept. For example, NLP class models often have a pre-tokenizer to transform text data of string type into discrete tensor data. CV class models also have similar processing logic to complete the cropping, scaling, transformation, and other processing of input images through preprocessing. On the one hand, considering that this part of preprocessing logic is decoupled from tensor reasoning of the depth model, on the other hand, the reason of the depth model has an independent technical system based on ONNX, so MetaSpore disassembled this part of preprocessing logic. NLP pretreatment Tokenizer has been integrated into the Query pretreatment service. MetaSpore dismantlement with a relatively general convention. Users only need to provide preprocessing logic files to realize the loading and prediction interface and export the necessary data and configuration files loaded into the preprocessing service. Subsequent CV preprocessing logic will also be integrated in this manner. The preprocessing service currently provides the gRPC interface invocation externally and is dependent on the Query preprocessing (QP) module in the retrieval algorithm service. After the user request reaches the retrieval algorithm service, it will be forwarded to the service to complete the data preprocessing and continue the subsequent processing. The ReadMe provides details on how the preprocessing service is started, how the preprocessing model exported offline to cloud storage enters the service, and how to debug the service. To further improve the efficiency and stability of model reasoning, MetaSpore Serving implements a Python preprocessing submodule. So MetaSpore can provide gRPC services through user-specified preprocessor.py, complete Tokenizer or CV-related preprocessing in NLP, and translate requests into a Tensor that deep models can handle. Finally, the model inference is carried out by MetaSpore, Serving subsequent sub-modules. Presented here on the lot code: https://github.com/meta-soul/MetaSpore/compare/add\python\preprocessor 1.2.2 Retrieval algorithm services Retrieval algorithm service is the core of the whole online service system, which is responsible for the triage of experiments, the assembly of algorithm chains such as preprocessing, recall, sorting, and the invocation of dependent component services. The whole retrieval algorithm service is developed based on the Java Spring framework and supports multi-mode retrieval scenarios of text search and text search graph. Due to good internal abstraction and modular design, it has high flexibility and can be migrated to similar application scenarios at a low cost. Here’s a quick guide to configuring the environment to set up the retrieval algorithm service. See ReadME for more details: Install dependent components. Use Maven to install the online-Serving component Search for service configurations. Copy the template configuration file and replace the MongoDB, Milvus, and other configurations based on the development/production environment. Install and configure Consul. Consul allows you to synchronize the search service configuration in real-time, including cutting the flow of experiments, recall parameters, and sorting parameters. The project’s configuration file shows the current configuration parameters of text search and text search. The parameter modelName in the stage of pretreatment and recall is the corresponding model exported in offline processing. Start the service. Once the above configuration is complete, the retrieval service can be started from the entry script. Once the service is started, you can test it! For example, for a user with userId=10 who wants to query “How to renew ID card,” access the text search service. 1.2.3 User Entry Service Considering that the retrieval algorithm service is in the form of the API interface, it is difficult to locate and trace the problem, especially for the text search image scene can intuitively display the retrieval results to facilitate the iterative optimization of the retrieval algorithm. This paper provides a lightweight Web UI interface for text search and image search, a search input box, and results in a display page for users. Developed by Flask, the service can be easily integrated with other retrieval applications. The service calls the retrieval algorithm service and displays the returned results on the page. It’s also easy to install and start the service. Once you’re done, go to http://127.0.0.1:8090 to see if the search UI service is working correctly. See the ReadME at the end of this article for details. Multimodal system demonstration The multimodal retrieval service can be started when offline processing and online service environment configuration have been completed following the above instructions. Examples of textual searches are shown below. Enter the entry of the text search map application, enter “cat” first, and you can see that the first three digits of the returned result are cats: https://preview.redd.it/d7syq47rdz291.png?width=1280&format=png&auto=webp&s=b43df9abd380b7d9a52e3045dd787f4feeb69635 If you add a color constraint to “cat” to retrieve “black cat,” you can see that it does return a black cat: &#x200B; https://preview.redd.it/aa7pxx8tdz291.png?width=1280&format=png&auto=webp&s=e3727c29d1bde6eea2e1cccf6c46d3cae3f4750e Further, strengthen the constraint on the search term, change it to “black cat on the bed,” and return results containing pictures of a black cat climbing on the bed: &#x200B; https://preview.redd.it/2mw4qpjudz291.png?width=1280&format=png&auto=webp&s=1cf1db667892b9b3a40451993680fbd6980b5520 The cat can still be found through the text search system after the color and scene modification in the above example. Conclusion The cutting-edge pre-training technology can bridge the semantic gap between different modes, and the HuggingFace community can greatly reduce the cost for developers to use the pre-training model. Combined with the technological ecology of MetaSpore online reasoning and online microservices provided by DMetaSpore, the pre-training model is no longer mere offline dabbling. Instead, it can truly achieve end-to-end implementation from cutting-edge technology to industrial scenarios, fully releasing the dividends of the pre-training large model. In the future, DMetaSoul will continue to improve and optimize the MetaSpore technology ecosystem: More automated and wider access to HuggingFace community ecology. MetaSpore will soon release a common model rollout mechanism to make HuggingFace ecologically accessible and will later integrate preprocessing services into online services. Multi-mode retrieval offline algorithm optimization. For multimodal retrieval scenarios, MetaSpore will continuously iteratively optimize offline algorithm components, including text recall/sort model, graphic recall/sort model, etc., to improve the accuracy and efficiency of the retrieval algorithm. For related code and reference documentation in this article, please visit: https://github.com/meta-soul/MetaSpore/tree/main/demo/multimodal/online Some images source: https://github.com/openai/CLIP/raw/main/CLIP.png https://www.sbert.net/examples/training/sts/README.html

6 principles to data architecture that facilitate innovation
reddit
LLM Vibe Score0
Human Vibe Score1
Competitive_Speech36This week

6 principles to data architecture that facilitate innovation

My team and I have been re-building our company's data architecture. In the process of doing so, I got together six key principles to transforming data architectures and thought I would share them, as a strong data architecture is crucial for businesses looking to stay competitive in the digital landscape, as it improves decision-making, time to market, and data security. When executed with efficiency, a resilient data architecture unleashes unparalleled degrees of agility. Principle 1: Agility and flexibility To quickly adjust to market fluctuations, businesses must create adaptable data infrastructures that can effortlessly manage an ever-growing influx of data. To accomplish this objective, we recommend to our clients to implement Enterprise Service Bus, Enterprise Data Warehouse, and Master Data Management integrated together. &#x200B; I believe the best option is this: \- By centralizing communication, ESB reduces the time and effort required to integrate new systems; \- EDW consolidates data from different sources, resulting in a 50% reduction in software implementation time; \- Finally, MDM ensures consistency and accuracy across the organization, leading to better decision-making and streamlined operations. Implementing these solutions can lead to reduced software implementation time, better ROI, and more manageable data architecture. By fostering a culture of collaboration and adopting modern technologies and practices, businesses can prioritize agility and flexibility in their data architecture to increase the pace of innovation. Principle 2: Modularity and reusability Data architecture that fosters modularity and reusability is essential for accelerating innovation within an organization. By breaking data architecture components into smaller, more manageable pieces, businesses can enable different teams to leverage existing architecture components, reducing redundancy and improving overall efficiency. MDM can promote modularity and reusability by creating a central repository for critical business data. This prevents duplication and errors, improving efficiency and decision-making. MDM enables a single source of truth for data, accessible across multiple systems, which promotes integration and scalability. MDM also provides standardized data models, rules, and governance policies that reduce development time, increase quality, and ensure proper management throughout the data’s lifecycle. Another way to achieve modularity in data architecture is through the use of microservices and scripts for Extract, Transform, and Load (ETL) processes. Adopting a structured methodology and framework can ensure these components are well-organized, making it easier for teams to collaborate and maintain the system. Microservices can also contribute to modularity and reusability in data architecture. These small, independent components can be developed, deployed, and scaled independently of one another. By utilizing microservices, organizations can update or replace individual components without affecting the entire system, improving flexibility and adaptability. Principle 3: Data quality and consistency The efficiency of operations depends on data’s quality, so a meticulously crafted data architecture plays a pivotal role in preserving it, empowering enterprises to make well-informed decisions based on credible information. Here are some key factors to consider that will help your company ensure quality: \- Implementing Master Data Management (MDM) – this way, by consolidating, cleansing, and standardizing data from multiple sources, your IT department will be able to create a single, unified view of the most important data entities (customers, products, and suppliers); \- Assigning data stewardship responsibilities to a small team or an individual specialist; \- Considering implementing data validation, data lineage, and data quality metrics; \- By implementing MDM and adopting a minimal data stewardship approach, organizations can maintain high-quality data that drives innovation and growth. Principle 4: Data governance Data governance is a strategic framework that goes beyond ensuring data quality and consistency. It includes ensuring data security, privacy, accessibility, regulatory compliance, and lifecycle management. Here are some key aspects of data governance: \- Implementing robust measures and controls to protect sensitive data from unauthorized access, breaches, and theft. This is only possible through including encryption, access controls, and intrusion detection systems into your company’s IT architecture; \- Adhering to data privacy regulations and guidelines, such as the General Data Protection Regulation (GDPR) and the California Consumer Privacy Act (CCPA); \- Defining stringent conditions for who has access to specific data assets to maintain control over data and ensure its accessibility only for legitimate purposes. Managing the entire lifecycle of data, from creation and storage to archiving and disposal, including defining policies for data retention, archiving, and deletion in compliance with legal and regulatory requirements. To facilitate effective data governance, organizations can leverage various tools and technologies, such as: \- Data cataloging tools: Solutions like Collibra, Alation, or Informatica Enterprise Data Catalog help organizations discover, understand, and manage their data assets. \- Data lineage tools: Tools like Talend, IBM InfoSphere, or Apache Atlas help track data’s origin, transformation, and usage, providing insights into data quality issues and potential areas for improvement. \- Data quality tools: Solutions like Informatica Data Quality, Trifacta, or SAS Data Quality help organizations maintain high-quality data by identifying and correcting errors, inconsistencies, and inaccuracies. \- Data security and privacy tools: Tools like Varonis, BigID, or Spirion help protect sensitive data and ensure compliance with data privacy regulations. Principle 5: Cloud-first approach A cloud-first approach prioritizes cloud-based solutions over on-premises ones when it comes to data management. Cloud-based data management pros: \- Virtually limitless scalability, so that organizations can grow and adapt to changing data requirements without significant infrastructure investments; \- The pay-as-you-go model of cloud services reduces maintenance costs usually associated with the on-premise choice; \- Greater flexibility for deploying and integrating new technologies and services; \- Cloud can be accessed from anywhere, at any time, turning team collaboration and remote work into a breeze; \- Built-in backup and disaster recovery capabilities, ensuring data safety and minimizing downtime in case of emergencies. Cloud-based data management cons: \- Cloud-first approach raises many data security, privacy, and compliance concerns; \- Transferring large data volumes to and from cloud is often time-consuming and results in increased latency for certain apps; \- Relying on a single cloud provider makes it difficult to switch them or move back to the on-premises option without significant funds and effort. Challenges that organizations that choose a cloud-first approach face: \- Integrating cloud-based systems with on-premises ones can be complex and time-consuming; \- Ensuring data governance and compliance in a multi-cloud or hybrid environment is also another problem reported by my clients. How EDW, ESB, and MDM promote cloud-first approach: A cloud-based EDW centralizes data from multiple sources, enabling a unified view of the organization’s data and simplifying data integration across cloud and on-premises systems. An ESB facilitates communication between disparate cloud and on-premises systems, streamlining data integration and promoting a modular architecture. Cloud-based MDM solutions are used for maintaining data quality and consistency across multiple data sources and environments. Principle 6: Automation and artificial intelligence Incorporating automation tools and AI technologies into data architecture can optimize processes and decision-making. Key Applications: \- Data ingestion and integration: Automation simplifies data schema updates and identifies data quality issues, while AI-assisted development helps create tailored connectors, scripts, and microservices. \- Data quality management: Machine learning algorithms improve data quality and consistency by automatically detecting and correcting inconsistencies and duplicates. \- Predictive analytics: AI and machine learning models analyze historical data to predict trends, identify opportunities, and uncover hidden patterns for better-informed decisions. How No-Code Tools and AI-Assisted Development Work: Business users define data requirements and workflows using no-code tools, enabling AI models to understand their needs. AI models process the information, generating recommendations for connector creation, ETL scripts, and microservices. Developers use AI-generated suggestions to accelerate development and tailor solutions to business needs. By combining automation, AI technologies, and no-code tools, organizations can streamline data architecture processes and bridge the gap between business users and developers, ultimately accelerating innovation. I share more tips on building an agile data architectures in my blog.

I Built Blainy - An AI Writing Tool for Students and Researchers
reddit
LLM Vibe Score0
Human Vibe Score1
silverglimmer1This week

I Built Blainy - An AI Writing Tool for Students and Researchers

Hello Everyone, I built Blainy, an AI writing tool designed to make writing easier and more efficient, based on my own experiences as a student working part-time and struggling to find the time for essays and assignments. Blainy is perfect for students, researchers, content creators, and bloggers. It addresses the gaps where most writing tools fall short and helps you write essays, assignments, research papers, product descriptions, blog content, and more with ease. I created this tool based on the problems I faced, so I genuinely want to know your review on this. Blainy's Features: AI Suggestions: This feature provides you with suggestions while you are writing, so you don't face the writer's block issue. This was the main issue I usually faced when writing my essays. You will get suggestions while you are writing, and if you don't like them, you can always ask for alternatives. AI Automation: If you want AI to write for you, you can choose this feature. It will write one to two paragraphs according to what you select. You can choose to write an introduction, conclusion, arguments, etc. If you just want it to write casually, select the "continue writing" feature, and it will write all on its own. Paraphrasing: If you want to paraphrase your text, you can do it on Blainy. You can also select different tones for writing, such as academic, friendly, simplicity, and more. Citations: By using this feature, you no longer need to search for citations on Google or ChatGPT. Blainy will load millions of citations for you in seconds. You can select any citation you want, and if you want to add a custom citation, you can do that too. Built-in Plagiarism Checker: Blainy includes a plagiarism checker to ensure that your content is original and plagiarism-free. PDF Chat: If you have any questions about a document that you are curious about or don't understand, you can use this feature. It will answer your question and help you summarize the whole article, and more. Best of all, We provide daily credits so you can access all these features for free with daily credits! We understand the unique challenges faced by students, including those with dyslexia and other writing difficulties. That's why we're working on adding features like a voice-to-text converter to assist students who struggle with writing. Your feedback is invaluable to us, so please don't hesitate to reach out and share your thoughts. We're also considering adding some free tools like paraphrasing to attract more users. If you have any suggestions for additional features that would be beneficial, please let me know. Your input can help us improve Blainy and make it even more valuable for everyone. If you have any good ideas that you think can help us in any way, please let me know. Thank you in advance for your support and feedback! Check it out: Blainy

My Marketing App made $10,000 in 2024. Here is how I target to make $100,000 in 2025:
reddit
LLM Vibe Score0
Human Vibe Score1
MonkDiThis week

My Marketing App made $10,000 in 2024. Here is how I target to make $100,000 in 2025:

You totally get me, I think. It’s a bizarre feeling when you build something, and people appreciate it and are even ready to pay! Pleasant though) In early 2024 my mate and I created a marketing tool that generates ads, content and strategy blocks with a click – Aiter.io. Users can just insert a URL, hit the button and everything is ready. TBH, I built this tool because I’m too lazy to chat with ChatGPT) https://preview.redd.it/ew2kud7ceyde1.png?width=1140&format=png&auto=webp&s=f3fe5b67075858cea3d52278e8063113efa3b97e In 2024 we made $10,000, here is what worked for us: AI directories. Still is the best channel of traffic and clients for us. We listed on TAAFT and other directories scrape TAAFT, so, eventually, we became listed on all major ones. I wrote a Reddit post earlier that explained this process in detail. Email marketing. Gosh, I thought it was dead – I have never been so wrong! We set up automatic emails that share marketing insights and they have a \~25% open rate + consistently convert people. It works great. Product marketing. Having a free version really helps with word-of-mouth and leads, which can be converted via email. Also, we consistently worked on product improvement. I’d say, that our free updates give people a feeling that the devs care about their stuff that’s why they are more confident investing in it. Google Ads. TBH, we had a shitty landing page all the time because were busy with the product. So, Google Ads didn’t work well for us. But we’ve launched the 2.0 version which has a better landing page, and will try it again. Influencers. Worked well for us, but we didn’t pay a dime for this. They just found our tool on directories and created videos about Aiter, so it was a sporadic marketing channel for us. We hope to change it in 2025. We see that our product works and attracts the audience, so we want to deliver and get more in 2025. Here is the plan: Product: add ad banners and video generation. So far, we generate only text data and it’s not so valuable in the time of ChatGPT and Claude. But to generate a high-quality ad banner is still challenging, so we put this on our roadmap. Another feature – one-click market analysis to get marketing insights. Become a TOP50 tool on TAAFT. We’ve become a top tool in our category (content generation) but will need to promote our profile on the profile far more aggressively to get into TOP50 Email marketing. We are fools because we almost didn’t have product emails that explain how it works. Will fix it. Also, we are considering participating more in paid newsletters, like collaborating with Substack influencers. Youtube marketing. Search for low-tail marketing keywords on YouTube and create videos on them, placing my product in them. Blog. Our new platform is Webflow which gives a lot of flexibility in terms of blogging. So, we will repeat the YouTube strategy with blogging. Paid marketing. With an updated landing page, we hope that paid campaigns will work better. We plan to launch campaigns that target different jobs to be done and customer objections to find the right message. Product Management. For 2025, our two key product metrics are retention and product activation rate. For this, we plan to simplify onboarding and make it simpler as well as conduct a lot of in-depth interviews to understand how we can retain users better. Funding. All of this exciting stuff requires money, so we are in the process of securing funding (fingers crossed). Having an indie project is exciting and invigorating. With all these activities, I hope we will achieve the goal of $100,000 in 2025. And what are your goals and marketing steps for 2025? Or maybe you could share some exciting marketing ideas I overlooked?

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I very rarely have stuff to post on Reddit, but I share how my project is going on, just random stuff, and memes on X. In case few might want to keep up 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2B products beats building B2C products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

Things I did to promote my product, and how they turned out
reddit
LLM Vibe Score0
Human Vibe Score1
laike9mThis week

Things I did to promote my product, and how they turned out

(I will share more updates in the future, you can find me on Twitter and/or Mastodon) Ask any ten indie developers about the toughest part of their job, and nine will likely say "marketing." I recently got a taste of this firsthand when I launched Xylect. Here's a rundown of my promotional attempts - hopefully, my experiences can help fellow developers out there. Podcast Community (✅ Success) I kicked things off by promoting Xylect in my podcast listener group. It wasn't a blockbuster, but I managed to sell a few copies and got some invaluable feedback from friends. Shoutout to those early supporters! Reddit r/macapps (✅ Success) Having had some luck promoting open-source projects on Reddit before, I decided to make r/macapps my first stop in the English-speaking world. I made an app to help you automate boring tasks with one click This post turned out to be a hit! I sold about ten copies and got a ton of useful feedback. Users pointed out compatibility issues with PopClip and suggested improvements for the website. One Italian user even requested localization, which I happily added. https://preview.redd.it/y4fuwh6hleqd1.png?width=959&format=png&auto=webp&s=7bb1b68cbf8a4f94998999e0832b9b7bd85bac67 https://preview.redd.it/8uu4cmyhleqd1.png?width=683&format=png&auto=webp&s=8f1744636aee8074b0e7491a334ef06076b143b0 I also got an intriguing email from a French user - more on that later. More Reddit Posts (❌ Failure) Riding high on my r/macapps success, I branched out to r/SideProject, r/Entrepreneur, and r/indiehackers. These subreddits frown upon direct self-promotion, so I took a softer approach with an article: The unexpected emotional cost of being an indiehacker While the article was heartfelt, it fell flat. Across all three posts, I got a grand total of three comments - two of which were complaints about the font size on mobile. Needless to say, I didn't sell a single copy. Hacker News (❌ Failure) As one of the tech world's major forums, I had to give Hacker News a shot. I wasn't too optimistic, given my past experiences there. Posting on HN feels like a mix of luck and dark magic. As expected, my post vanished without a trace - no comments, no sales. I might give it another go someday. If you're curious, you can check out my previous HN submissions. Tools Directory Websites (❌ Failure) These sites have a simple premise: you list your app, they display it. Seemed like an easy way to get some backlinks, right? Well, I learned the hard way that it's not that simple. I stumbled upon a Reddit post where someone claimed to have made a killing with their directory site in just a few days. The catch? Each listing cost $19. The site had a handful of apps listed, so I thought, "Why not? Early bird gets the worm." I paid up and listed Xylect. Spoiler alert: all I got was $19 poorer 🥲 Lesson learned: These directory sites won't magically sell your product. At best, they're just glorified backlinks. There might be some value in paid promotions on these platforms, but I can't speak to that from experience. V2EX (❌ Failure) After striking out in the English-speaking world, I turned my attention to the Chinese market, starting with V2EX (think of it as China's hybrid of HN and Reddit). This turned out to be my most unexpected flop. Here's the post: [\[Launch Discount\] Mac's most powerful AI search (Perplexity + Wikipedia + Google), boost your efficiency tenfold with one click. No API key required, no prompt needed, no token limit 🔥 - V2EX](https://www.v2ex.com/t/1064930?p=1#reply36) I'd seen decent engagement on other promo posts, so I had high hopes. I posted late at night (US time) and went to bed dreaming of waking up to a flood of comments. Reality check: The next morning, I had exactly one reply - from Kilerd, a loyal podcast listener showing some love. I was baffled. After re-reading my post, I realized I'd missed a crucial element: promo codes. A quick scan of popular posts confirmed my suspicion. Nearly every successful promo post was offering codes, and most comments were just base64-encoded email addresses. Talk about a facepalm moment. I scrambled to add a note about an upcoming free trial and invited users to drop their emails. This got the ball rolling with some code requests, but by then, the damage was done. The post fizzled out, and I didn't sell a single copy 🫠 A French Friend's Newsletter (✅ Success) At this point, my promotional efforts were looking pretty grim. My sales chart had a depressing stretch of flatline. But then, a glimmer of hope appeared in my inbox. Remember that French user I mentioned earlier? He ran a newsletter called vvmac and offered to feature Xylect if I added French support and sent him a free license. It was an offer I couldn't refuse. What followed was a crash course in French localization (thank you, Claude!) and the start of an incredible partnership. This guy was the most thorough beta tester I've ever encountered. We exchanged over sixty emails, covering everything from translations to UI tweaks to bug fixes. His response time was lightning-fast - I'd fix a bug, and five minutes later, he'd confirm it was sorted. The result? A much-improved Xylect and a glowing feature in his newsletter. https://preview.redd.it/ylcq2wxoleqd1.png?width=991&format=png&auto=webp&s=ee395110f50417d5c7f61318f27bf3dc30247809 I'm still in awe of his dedication. He single-handedly transformed Xylect from a buggy mess into a polished product. I'll be forever grateful for his help. The newsletter feature led to a few more sales, but honestly, that felt like a bonus at that point. Influencers (❌ Failure) I knew from the start that to really make waves, I'd need influencer backing. So, I added a note offering free licenses to content creators willing to collaborate. https://preview.redd.it/tyb2m1rqleqd1.png?width=799&format=png&auto=webp&s=56eabf126e772515322595613c546e6ba69fb431 I did get one taker: Hey, I'll be honest, I am not a huge content creator but I think I put a lot of effort in evaluating and figuring out which apps work... So I was wondering if I could get a license in case you are willing to share it. Thank you for considering. Have a great weekend. But I knew I needed to aim higher. With the new French localization, I thought I'd try my luck with some French-speaking Mac YouTubers. I crafted emails highlighting how Xylect could help their French audience with English content. https://preview.redd.it/07oqzemrleqd1.png?width=542&format=png&auto=webp&s=3d160c1d149f28e9029816a277c6ab2496fcd57e After days of silence, I got one reply. It was... not what I was hoping for: Hi, Thank you for your proposal. I can help you to promote your service on Tiktok, Instagram et YouTube, with unique short video. Price for this project is 3500€. Unless I've completely lost my marbles, there's no way I'm dropping 3500€ on promotion. Sure, given their follower count (YouTube: 348K, TikTok: 2.7M, Instagram: 400K), it's not an outrageous ask. For some products, it might even be worth it. But for Xylect? No way. I also reached out to a Chinese influencer on Xiaohongshu, but they weren't interested. Back to the drawing board. Conclusion If you've made it this far, you've probably realized this isn't exactly a success story. My search for effective promotional channels came up largely empty-handed. I'd naively thought that my success with open-source projects would translate seamlessly to the indie dev world. Boy, was I wrong. As I mentioned in my previous article, open-source projects create a dynamic where users feel indebted to developers for their free labor. But in the commercial world of indie development, that dynamic completely flips. While this experience was often frustrating, it was also enlightening - which was kind of the point. As my first foray into indie development, my main goal was to learn the ropes and understand the process. Making money would've been nice, sure, but it wasn't my primary focus. Thanks for sticking with me through this post. I will share more updates in the future, you can follow me on  Twitter and/or Mastodon.

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I very rarely have stuff to post on Reddit, but I share how my project is going on, just random stuff, and memes on X. In case few might want to keep up 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2B products beats building B2C products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

I’ve built a gaming recommendation and exploration platform called Which Game Next
reddit
LLM Vibe Score0
Human Vibe Score0.714
kasperooThis week

I’ve built a gaming recommendation and exploration platform called Which Game Next

Hello there! Me and a few of my best friends are software engineers, and we’ve been working part-time on developing a side project for the past 12 months. It’s called www.whichgamenext.com, and we’ve recently launched into open beta for everyone to check out. Your feedback would be invaluable to us! Our aim has been to build a gaming recommendation engine, alongside providing market oversight for where you can legally and officially purchase or obtain modern games from multiple stores and/or subscriptions. It’s often difficult to figure out what you have access to if you only have a single specific subscription, like Game Pass PC, or if you’re only interested in games on GOG/Nintendo (what a mix!). We started by identifying the available digital stores and subscriptions and slowly compiling our database using multiple automated services to gather data on these games. Think JustWatch, but for games! One major service we’ve partnered with is IGDB, which has been supplying us with JSON data dumps that served as the initial seed for our game data. A massive thank you to them for their continued support! With the data in place, we’ve been focusing on exploring new features. So far, this has included private and public user-generated lists, personal backlog tracking, and the ability to like or dislike games. We’re now improving our recommendation engine, tackling the complexities that come with it, and having a lot of fun along the way. We’re utilising modern AI strategies and solving fascinating problems related to large-scale data aggregation. We truly can’t wait to share this fantastic work! In addition to this, you can soon expect curated collections, articles about games, and supporting links to help you make informed, unbiased purchasing decisions. Your shared data will drive the recommendations. But it doesn’t stop there—we have plenty of other features on our radar, such as importing games from your favourite stores, syncing your gameplay time, surfacing data like “How Long to Beat,” and creating new and exciting ways to interact with this growing community! This is a passion project created by a group of gamers who want to spend their time and money wisely, without purchasing biases. Since it’s a side project, we mostly work on it at night, but we’re excited to grow the community, share our vision, and, who knows, maybe one day make it our full-time job! Let’s dive into the technical details: • Monorepo architecture: This speeds up development by sharing libraries, living style guides, configs, etc. Nx.js has been brilliant, enabling us to create a dependency graph of changes and only build/deploy what’s modified in a PR. • AWS: We’re using the free tier (with a few exceptions where we pay for smaller services). Achieving self-sufficiency is critical for us. Additionally, we applied to the AWS Startup Foundation programme and received $1,000 in AWS credits, which has been incredibly helpful! • Infrastructure: Fully deployed as code with Terraform. • Backends: Built using Express and Nest.js, split into around 40 projects and counting! Each project plays a unique role in gathering and syncing game data. • Scalability: Designed from the ground up, utilising AWS Lambdas with auto-scaling and load balancing. • Databases: We use Postgres with RDS and DynamoDB for storing various data. • Frontend stack: Built with React, Next.js, Tailwind, Zustand, TanStack Query, Jest, and Storybook. • CI/CD: Managed with GitHub Actions and Amplify hooks for deploying the frontends. • Admin portal: We’ve built a bespoke CMS to control the main website. It synchronises with external services, tracks game data changes, and allows us to selectively apply ‘patches’ from sites like IGDB. The system also includes data override and rollback capabilities, ensuring we maintain control over game data. • Automation: Partially automated, so manual intervention is rarely needed. • Scraping tools: Fully integrated into the admin portal with log trail capabilities. • Cloudflare: Used for on-the-fly image transformations; we’re considering moving to it full-time as our CDN for free WebP conversions. • Authentication: Handled by Cognito, with a custom frontend built from scratch. Key learnings so far: • AWS cold starts: Not ideal! While the platform is still new, we ping endpoints to keep them responsive. This won’t be an issue once traffic increases. • Lambda memory matters: We learned the hard way that low-memory configurations can delay responses by 2-3 seconds. • DynamoDB partition keys: If not designed correctly from the start, you might have to start over (yes, we’ve been there!). • GitHub Actions: Setting up node\_modules cache reuse takes time, but it’s worth it—don’t give up! We don’t know where this project will take us yet, but it’s been a fantastic journey so far. We’ve learned a lot, explored technologies we don’t typically use in our day jobs, and built something we’re genuinely passionate about. Your feedback would mean the world to us. What do you think of what we’ve done so far? What would you like to see added? Is this a service you’d use? Do you see the value in it as we do? Thanks for reading, and we hope to see you in the comments! (or our newly created /r/whichgamenext

I Built Blainy - An AI Writing Tool for Students and Researchers
reddit
LLM Vibe Score0
Human Vibe Score1
silverglimmer1This week

I Built Blainy - An AI Writing Tool for Students and Researchers

Hello Everyone, I built Blainy, an AI writing tool designed to make writing easier and more efficient, based on my own experiences as a student working part-time and struggling to find the time for essays and assignments. Blainy is perfect for students, researchers, content creators, and bloggers. It addresses the gaps where most writing tools fall short and helps you write essays, assignments, research papers, product descriptions, blog content, and more with ease. I created this tool based on the problems I faced, so I genuinely want to know your review on this. Blainy's Features: AI Suggestions: This feature provides you with suggestions while you are writing, so you don't face the writer's block issue. This was the main issue I usually faced when writing my essays. You will get suggestions while you are writing, and if you don't like them, you can always ask for alternatives. AI Automation: If you want AI to write for you, you can choose this feature. It will write one to two paragraphs according to what you select. You can choose to write an introduction, conclusion, arguments, etc. If you just want it to write casually, select the "continue writing" feature, and it will write all on its own. Paraphrasing: If you want to paraphrase your text, you can do it on Blainy. You can also select different tones for writing, such as academic, friendly, simplicity, and more. Citations: By using this feature, you no longer need to search for citations on Google or ChatGPT. Blainy will load millions of citations for you in seconds. You can select any citation you want, and if you want to add a custom citation, you can do that too. Built-in Plagiarism Checker: Blainy includes a plagiarism checker to ensure that your content is original and plagiarism-free. PDF Chat: If you have any questions about a document that you are curious about or don't understand, you can use this feature. It will answer your question and help you summarize the whole article, and more. Best of all, We provide daily credits so you can access all these features for free with daily credits! We understand the unique challenges faced by students, including those with dyslexia and other writing difficulties. That's why we're working on adding features like a voice-to-text converter to assist students who struggle with writing. Your feedback is invaluable to us, so please don't hesitate to reach out and share your thoughts. We're also considering adding some free tools like paraphrasing to attract more users. If you have any suggestions for additional features that would be beneficial, please let me know. Your input can help us improve Blainy and make it even more valuable for everyone. If you have any good ideas that you think can help us in any way, please let me know. Thank you in advance for your support and feedback! Check it out: Blainy

[D] Overwhelmed by fast advances in recent weeks
reddit
LLM Vibe Score0
Human Vibe Score1
iamx9000againThis week

[D] Overwhelmed by fast advances in recent weeks

I was watching the GTC keynote and became entirely overwhelmed by the amount of progress achieved from last year. I'm wondering how everyone else feels. &#x200B; Firstly, the entire ChatGPT, GPT-3/GPT-4 chaos has been going on for a few weeks, with everyone scrambling left and right to integrate chatbots into their apps, products, websites. Twitter is flooded with new product ideas, how to speed up the process from idea to product, countless promp engineering blogs, tips, tricks, paid courses. &#x200B; Not only was ChatGPT disruptive, but a few days later, Microsoft and Google also released their models and integrated them into their search engines. Microsoft also integrated its LLM into its Office suite. It all happenned overnight. I understand that they've started integrating them along the way, but still, it seems like it hapenned way too fast. This tweet encompases the past few weeks perfectly https://twitter.com/AlphaSignalAI/status/1638235815137386508 , on a random Tuesday countless products are released that seem revolutionary. &#x200B; In addition to the language models, there are also the generative art models that have been slowly rising in mainstream recognition. Now Midjourney AI is known by a lot of people who are not even remotely connected to the AI space. &#x200B; For the past few weeks, reading Twitter, I've felt completely overwhelmed, as if the entire AI space is moving beyond at lightning speed, whilst around me we're just slowly training models, adding some data, and not seeing much improvement, being stuck on coming up with "new ideas, that set us apart". &#x200B; Watching the GTC keynote from NVIDIA I was again, completely overwhelmed by how much is being developed throughout all the different domains. The ASML EUV (microchip making system) was incredible, I have no idea how it does lithography and to me it still seems like magic. The Grace CPU with 2 dies (although I think Apple was the first to do it?) and 100 GB RAM, all in a small form factor. There were a lot more different hardware servers that I just blanked out at some point. The omniverse sim engine looks incredible, almost real life (I wonder how much of a domain shift there is between real and sim considering how real the sim looks). Beyond it being cool and usable to train on synthetic data, the car manufacturers use it to optimize their pipelines. This change in perspective, of using these tools for other goals than those they were designed for I find the most interesting. &#x200B; The hardware part may be old news, as I don't really follow it, however the software part is just as incredible. NVIDIA AI foundations (language, image, biology models), just packaging everything together like a sandwich. Getty, Shutterstock and Adobe will use the generative models to create images. Again, already these huge juggernauts are already integrated. &#x200B; I can't believe the point where we're at. We can use AI to write code, create art, create audiobooks using Britney Spear's voice, create an interactive chatbot to converse with books, create 3D real-time avatars, generate new proteins (?i'm lost on this one), create an anime and countless other scenarios. Sure, they're not perfect, but the fact that we can do all that in the first place is amazing. &#x200B; As Huang said in his keynote, companies want to develop "disruptive products and business models". I feel like this is what I've seen lately. Everyone wants to be the one that does something first, just throwing anything and everything at the wall and seeing what sticks. &#x200B; In conclusion, I'm feeling like the world is moving so fast around me whilst I'm standing still. I want to not read anything anymore and just wait until everything dies down abit, just so I can get my bearings. However, I think this is unfeasible. I fear we'll keep going in a frenzy until we just burn ourselves at some point. &#x200B; How are you all fairing? How do you feel about this frenzy in the AI space? What are you the most excited about?

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[P]MMML | Deploy HuggingFace training model rapidly based on MetaSpore
reddit
LLM Vibe Score0
Human Vibe Score1
qazmkoppThis week

[P]MMML | Deploy HuggingFace training model rapidly based on MetaSpore

A few days ago, HuggingFace announced a $100 million Series C funding round, which was big news in open source machine learning and could be a sign of where the industry is headed. Two days before the HuggingFace funding announcement, open-source machine learning platform MetaSpore released a demo based on the HuggingFace Rapid deployment pre-training model. As deep learning technology makes innovative breakthroughs in computer vision, natural language processing, speech understanding, and other fields, more and more unstructured data are perceived, understood, and processed by machines. These advances are mainly due to the powerful learning ability of deep learning. Through pre-training of deep models on massive data, the models can capture the internal data patterns, thus helping many downstream tasks. With the industry and academia investing more and more energy in the research of pre-training technology, the distribution warehouses of pre-training models such as HuggingFace and Timm have emerged one after another. The open-source community release pre-training significant model dividends at an unprecedented speed. In recent years, the data form of machine modeling and understanding has gradually evolved from single-mode to multi-mode, and the semantic gap between different modes is being eliminated, making it possible to retrieve data across modes. Take CLIP, OpenAI’s open-source work, as an example, to pre-train the twin towers of images and texts on a dataset of 400 million pictures and texts and connect the semantics between pictures and texts. Many researchers in the academic world have been solving multimodal problems such as image generation and retrieval based on this technology. Although the frontier technology through the semantic gap between modal data, there is still a heavy and complicated model tuning, offline data processing, high performance online reasoning architecture design, heterogeneous computing, and online algorithm be born multiple processes and challenges, hindering the frontier multimodal retrieval technologies fall to the ground and pratt &whitney. DMetaSoul aims at the above technical pain points, abstracting and uniting many links such as model training optimization, online reasoning, and algorithm experiment, forming a set of solutions that can quickly apply offline pre-training model to online. This paper will introduce how to use the HuggingFace community pre-training model to conduct online reasoning and algorithm experiments based on MetaSpore technology ecology so that the benefits of the pre-training model can be fully released to the specific business or industry and small and medium-sized enterprises. And we will give the text search text and text search graph two multimodal retrieval demonstration examples for your reference. Multimodal semantic retrieval The sample architecture of multimodal retrieval is as follows: Our multimodal retrieval system supports both text search and text search application scenarios, including offline processing, model reasoning, online services, and other core modules: &#x200B; https://preview.redd.it/w4v4c7vcez291.png?width=1834&format=png&auto=webp&s=0687efb1fddb26e8e30cb844d398ec712b947f31 Offline processing, including offline data processing processes for different application scenarios of text search and text search, including model tuning, model export, data index database construction, data push, etc. Model inference. After the offline model training, we deployed our NLP and CV large models based on the MetaSpore Serving framework. MetaSpore Serving helps us conveniently perform online inference, elastic scheduling, load balancing, and resource scheduling in heterogeneous environments. Online services. Based on MetaSpore’s online algorithm application framework, MetaSpore has a complete set of reusable online search services, including Front-end retrieval UI, multimodal data preprocessing, vector recall and sorting algorithm, AB experimental framework, etc. MetaSpore also supports text search by text and image scene search by text and can be migrated to other application scenarios at a low cost. The HuggingFace open source community has provided several excellent baseline models for similar multimodal retrieval problems, which are often the starting point for actual optimization in the industry. MetaSpore also uses the pre-training model of the HuggingFace community in its online services of searching words by words and images by words. Searching words by words is based on the semantic similarity model of the question and answer field optimized by MetaSpore, and searching images by words is based on the community pre-training model. These community open source pre-training models are exported to the general ONNX format and loaded into MetaSpore Serving for online reasoning. The following sections will provide a detailed description of the model export and online retrieval algorithm services. The reasoning part of the model is standardized SAAS services with low coupling with the business. Interested readers can refer to my previous post: The design concept of MetaSpore, a new generation of the one-stop machine learning platform. 1.1 Offline Processing Offline processing mainly involves the export and loading of online models and index building and pushing of the document library. You can follow the step-by-step instructions below to complete the offline processing of text search and image search and see how the offline pre-training model achieves reasoning at MetaSpore. 1.1.1 Search text by text Traditional text retrieval systems are based on literal matching algorithms such as BM25. Due to users’ diverse query words, a semantic gap between query words and documents is often encountered. For example, users misspell “iPhone” as “Phone,” and search terms are incredibly long, such as “1 \~ 3 months old baby autumn small size bag pants”. Traditional text retrieval systems will use spelling correction, synonym expansion, search terms rewriting, and other means to alleviate the semantic gap but fundamentally fail to solve this problem. Only when the retrieval system fully understands users’ query terms and documents can it meet users’ retrieval demands at the semantic level. With the continuous progress of pre-training and representational learning technology, some commercial search engines continue to integrate semantic vector retrieval methods based on symbolic learning into the retrieval ecology. Semantic retrieval model This paper introduces a set of semantic vector retrieval applications. MetaSpore built a set of semantic retrieval systems based on encyclopedia question and answer data. MetaSpore adopted the Sentence-Bert model as the semantic vector representation model, which fine-tunes the twin tower BERT in supervised or unsupervised ways to make the model more suitable for retrieval tasks. The model structure is as follows: The query-Doc symmetric two-tower model is used in text search and question and answer retrieval. The vector representation of online Query and offline DOC share the same vector representation model, so it is necessary to ensure the consistency of the offline DOC library building model and online Query inference model. The case uses MetaSpore’s text representation model Sbert-Chinese-QMC-domain-V1, optimized in the open-source semantically similar data set. This model will express the question and answer data as a vector in offline database construction. The user query will be expressed as a vector by this model in online retrieval, ensuring that query-doc in the same semantic space, users’ semantic retrieval demands can be guaranteed by vector similarity metric calculation. Since the text presentation model does vector encoding for Query online, we need to export the model for use by the online service. Go to the q&A data library code directory and export the model concerning the documentation. In the script, Pytorch Tracing is used to export the model. The models are exported to the “./export “directory. The exported models are mainly ONNX models used for wired reasoning, Tokenizer, and related configuration files. The exported models are loaded into MetaSpore Serving by the online Serving system described below for model reasoning. Since the exported model will be copied to the cloud storage, you need to configure related variables in env.sh. \Build library based on text search \ The retrieval database is built on the million-level encyclopedia question and answer data set. According to the description document, you need to download the data and complete the database construction. The question and answer data will be coded as a vector by the offline model, and then the database construction data will be pushed to the service component. The whole process of database construction is described as follows: Preprocessing, converting the original data into a more general JSonline format for database construction; Build index, use the same model as online “sbert-Chinese-qmc-domain-v1” to index documents (one document object per line); Push inverted (vector) and forward (document field) data to each component server. The following is an example of the database data format. After offline database construction is completed, various data are pushed to corresponding service components, such as Milvus storing vector representation of documents and MongoDB storing summary information of documents. Online retrieval algorithm services will use these service components to obtain relevant data. 1.1.2 Search by text Text and images are easy for humans to relate semantically but difficult for machines. First of all, from the perspective of data form, the text is the discrete ID type of one-dimensional data based on words and words. At the same time, images are continuous two-dimensional or three-dimensional data. Secondly, the text is a subjective creation of human beings, and its expressive ability is vibrant, including various turning points, metaphors, and other expressions, while images are machine representations of the objective world. In short, bridging the semantic gap between text and image data is much more complex than searching text by text. The traditional text search image retrieval technology generally relies on the external text description data of the image or the nearest neighbor retrieval technology and carries out the retrieval through the image associated text, which in essence degrades the problem to text search. However, it will also face many issues, such as obtaining the associated text of pictures and whether the accuracy of text search by text is high enough. The depth model has gradually evolved from single-mode to multi-mode in recent years. Taking the open-source project of OpenAI, CLIP, as an example, train the model through the massive image and text data of the Internet and map the text and image data into the same semantic space, making it possible to implement the text and image search technology based on semantic vector. CLIP graphic model The text search pictures introduced in this paper are implemented based on semantic vector retrieval, and the CLIP pre-training model is used as the two-tower retrieval architecture. Because the CLIP model has trained the semantic alignment of the twin towers’ text and image side models on the massive graphic and text data, it is particularly suitable for the text search graph scene. Due to the different image and text data forms, the Query-Doc asymmetric twin towers model is used for text search image retrieval. The image-side model of the twin towers is used for offline database construction, and the text-side model is used for the online return. In the final online retrieval, the database data of the image side model will be searched after the text side model encodes Query, and the CLIP pre-training model guarantees the semantic correlation between images and texts. The model can draw the graphic pairs closer in vector space by pre-training on a large amount of visual data. Here we need to export the text-side model for online MetaSpore Serving inference. Since the retrieval scene is based on Chinese, the CLIP model supporting Chinese understanding is selected. The exported content includes the ONNX model used for online reasoning and Tokenizer, similar to the text search. MetaSpore Serving can load model reasoning through the exported content. Build library on Image search You need to download the Unsplash Lite library data and complete the construction according to the instructions. The whole process of database construction is described as follows: Preprocessing, specify the image directory, and then generate a more general JSOnline file for library construction; Build index, use OpenAI/Clip-Vit-BASE-Patch32 pre-training model to index the gallery, and output one document object for each line of index data; Push inverted (vector) and forward (document field) data to each component server. Like text search, after offline database construction, relevant data will be pushed to service components, called by online retrieval algorithm services to obtain relevant data. 1.2 Online Services The overall online service architecture diagram is as follows: https://preview.redd.it/jfsl8hdfez291.png?width=1280&format=png&auto=webp&s=a858e2304a0c93e78ba5429612ca08cbee69b35a Multi-mode search online service system supports application scenarios such as text search and text search. The whole online service consists of the following parts: Query preprocessing service: encapsulate preprocessing logic (including text/image, etc.) of pre-training model, and provide services through gRPC interface; Retrieval algorithm service: the whole algorithm processing link includes AB experiment tangent flow configuration, MetaSpore Serving call, vector recall, sorting, document summary, etc.; User entry service: provides a Web UI interface for users to debug and track down problems in the retrieval service. From a user request perspective, these services form invocation dependencies from back to front, so to build up a multimodal sample, you need to run each service from front to back first. Before doing this, remember to export the offline model, put it online and build the library first. This article will introduce the various parts of the online service system and make the whole service system step by step according to the following guidance. See the ReadME at the end of this article for more details. 1.2.1 Query preprocessing service Deep learning models tend to be based on tensors, but NLP/CV models often have a preprocessing part that translates raw text and images into tensors that deep learning models can accept. For example, NLP class models often have a pre-tokenizer to transform text data of string type into discrete tensor data. CV class models also have similar processing logic to complete the cropping, scaling, transformation, and other processing of input images through preprocessing. On the one hand, considering that this part of preprocessing logic is decoupled from tensor reasoning of the depth model, on the other hand, the reason of the depth model has an independent technical system based on ONNX, so MetaSpore disassembled this part of preprocessing logic. NLP pretreatment Tokenizer has been integrated into the Query pretreatment service. MetaSpore dismantlement with a relatively general convention. Users only need to provide preprocessing logic files to realize the loading and prediction interface and export the necessary data and configuration files loaded into the preprocessing service. Subsequent CV preprocessing logic will also be integrated in this manner. The preprocessing service currently provides the gRPC interface invocation externally and is dependent on the Query preprocessing (QP) module in the retrieval algorithm service. After the user request reaches the retrieval algorithm service, it will be forwarded to the service to complete the data preprocessing and continue the subsequent processing. The ReadMe provides details on how the preprocessing service is started, how the preprocessing model exported offline to cloud storage enters the service, and how to debug the service. To further improve the efficiency and stability of model reasoning, MetaSpore Serving implements a Python preprocessing submodule. So MetaSpore can provide gRPC services through user-specified preprocessor.py, complete Tokenizer or CV-related preprocessing in NLP, and translate requests into a Tensor that deep models can handle. Finally, the model inference is carried out by MetaSpore, Serving subsequent sub-modules. Presented here on the lot code: https://github.com/meta-soul/MetaSpore/compare/add\python\preprocessor 1.2.2 Retrieval algorithm services Retrieval algorithm service is the core of the whole online service system, which is responsible for the triage of experiments, the assembly of algorithm chains such as preprocessing, recall, sorting, and the invocation of dependent component services. The whole retrieval algorithm service is developed based on the Java Spring framework and supports multi-mode retrieval scenarios of text search and text search graph. Due to good internal abstraction and modular design, it has high flexibility and can be migrated to similar application scenarios at a low cost. Here’s a quick guide to configuring the environment to set up the retrieval algorithm service. See ReadME for more details: Install dependent components. Use Maven to install the online-Serving component Search for service configurations. Copy the template configuration file and replace the MongoDB, Milvus, and other configurations based on the development/production environment. Install and configure Consul. Consul allows you to synchronize the search service configuration in real-time, including cutting the flow of experiments, recall parameters, and sorting parameters. The project’s configuration file shows the current configuration parameters of text search and text search. The parameter modelName in the stage of pretreatment and recall is the corresponding model exported in offline processing. Start the service. Once the above configuration is complete, the retrieval service can be started from the entry script. Once the service is started, you can test it! For example, for a user with userId=10 who wants to query “How to renew ID card,” access the text search service. 1.2.3 User Entry Service Considering that the retrieval algorithm service is in the form of the API interface, it is difficult to locate and trace the problem, especially for the text search image scene can intuitively display the retrieval results to facilitate the iterative optimization of the retrieval algorithm. This paper provides a lightweight Web UI interface for text search and image search, a search input box, and results in a display page for users. Developed by Flask, the service can be easily integrated with other retrieval applications. The service calls the retrieval algorithm service and displays the returned results on the page. It’s also easy to install and start the service. Once you’re done, go to http://127.0.0.1:8090 to see if the search UI service is working correctly. See the ReadME at the end of this article for details. Multimodal system demonstration The multimodal retrieval service can be started when offline processing and online service environment configuration have been completed following the above instructions. Examples of textual searches are shown below. Enter the entry of the text search map application, enter “cat” first, and you can see that the first three digits of the returned result are cats: https://preview.redd.it/0n5nuyvhez291.png?width=1280&format=png&auto=webp&s=1e9c054f541d53381674b8d6001b4bf524506bd2 If you add a color constraint to “cat” to retrieve “black cat,” you can see that it does return a black cat: https://preview.redd.it/rzc0qjyjez291.png?width=1280&format=png&auto=webp&s=d5bcc503ef0fb3360c7740e60e295cf372dcad47 Further, strengthen the constraint on the search term, change it to “black cat on the bed,” and return results containing pictures of a black cat climbing on the bed: &#x200B; https://preview.redd.it/c4b2q8olez291.png?width=1280&format=png&auto=webp&s=4f3817b0b9f07e1e68d1d4a8281702ba3834a00a The cat can still be found through the text search system after the color and scene modification in the above example. Conclusion The cutting-edge pre-training technology can bridge the semantic gap between different modes, and the HuggingFace community can greatly reduce the cost for developers to use the pre-training model. Combined with the technological ecology of MetaSpore online reasoning and online microservices provided by DMetaSpore, the pre-training model is no longer mere offline dabbling. Instead, it can truly achieve end-to-end implementation from cutting-edge technology to industrial scenarios, fully releasing the dividends of the pre-training large model. In the future, DMetaSoul will continue to improve and optimize the MetaSpore technology ecosystem: More automated and wider access to HuggingFace community ecology. MetaSpore will soon release a common model rollout mechanism to make HuggingFace ecologically accessible and will later integrate preprocessing services into online services. Multi-mode retrieval offline algorithm optimization. For multimodal retrieval scenarios, MetaSpore will continuously iteratively optimize offline algorithm components, including text recall/sort model, graphic recall/sort model, etc., to improve the accuracy and efficiency of the retrieval algorithm. For related code and reference documentation in this article, please visit: https://github.com/meta-soul/MetaSpore/tree/main/demo/multimodal/online Some images source: https://github.com/openai/CLIP/raw/main/CLIP.png https://www.sbert.net/examples/training/sts/README.html

[P]MMML | Deploy HuggingFace training model rapidly based on MetaSpore
reddit
LLM Vibe Score0
Human Vibe Score1
qazmkoppThis week

[P]MMML | Deploy HuggingFace training model rapidly based on MetaSpore

A few days ago, HuggingFace announced a $100 million Series C funding round, which was big news in open source machine learning and could be a sign of where the industry is headed. Two days before the HuggingFace funding announcement, open-source machine learning platform MetaSpore released a demo based on the HuggingFace Rapid deployment pre-training model. As deep learning technology makes innovative breakthroughs in computer vision, natural language processing, speech understanding, and other fields, more and more unstructured data are perceived, understood, and processed by machines. These advances are mainly due to the powerful learning ability of deep learning. Through pre-training of deep models on massive data, the models can capture the internal data patterns, thus helping many downstream tasks. With the industry and academia investing more and more energy in the research of pre-training technology, the distribution warehouses of pre-training models such as HuggingFace and Timm have emerged one after another. The open-source community release pre-training significant model dividends at an unprecedented speed. In recent years, the data form of machine modeling and understanding has gradually evolved from single-mode to multi-mode, and the semantic gap between different modes is being eliminated, making it possible to retrieve data across modes. Take CLIP, OpenAI’s open-source work, as an example, to pre-train the twin towers of images and texts on a dataset of 400 million pictures and texts and connect the semantics between pictures and texts. Many researchers in the academic world have been solving multimodal problems such as image generation and retrieval based on this technology. Although the frontier technology through the semantic gap between modal data, there is still a heavy and complicated model tuning, offline data processing, high performance online reasoning architecture design, heterogeneous computing, and online algorithm be born multiple processes and challenges, hindering the frontier multimodal retrieval technologies fall to the ground and pratt &whitney. DMetaSoul aims at the above technical pain points, abstracting and uniting many links such as model training optimization, online reasoning, and algorithm experiment, forming a set of solutions that can quickly apply offline pre-training model to online. This paper will introduce how to use the HuggingFace community pre-training model to conduct online reasoning and algorithm experiments based on MetaSpore technology ecology so that the benefits of the pre-training model can be fully released to the specific business or industry and small and medium-sized enterprises. And we will give the text search text and text search graph two multimodal retrieval demonstration examples for your reference. Multimodal semantic retrieval The sample architecture of multimodal retrieval is as follows: Our multimodal retrieval system supports both text search and text search application scenarios, including offline processing, model reasoning, online services, and other core modules: &#x200B; https://preview.redd.it/w4v4c7vcez291.png?width=1834&format=png&auto=webp&s=0687efb1fddb26e8e30cb844d398ec712b947f31 Offline processing, including offline data processing processes for different application scenarios of text search and text search, including model tuning, model export, data index database construction, data push, etc. Model inference. After the offline model training, we deployed our NLP and CV large models based on the MetaSpore Serving framework. MetaSpore Serving helps us conveniently perform online inference, elastic scheduling, load balancing, and resource scheduling in heterogeneous environments. Online services. Based on MetaSpore’s online algorithm application framework, MetaSpore has a complete set of reusable online search services, including Front-end retrieval UI, multimodal data preprocessing, vector recall and sorting algorithm, AB experimental framework, etc. MetaSpore also supports text search by text and image scene search by text and can be migrated to other application scenarios at a low cost. The HuggingFace open source community has provided several excellent baseline models for similar multimodal retrieval problems, which are often the starting point for actual optimization in the industry. MetaSpore also uses the pre-training model of the HuggingFace community in its online services of searching words by words and images by words. Searching words by words is based on the semantic similarity model of the question and answer field optimized by MetaSpore, and searching images by words is based on the community pre-training model. These community open source pre-training models are exported to the general ONNX format and loaded into MetaSpore Serving for online reasoning. The following sections will provide a detailed description of the model export and online retrieval algorithm services. The reasoning part of the model is standardized SAAS services with low coupling with the business. Interested readers can refer to my previous post: The design concept of MetaSpore, a new generation of the one-stop machine learning platform. 1.1 Offline Processing Offline processing mainly involves the export and loading of online models and index building and pushing of the document library. You can follow the step-by-step instructions below to complete the offline processing of text search and image search and see how the offline pre-training model achieves reasoning at MetaSpore. 1.1.1 Search text by text Traditional text retrieval systems are based on literal matching algorithms such as BM25. Due to users’ diverse query words, a semantic gap between query words and documents is often encountered. For example, users misspell “iPhone” as “Phone,” and search terms are incredibly long, such as “1 \~ 3 months old baby autumn small size bag pants”. Traditional text retrieval systems will use spelling correction, synonym expansion, search terms rewriting, and other means to alleviate the semantic gap but fundamentally fail to solve this problem. Only when the retrieval system fully understands users’ query terms and documents can it meet users’ retrieval demands at the semantic level. With the continuous progress of pre-training and representational learning technology, some commercial search engines continue to integrate semantic vector retrieval methods based on symbolic learning into the retrieval ecology. Semantic retrieval model This paper introduces a set of semantic vector retrieval applications. MetaSpore built a set of semantic retrieval systems based on encyclopedia question and answer data. MetaSpore adopted the Sentence-Bert model as the semantic vector representation model, which fine-tunes the twin tower BERT in supervised or unsupervised ways to make the model more suitable for retrieval tasks. The model structure is as follows: The query-Doc symmetric two-tower model is used in text search and question and answer retrieval. The vector representation of online Query and offline DOC share the same vector representation model, so it is necessary to ensure the consistency of the offline DOC library building model and online Query inference model. The case uses MetaSpore’s text representation model Sbert-Chinese-QMC-domain-V1, optimized in the open-source semantically similar data set. This model will express the question and answer data as a vector in offline database construction. The user query will be expressed as a vector by this model in online retrieval, ensuring that query-doc in the same semantic space, users’ semantic retrieval demands can be guaranteed by vector similarity metric calculation. Since the text presentation model does vector encoding for Query online, we need to export the model for use by the online service. Go to the q&A data library code directory and export the model concerning the documentation. In the script, Pytorch Tracing is used to export the model. The models are exported to the “./export “directory. The exported models are mainly ONNX models used for wired reasoning, Tokenizer, and related configuration files. The exported models are loaded into MetaSpore Serving by the online Serving system described below for model reasoning. Since the exported model will be copied to the cloud storage, you need to configure related variables in env.sh. \Build library based on text search \ The retrieval database is built on the million-level encyclopedia question and answer data set. According to the description document, you need to download the data and complete the database construction. The question and answer data will be coded as a vector by the offline model, and then the database construction data will be pushed to the service component. The whole process of database construction is described as follows: Preprocessing, converting the original data into a more general JSonline format for database construction; Build index, use the same model as online “sbert-Chinese-qmc-domain-v1” to index documents (one document object per line); Push inverted (vector) and forward (document field) data to each component server. The following is an example of the database data format. After offline database construction is completed, various data are pushed to corresponding service components, such as Milvus storing vector representation of documents and MongoDB storing summary information of documents. Online retrieval algorithm services will use these service components to obtain relevant data. 1.1.2 Search by text Text and images are easy for humans to relate semantically but difficult for machines. First of all, from the perspective of data form, the text is the discrete ID type of one-dimensional data based on words and words. At the same time, images are continuous two-dimensional or three-dimensional data. Secondly, the text is a subjective creation of human beings, and its expressive ability is vibrant, including various turning points, metaphors, and other expressions, while images are machine representations of the objective world. In short, bridging the semantic gap between text and image data is much more complex than searching text by text. The traditional text search image retrieval technology generally relies on the external text description data of the image or the nearest neighbor retrieval technology and carries out the retrieval through the image associated text, which in essence degrades the problem to text search. However, it will also face many issues, such as obtaining the associated text of pictures and whether the accuracy of text search by text is high enough. The depth model has gradually evolved from single-mode to multi-mode in recent years. Taking the open-source project of OpenAI, CLIP, as an example, train the model through the massive image and text data of the Internet and map the text and image data into the same semantic space, making it possible to implement the text and image search technology based on semantic vector. CLIP graphic model The text search pictures introduced in this paper are implemented based on semantic vector retrieval, and the CLIP pre-training model is used as the two-tower retrieval architecture. Because the CLIP model has trained the semantic alignment of the twin towers’ text and image side models on the massive graphic and text data, it is particularly suitable for the text search graph scene. Due to the different image and text data forms, the Query-Doc asymmetric twin towers model is used for text search image retrieval. The image-side model of the twin towers is used for offline database construction, and the text-side model is used for the online return. In the final online retrieval, the database data of the image side model will be searched after the text side model encodes Query, and the CLIP pre-training model guarantees the semantic correlation between images and texts. The model can draw the graphic pairs closer in vector space by pre-training on a large amount of visual data. Here we need to export the text-side model for online MetaSpore Serving inference. Since the retrieval scene is based on Chinese, the CLIP model supporting Chinese understanding is selected. The exported content includes the ONNX model used for online reasoning and Tokenizer, similar to the text search. MetaSpore Serving can load model reasoning through the exported content. Build library on Image search You need to download the Unsplash Lite library data and complete the construction according to the instructions. The whole process of database construction is described as follows: Preprocessing, specify the image directory, and then generate a more general JSOnline file for library construction; Build index, use OpenAI/Clip-Vit-BASE-Patch32 pre-training model to index the gallery, and output one document object for each line of index data; Push inverted (vector) and forward (document field) data to each component server. Like text search, after offline database construction, relevant data will be pushed to service components, called by online retrieval algorithm services to obtain relevant data. 1.2 Online Services The overall online service architecture diagram is as follows: https://preview.redd.it/jfsl8hdfez291.png?width=1280&format=png&auto=webp&s=a858e2304a0c93e78ba5429612ca08cbee69b35a Multi-mode search online service system supports application scenarios such as text search and text search. The whole online service consists of the following parts: Query preprocessing service: encapsulate preprocessing logic (including text/image, etc.) of pre-training model, and provide services through gRPC interface; Retrieval algorithm service: the whole algorithm processing link includes AB experiment tangent flow configuration, MetaSpore Serving call, vector recall, sorting, document summary, etc.; User entry service: provides a Web UI interface for users to debug and track down problems in the retrieval service. From a user request perspective, these services form invocation dependencies from back to front, so to build up a multimodal sample, you need to run each service from front to back first. Before doing this, remember to export the offline model, put it online and build the library first. This article will introduce the various parts of the online service system and make the whole service system step by step according to the following guidance. See the ReadME at the end of this article for more details. 1.2.1 Query preprocessing service Deep learning models tend to be based on tensors, but NLP/CV models often have a preprocessing part that translates raw text and images into tensors that deep learning models can accept. For example, NLP class models often have a pre-tokenizer to transform text data of string type into discrete tensor data. CV class models also have similar processing logic to complete the cropping, scaling, transformation, and other processing of input images through preprocessing. On the one hand, considering that this part of preprocessing logic is decoupled from tensor reasoning of the depth model, on the other hand, the reason of the depth model has an independent technical system based on ONNX, so MetaSpore disassembled this part of preprocessing logic. NLP pretreatment Tokenizer has been integrated into the Query pretreatment service. MetaSpore dismantlement with a relatively general convention. Users only need to provide preprocessing logic files to realize the loading and prediction interface and export the necessary data and configuration files loaded into the preprocessing service. Subsequent CV preprocessing logic will also be integrated in this manner. The preprocessing service currently provides the gRPC interface invocation externally and is dependent on the Query preprocessing (QP) module in the retrieval algorithm service. After the user request reaches the retrieval algorithm service, it will be forwarded to the service to complete the data preprocessing and continue the subsequent processing. The ReadMe provides details on how the preprocessing service is started, how the preprocessing model exported offline to cloud storage enters the service, and how to debug the service. To further improve the efficiency and stability of model reasoning, MetaSpore Serving implements a Python preprocessing submodule. So MetaSpore can provide gRPC services through user-specified preprocessor.py, complete Tokenizer or CV-related preprocessing in NLP, and translate requests into a Tensor that deep models can handle. Finally, the model inference is carried out by MetaSpore, Serving subsequent sub-modules. Presented here on the lot code: https://github.com/meta-soul/MetaSpore/compare/add\python\preprocessor 1.2.2 Retrieval algorithm services Retrieval algorithm service is the core of the whole online service system, which is responsible for the triage of experiments, the assembly of algorithm chains such as preprocessing, recall, sorting, and the invocation of dependent component services. The whole retrieval algorithm service is developed based on the Java Spring framework and supports multi-mode retrieval scenarios of text search and text search graph. Due to good internal abstraction and modular design, it has high flexibility and can be migrated to similar application scenarios at a low cost. Here’s a quick guide to configuring the environment to set up the retrieval algorithm service. See ReadME for more details: Install dependent components. Use Maven to install the online-Serving component Search for service configurations. Copy the template configuration file and replace the MongoDB, Milvus, and other configurations based on the development/production environment. Install and configure Consul. Consul allows you to synchronize the search service configuration in real-time, including cutting the flow of experiments, recall parameters, and sorting parameters. The project’s configuration file shows the current configuration parameters of text search and text search. The parameter modelName in the stage of pretreatment and recall is the corresponding model exported in offline processing. Start the service. Once the above configuration is complete, the retrieval service can be started from the entry script. Once the service is started, you can test it! For example, for a user with userId=10 who wants to query “How to renew ID card,” access the text search service. 1.2.3 User Entry Service Considering that the retrieval algorithm service is in the form of the API interface, it is difficult to locate and trace the problem, especially for the text search image scene can intuitively display the retrieval results to facilitate the iterative optimization of the retrieval algorithm. This paper provides a lightweight Web UI interface for text search and image search, a search input box, and results in a display page for users. Developed by Flask, the service can be easily integrated with other retrieval applications. The service calls the retrieval algorithm service and displays the returned results on the page. It’s also easy to install and start the service. Once you’re done, go to http://127.0.0.1:8090 to see if the search UI service is working correctly. See the ReadME at the end of this article for details. Multimodal system demonstration The multimodal retrieval service can be started when offline processing and online service environment configuration have been completed following the above instructions. Examples of textual searches are shown below. Enter the entry of the text search map application, enter “cat” first, and you can see that the first three digits of the returned result are cats: https://preview.redd.it/0n5nuyvhez291.png?width=1280&format=png&auto=webp&s=1e9c054f541d53381674b8d6001b4bf524506bd2 If you add a color constraint to “cat” to retrieve “black cat,” you can see that it does return a black cat: https://preview.redd.it/rzc0qjyjez291.png?width=1280&format=png&auto=webp&s=d5bcc503ef0fb3360c7740e60e295cf372dcad47 Further, strengthen the constraint on the search term, change it to “black cat on the bed,” and return results containing pictures of a black cat climbing on the bed: &#x200B; https://preview.redd.it/c4b2q8olez291.png?width=1280&format=png&auto=webp&s=4f3817b0b9f07e1e68d1d4a8281702ba3834a00a The cat can still be found through the text search system after the color and scene modification in the above example. Conclusion The cutting-edge pre-training technology can bridge the semantic gap between different modes, and the HuggingFace community can greatly reduce the cost for developers to use the pre-training model. Combined with the technological ecology of MetaSpore online reasoning and online microservices provided by DMetaSpore, the pre-training model is no longer mere offline dabbling. Instead, it can truly achieve end-to-end implementation from cutting-edge technology to industrial scenarios, fully releasing the dividends of the pre-training large model. In the future, DMetaSoul will continue to improve and optimize the MetaSpore technology ecosystem: More automated and wider access to HuggingFace community ecology. MetaSpore will soon release a common model rollout mechanism to make HuggingFace ecologically accessible and will later integrate preprocessing services into online services. Multi-mode retrieval offline algorithm optimization. For multimodal retrieval scenarios, MetaSpore will continuously iteratively optimize offline algorithm components, including text recall/sort model, graphic recall/sort model, etc., to improve the accuracy and efficiency of the retrieval algorithm. For related code and reference documentation in this article, please visit: https://github.com/meta-soul/MetaSpore/tree/main/demo/multimodal/online Some images source: https://github.com/openai/CLIP/raw/main/CLIP.png https://www.sbert.net/examples/training/sts/README.html

[D] AI regulation: a review of NTIA's "AI Accountability Policy" doc
reddit
LLM Vibe Score0
Human Vibe Score0.667
elehman839This week

[D] AI regulation: a review of NTIA's "AI Accountability Policy" doc

How will governments respond to the rapid rise of AI? How can sensible regulation keep pace with AI technology? These questions interest many of us! One early US government response has come from the National Telecommunications and Information Administration (NTIA). Specifically, the NTIA published an "AI Accountability Policy Request for Comment" on April 11, 2023. I read the NTIA document carefully, and I'm sharing my observations here for others interested in AI regulation. You can, of course, read the original materials and form your own opinions. Moreover, you can share those opinions not only on this post, but also with the NTIA itself until June 12, 2023. As background, the NTIA (homepage, Wikipedia) consists of a few hundred people within the Department of Commerce. The official mission of the NTIA is "advising the President on telecommunications and information policy issues". Topics covered by NTIA include broadband internet access, spectrum management, internet health, and now artificial intelligence. I do not know whether the NTIA will ultimately drive thinking around AI regulation in the United States or they are just a spunky lot who got something on paper early. The NTIA document is not a specific policy proposal, but rather a thoughtful discussion of AI regulation, followed by a long list of questions on which the NTIA seeks input. This format seems appropriate right now, as we're all trying to make sense of a fast-changing world. The NTIA document leans heavily on two others: the Blueprint for an AI Bill of Rights from the White House Office of Science and Technology and the AI Risk Management Framework from the National Institute of Standards and Technology (NIST). Without going into these two in depth, even tiny snippets convey their differing audiences and flavors: White House Blueprint: "You should be protected from safe and ineffective systems." NIST Framework: "Risk refers to the composite measure of an event’s probability of occurring and the magnitude or degree of the consequences of the corresponding event." Now, turning back to the NTIA document itself, I'll comment on three aspects (1) scope, (2) problems addressed, and (3) solutions contemplated. Scope is critical to understanding the NTIA document, and is probably worth keeping in mind in all near-term discussion of AI regulation. Over the past several years, at least two different technologies have been called "AI". The document mentions both, but the emphasis is NOT on the one you're probably thinking about. In more detail: A few years ago, regulators began scrutinizing "automated decisions systems", which passed as "AI" in those ancient times. An example would be an ML model used by a bank to decide whether or not you get a loan. That model might take in all sorts of information about you, combine it in mysterious ML ways, and reject your loan request. Then you might wonder, "Did that system effectively use my address and name to deduce that I am black and then reject my loan request on the basis of race?" There is some evidence of that happening, and this seems like an injustice. So perhaps such systems should be audited and certified so people know this won't happen. This is the focus of the document. These days, AI more commonly refers to open-ended systems that can engage on a wide range of topics and approximate human intelligence. The document briefly mentions generative AI models, large language models, ChatGPT, and "foundational models" (sic), but this is not the focus. The passing mentions may obscure this, unfortunately. In my opinion, these two notions of "AI" are radically different, and many of the differences matter from a regulatory perspective. Yet NTIA lumps both under a sweeping definition of an "AI system" as "an engineered or machine-based system that can, for a given set of objectives, generate outputs such as predictions, recommendations, or decisions influencing real or virtual environments." (Hmm, this includes my Magic 8-Ball…) Keep scope in mind as we turn to the next aspect: the problems under discussion. Now, NTIA's goal is to solicit input, so considering a wide range of potential problems associated with AI makes sense. Consistent with that, the document refers to democratic values, civil rights, civil liberties, and privacy. And citing the NIST doc, NTIA vaguely notes "a wide range of potential AI risks". Also, AI systems should be "valid and reliable, safe, secure and resilient, accountable and transparent, explainable and interpretable, privacy-enhanced, and fair with their harmful bias managed". And they should call their mothers \every\ week. (Okay, I made that one up.) A few comments on this formulation of the problem. First, these concerns feel more applicable to older-style AI. This includes automated decisions systems, like for a bank loan or for a prison parole recommendation. Sure, I believe such systems should operate in ways consistent with our consensus societal values, and further regulation may be needed to achieve that. But, hello! There's also another, newer class of AI that poses additional challenges. And I don't see those discussed in the NTIA document. Such challenges might include: People losing jobs because AI takes their work. Ensuring malicious people don't use AI tools to wreak havoc on the world. Sorting out intellectual property issues around AI to ensure both rapid progress in the field and respect for creators' rights. Ensuring laws appropriately assign culpability to humans when AIs cause harm. Planning for an incident analogous to the first internet worm, where an AI goes rogue, wreaks some havoc, and everyone is shocked (before it happens 28,385 more times). Bottom line: when I cntrl-F the doc for "robotic overlords", I get zero hits. ZERO. This is why I now believe scope is so important when considering efforts to regulate AI: are we talking about old-school AI or 2023-era AI or what? Because they are pretty different. The last aspect I'll address is the solutions contemplated. Again, NTIA's goal is to stimulate discussion, not propose something specific. Nevertheless, there is a strong push in one particular direction: unlike, "robotic overlord", the word "audit" appears more than 100 times along with many instances of "assessment" and "certification". On one hand, this approach makes sense. Suppose you want to ensure that a bank loan system is fair, that a social media platform isn't spreading misinformation, that a search engine is returning accurate results, etc. Then someone, somewhere has to assess or audit that system and look for problems. That audit might be done by the creator of the system or a third-party auditing agency. Such audits could be incentivized by mandates, prizes, or shiny gold stars. The government might help by fostering development of auditing tools and data. The NTIA is open to all such possibilities and seeks input on how to proceed. On the other hand, this seems like a tactic best suited to automated decision systems operated by financial institutions, government agencies, and the like. Such formal processes seem a poor fit for the current AI wave. For example: Auditing will take time and money. That's something a bank might pay for a system that will run for years. For something fine-tuned over the weekend at a startup or by some guy living in his mother's basement, that's probably not going to happen. Auditing a straightforward decision system seems far easier than assessing an open-ended AI. Beyond basic practicality, the AI could be taught to lie when it senses an audit. Also, auditing procedures (like the NTIA doc itself) will presumably be online, which means that AIs will read them and could potentially respond. Most current ML models fix parameters after training, but I think we'll soon see some models whose parameters evolve as they engage with the world. Auditing such a system that varies continuously over time seems especially difficult. Auditing a foundation model probably tells you little about derivative models. A sweet-hearted model can surely be made into monster with moderate additional training; you don't need to teach the model new cognitive skills, just repurpose existing ones to new ends. More generally, auditing doesn't address many of my concerns about AI regulation (see list above). For example, auditing sort of assumes a basically responsible actor (bank, government agency, big tech company), but AI could be misused by malicious people who, naturally, will not seek a responsible outside assessment. In any case, for both old-school and modern AI, auditing is only one line of defense, and that's not enough. You can audit until you're blue in the face, stuff will still get through, and AI systems will still cause some harm. So what's the next line of defense? For example, is our legal system ready to sensibly assign culpability to humans for AI-related incidents? In summary, the critical problem with the NTIA document is that it creates a largely false appearance of US government engagement with the new class of AI technology. As a result, people could wrongly believe that the US government is already responding to the rise of AI, and fail to advocate for actual, effective engagement. That said, the NTIA document does address important issues around a prominent technology sometimes (formerly?) called "AI". Even there, however, the proposed approach (auditing) seems like an overly-fragile, single line of defense.

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[D] Overwhelmed by fast advances in recent weeks
reddit
LLM Vibe Score0
Human Vibe Score1
iamx9000againThis week

[D] Overwhelmed by fast advances in recent weeks

I was watching the GTC keynote and became entirely overwhelmed by the amount of progress achieved from last year. I'm wondering how everyone else feels. &#x200B; Firstly, the entire ChatGPT, GPT-3/GPT-4 chaos has been going on for a few weeks, with everyone scrambling left and right to integrate chatbots into their apps, products, websites. Twitter is flooded with new product ideas, how to speed up the process from idea to product, countless promp engineering blogs, tips, tricks, paid courses. &#x200B; Not only was ChatGPT disruptive, but a few days later, Microsoft and Google also released their models and integrated them into their search engines. Microsoft also integrated its LLM into its Office suite. It all happenned overnight. I understand that they've started integrating them along the way, but still, it seems like it hapenned way too fast. This tweet encompases the past few weeks perfectly https://twitter.com/AlphaSignalAI/status/1638235815137386508 , on a random Tuesday countless products are released that seem revolutionary. &#x200B; In addition to the language models, there are also the generative art models that have been slowly rising in mainstream recognition. Now Midjourney AI is known by a lot of people who are not even remotely connected to the AI space. &#x200B; For the past few weeks, reading Twitter, I've felt completely overwhelmed, as if the entire AI space is moving beyond at lightning speed, whilst around me we're just slowly training models, adding some data, and not seeing much improvement, being stuck on coming up with "new ideas, that set us apart". &#x200B; Watching the GTC keynote from NVIDIA I was again, completely overwhelmed by how much is being developed throughout all the different domains. The ASML EUV (microchip making system) was incredible, I have no idea how it does lithography and to me it still seems like magic. The Grace CPU with 2 dies (although I think Apple was the first to do it?) and 100 GB RAM, all in a small form factor. There were a lot more different hardware servers that I just blanked out at some point. The omniverse sim engine looks incredible, almost real life (I wonder how much of a domain shift there is between real and sim considering how real the sim looks). Beyond it being cool and usable to train on synthetic data, the car manufacturers use it to optimize their pipelines. This change in perspective, of using these tools for other goals than those they were designed for I find the most interesting. &#x200B; The hardware part may be old news, as I don't really follow it, however the software part is just as incredible. NVIDIA AI foundations (language, image, biology models), just packaging everything together like a sandwich. Getty, Shutterstock and Adobe will use the generative models to create images. Again, already these huge juggernauts are already integrated. &#x200B; I can't believe the point where we're at. We can use AI to write code, create art, create audiobooks using Britney Spear's voice, create an interactive chatbot to converse with books, create 3D real-time avatars, generate new proteins (?i'm lost on this one), create an anime and countless other scenarios. Sure, they're not perfect, but the fact that we can do all that in the first place is amazing. &#x200B; As Huang said in his keynote, companies want to develop "disruptive products and business models". I feel like this is what I've seen lately. Everyone wants to be the one that does something first, just throwing anything and everything at the wall and seeing what sticks. &#x200B; In conclusion, I'm feeling like the world is moving so fast around me whilst I'm standing still. I want to not read anything anymore and just wait until everything dies down abit, just so I can get my bearings. However, I think this is unfeasible. I fear we'll keep going in a frenzy until we just burn ourselves at some point. &#x200B; How are you all fairing? How do you feel about this frenzy in the AI space? What are you the most excited about?

The delicate balance of building an online community business
reddit
LLM Vibe Score0
Human Vibe Score0.895
matthewbarbyThis week

The delicate balance of building an online community business

Hey /r/Entrepreneur 👋 Just under two years ago I launched an online community business called Traffic Think Tank with two other co-founders, Nick Eubanks and Ian Howells. As a Traffic Think Tank customer you (currently) pay $119 a month to get access to our online community, which is run through Slack. The community is focused on helping you learn various aspects of marketing, with a particular focus on search engine optimization (SEO). Alongside access to the Slack community, we publish new educational video content from outside experts every week that all customers have access to. At the time of writing, Traffic Think Tank has around 650 members spanning across 17 of the 24 different global time zones. I was on a business trip over in Sydney recently, and during my time there I met up with some of our Australia-based community members. During dinner I was asked by several of them how the idea for Traffic Think Tank came about and what steps we took to validate that the idea was worth pursuing.  This is what I told them… How it all began It all started with a personal need. Nick, an already successful entrepreneur and owner of a marketing agency, had tested out an early version Traffic Think Tank in early 2017. He offered real-time consulting for around ten customers that he ran from Slack. He would publish some educational videos and offer his advice on projects that the members were running. The initial test went well, but it was tough to maintain on his own and he had to charge a fairly high price to make it worth his time. That’s when he spoke to me and Ian about turning this idea into something much bigger. Both Ian and I offered something slightly different to Nick. We’ve both spent time in senior positions at marketing agencies, but currently hold senior director positions in 2,000+ public employee companies (HubSpot and LendingTree). Alongside this, as a trio we could really ramp up the quality and quantity of content within the community, spread out the administrative workload and just generally have more resources to throw at getting this thing off the ground. Admittedly, Nick was much more optimistic about the potential of Traffic Think Tank – something I’m very thankful for now – whereas Ian and I were in the camp of “you’re out of your mind if you think hundreds of people are going to pay us to be a part of a Slack channel”. To validate the idea at scale, we decided that we’d get an initial MVP of the community up and running with a goal of reaching 100 paying customers in the first six months. If we achieved that, we’d validated that it was a viable business and we would continue to pursue it. If not, we’d kill it. We spent the next month building out the initial tech stack that enabled us to accept payments, do basic user management to the Slack channel, and get a one-page website up and running with information on what Traffic Think Tank was all about.  After this was ready, we doubled down on getting some initial content created for members – I mean, we couldn’t have people just land in an empty Slack channel, could we? We created around ten initial videos, 20 or so articles and then some long threads full of useful information within the Slack channel so that members would have some content to pour into right from the beginning.  Then, it was time to go live. The first 100 customers Fortunately, both Nick and I had built a somewhat substantial following in the SEO space over the previous 5-10 years, so we at least had a large email list to tap into (a total of around 40,000 people). We queued up some launch emails, set an initial price of $99 per month and pressed send. [\[LINK\] The launch email I sent to my subscribers announcing Traffic Think Tank](https://mailchi.mp/matthewbarby/future-of-marketing-1128181) What we didn’t expect was to sell all of the initial 100 membership spots in the first 72 hours. “Shit. What do we do now? Are we ready for this many people? Are we providing them with enough value? What if something breaks in our tech stack? What if they don’t like the content? What if everyone hates Slack?” All of these were thoughts running through my head. This brings me to the first great decision we made: we closed down new membership intake for 3 months so that we could focus completely on adding value to the first cohort of users. The right thing at the right time SEO is somewhat of a dark art to many people that are trying to learn about it for the first time. There’s hundreds of thousands (possibly millions) of articles and videos online that talk about how to do SEO.  Some of it’s good advice; a lot of it is very bad advice.  Add to this that the barrier to entry of claiming to be an “expert” in SEO is practically non-existent and you have a recipe for disaster. This is why, for a long time, individuals involved in SEO have flocked in their masses to online communities for information and to bounce ideas off of others in the space. Forums like SEObook, Black Hat World, WickedFire, Inbound.org, /r/BigSEO, and many more have, at one time, been called home by many SEOs.  In recent times, these communities have either been closed down or just simply haven’t adapted to the changing needs of the community – one of those needs being real-time feedback on real-world problems.  The other big need that we all spotted and personally had was the ability to openly share the things that are working – and the things that aren’t – in SEO within a private forum. Not everyone wanted to share their secret sauce with the world. One of the main reasons we chose Slack as the platform to run our community on was the fact that it solved these two core needs. It gave the ability to communicate in real-time across multiple devices, and all of the information shared within it was outside of the public domain. The other problem that plagued a lot of these early communities was spam. Most of them were web-based forums that were free to access. That meant they became a breeding ground for people trying to either sell their services or promote their own content – neither of which is conducive to building a thriving community. This was our main motivation for charging a monthly fee to access Traffic Think Tank. We spent a lot of time thinking through pricing. It needed to be enough money that people would be motivated to really make use of their membership and act in a way that’s beneficial to the community, but not too much money that it became cost prohibitive to the people that would benefit from it the most. Considering that most of our members would typically spend between $200-800 per month on SEO software, $99 initially felt like the perfect balance. Growing pains The first three months of running the community went by without any major hiccups. Members were incredibly patient with us, gave us great feedback and were incredibly helpful and accommodating to other members. Messages were being posted every day, with Nick, Ian and myself seeding most of the engagement at this stage.  With everything going smoothly, we decided that it was time to open the doors to another intake of new members. At this point we’d accumulated a backlog of people on our waiting list, so we knew that simply opening our doors would result in another large intake. Adding more members to a community has a direct impact on the value that each member receives. For Traffic Think Tank in particular, the value for members comes from three areas: The ability to have your questions answered by me, Nick and Ian, as well as other members of the community. The access to a large library of exclusive content. The ability to build connections with the wider community. In the early stages of membership growth, there was a big emphasis on the first of those three points. We didn’t have an enormous content library, nor did we have a particularly large community of members, so a lot of the value came from getting a lot of one-to-one time with the community founders. [\[IMAGE\] Screenshot of engagement within the Traffic Think Tank Slack community](https://cdn.shortpixel.ai/client/qglossy,retimg,w_1322/https://www.matthewbarby.com/wp-content/uploads/2019/08/Community-Engagement-in-Traffic-Think-Tank.png) The good thing about having 100 members was that it was just about feasible to give each and every member some one-to-one time within the month, which really helped us to deliver those moments of delight that the community needed early on. Two-and-a-half months after we launched Traffic Think Tank, we opened the doors to another 250 people, taking our total number of members to 350. This is where we experienced our first growing pains.  Our original members had become used to being able to drop us direct messages and expect an almost instant response, but this wasn’t feasible anymore. There were too many people, and we needed to create a shift in behavior. We needed more value to come from the community engaging with one another or we’d never be able to scale beyond this level. We started to really pay attention to engagement metrics; how many people were logging in every day, and of those, how many were actually posting messages within public channels.  We asked members that were logging in a lot but weren’t posting (the “lurkers”) why that was the case. We also asked the members that engaged in the community the most what motivated them to post regularly. We learned a lot from doing this. We found that the large majority of highly-engaged members had much more experience in SEO, whereas most of the “lurkers” were beginners. This meant that most of the information being shared in the community was very advanced, with a lot of feedback from the beginners in the group being that they “didn’t want to ask a stupid question”.  As managers of the community, we needed to facilitate conversations that catered to all of our members, not just those at a certain level of skill. To tackle this problem, we created a number of new channels that had a much deeper focus on beginner topics so novice members had a safe place to ask questions without judgment.  We also started running live video Q&As each month where we’d answer questions submitted by the community. This gave our members one-on-one time with me, Nick and Ian, but spread the value of these conversations across the whole community rather than them being hidden within private messages. As a result of these changes, we found that the more experienced members in the community were really enjoying sharing their knowledge with those with less experience. The number of replies within each question thread was really starting to increase, and the community started to shift away from just being a bunch of threads created by me, Nick and Ian to a thriving forum of diverse topics compiled by a diverse set of individuals. This is what we’d always wanted. A true community. It was starting to happen. [\[IMAGE\] Chart showing community engagement vs individual member value](https://cdn.shortpixel.ai/client/qglossy,retimg,w_1602/https://www.matthewbarby.com/wp-content/uploads/2019/08/Community-Engagement-Balance-Graph.jpg) At the same time, we started to realize that we’ll eventually reach a tipping point where there’ll be too much content for us to manage and our members to engage with. When we reach this point, the community will be tough to follow and the quality of any given post will go down. Not only that, but the community will become increasingly difficult to moderate. We’re not there yet, but we recognize that this will come, and we’ll have to adjust our model again. Advocating advocacy As we started to feel more comfortable about the value that members were receiving, we made the decision to indefinitely open for new members. At the same time, we increased the price of membership (from $99 a month to $119) in a bid to strike the right balance between profitability as a business and to slow down the rate at which we were reaching the tipping point of community size. We also made the decision to repay all of our early adopters by grandfathering them in to the original pricing – and committing to always do this in the future. Despite the price increase, we saw a continued flow of new members come into the community. The craziest part about this was that we were doing practically no marketing activities to encourage new members– this was all coming from word of mouth. Our members were getting enough value from the community that they were recommending it to their friends, colleagues and business partners.  The scale at which this was happening really took us by surprise and it told us one thing very clearly: delivering more value to members resulted in more value being delivered to the business. This is a wonderful dynamic to have because it perfectly aligns the incentives on both sides. We’d said from the start that we wouldn’t sacrifice value to members for more revenue – this is something that all three of us felt very strongly about. First and foremost, we wanted to create a community that delivered value to its members and was run in a way that aligned with our values as people. If we could find a way to stimulate brand advocacy, while also tightening the bonds between all of our individual community members, we’d be boosting both customer retention and customer acquisition in the same motion. This became our next big focus. [\[TWEET\] Adam, one of our members wore his Traffic Think Tank t-shirt in the Sahara desert](https://twitter.com/AdamGSteele/status/1130892481099382784) We started with some simple things: We shipped out Traffic Think Tank branded T-shirts to all new members. We’d call out each of the individuals that would submit questions to our live Q&A sessions and thank them live on air. We set up a new channel that was dedicated to sharing a quick introduction to who you are, what you do and where you’re based for all new members. We’d created a jobs channel and a marketplace for selling, buying and trading services with other members. Our monthly “blind dates” calls were started where you’d be randomly grouped with 3-4 other community members so that you could hop on a call to get to know each other better. The Traffic Think Tank In Real Life (IRL)* channel was born, which enabled members to facilitate in-person meetups with each other. In particular, we saw that as members started to meet in person or via calls the community itself was feeling more and more like a family. It became much closer knit and some members started to build up a really positive reputation for being particularly helpful to other members, or for having really strong knowledge in a specific area. [\[TWEET\] Dinner with some of the Traffic Think Tank members in Brighton, UK](https://twitter.com/matthewbarby/status/1117175584080134149) Nick, Ian and I would go out of our way to try and meet with members in real life wherever we could. I was taken aback by how appreciative people were for us doing this, and it also served as an invaluable way to gain honest feedback from members. There was another trend that we’d observed that we didn’t really expect to happen. More and more members were doing business with each another. We’ve had people find new jobs through the community, sell businesses to other members, launch joint ventures together and bring members in as consultants to their business. This has probably been the most rewarding thing to watch, and it was clear that the deeper relationships that our members were forming were resulting in an increased level of trust to work with each other. We wanted to harness this and take it to a new level. This brought us to arguably the best decision we’ve made so far running Traffic Think Tank… we were going to run a big live event for our members. I have no idea what I’m doing It’s the first week of January 2019 and we’re less than three weeks away from Traffic Think Tank LIVE, our first ever in-person event hosting 150 people, most of which are Traffic Think Tank members. It's like an ongoing nightmare I can’t wake up from. That was Nick’s response in our private admin channel to myself and Ian when I asked if they were finding the run-up to the event as stressful as I was. I think that all three of us were riding on such a high from how the community was growing that we felt like we could do anything. Running an event? How hard can it be? Well, turns out it’s really hard. We had seven different speakers flying over from around the world to speak at the event, there was a pre- and after event party, and we’d planned a charity dinner where we would take ten attendees (picked at random via a raffle) out for a fancy meal. Oh, and Nick, Ian and I were hosting a live Q&A session on stage. It wasn’t until precisely 48 hours before the event that we’d realized we didn’t have any microphones, nor had a large amount of the swag we’d ordered arrived. Plus, a giant storm had hit Philly causing a TON of flight cancellations. Perfect. Just perfect. This was honestly the tip of the iceberg. We hadn’t thought about who was going to run the registration desk, who would be taking photos during the event and who would actually field questions from the audience while all three of us sat on stage for our live Q&A panel. Turns out that the answer to all of those questions were my wife, Laura, and Nick’s wife, Kelley. Thankfully, they were on hand to save our asses. The weeks running up to the event were honestly some of the most stressful of my life. We sold around 50% of our ticket allocation within the final two weeks before the event. All of the event organizers told us this would happen, but did we believe them? Hell no!  Imagine having two weeks until the big day and as it stood half of the room would be completely empty. I was ready to fly most of my extended family over just to make it look remotely busy. [\[IMAGE\] One of our speakers, Ryan Stewart, presenting at Traffic Think Tank LIVE](https://cdn.shortpixel.ai/client/qglossy,retimg,w_1920/https://www.matthewbarby.com/wp-content/uploads/2019/08/Traffic-Think-Tank-LIVE-Ryan-Presenting.jpg) Thankfully, if all came together. We managed to acquire some microphones, the swag arrived on the morning of the event, all of our speakers were able to make it on time and the weather just about held up so that our entire allocation of ticket holders was able to make it to the event. We pooled together and I’m proud to say that the event was a huge success. While we made a substantial financial loss on the event itself, January saw a huge spike in new members, which more than recouped our losses. Not only that, but we got to hang out with a load of our members all day while they said really nice things about the thing we’d built. It was both exhausting and incredibly rewarding. Bring on Traffic Think Tank LIVE 2020! (This time we’re hiring an event manager...)   The road ahead Fast forward to today (August 2019) and Traffic Think Tank has over 650 members. The biggest challenges that we’re tackling right now include making sure the most interesting conversations and best content surfaces to the top of the community, making Slack more searchable (this is ultimately one of its flaws as a platform) and giving members a quicker way to find the exclusive content that we create. You’ll notice there’s a pretty clear theme here. In the past 30 days, 4,566 messages were posted in public channels inside Traffic Think Tank. If you add on any messages posted inside private direct messages, this number rises to 21,612. That’s a lot of messages. To solve these challenges and enable further scale in the future, we’ve invested a bunch of cash and our time into building out a full learning management system (LMS) that all members will get access to alongside the Slack community. The LMS will be a web-based portal that houses all of the video content we produce. It will also  provide an account admin section where users can update or change their billing information (they have to email us to do this right now, which isn’t ideal), a list of membership perks and discounts with our partners, and a list of links to some of the best threads within Slack – when clicked, these will drop you directly into Slack. [\[IMAGE\] Designs for the new learning management system (LMS)](https://cdn.shortpixel.ai/client/qglossy,retimg,w_2378/https://www.matthewbarby.com/wp-content/uploads/2019/08/Traffic-Think-Tank-LMS.png) It’s not been easy, but we’re 95% of the way through this and I’m certain that it will have a hugely positive impact on the experience for our members. Alongside this we hired a community manager, Liz, who supports with any questions that our members have, coordinates with external experts to arrange webinars for the community, helps with new member onboarding, and has tightened up some of our processes around billing and general accounts admin. This was a great decision. Finally, we’ve started planning next year’s live event, which we plan to more than double in size to 350 attendees, and we decided to pick a slightly warmer location in Miami this time out. Stay tuned for me to have a complete meltdown 3 weeks from the event. Final thoughts When I look back on the journey we’ve had so far building Traffic Think Tank, there’s one very important piece to this puzzle that’s made all of this work that I’ve failed to mention so far: co-founder alignment. Building a community is a balancing act that relies heavily on those in charge being completely aligned. Nick, Ian and I completely trust each other and more importantly, are philosophically aligned on how we want to run and grow the community. If we didn’t have this, the friction between us could tear apart the entire community. Picking the right people to work with is important in any company, but when your business is literally about bringing people together, there’s no margin for error here.  While I’m sure there will be many more challenges ahead, knowing that we all trust each other to make decisions that fall in line with each of our core values makes these challenges dramatically easier to overcome. Finally, I’d like to thank all of our members for making the community what it is today – it’d be nothing without you and I promise that we’ll never take that for granted. &#x200B; I originally posted this on my blog here. Welcoming all of your thoughts, comments, questions and I'll do my best to answer them :)

Started a content marketing agency 6 years ago - $0 to $5,974,324 (2023 update)
reddit
LLM Vibe Score0
Human Vibe Score1
mr_t_forhireThis week

Started a content marketing agency 6 years ago - $0 to $5,974,324 (2023 update)

Hey friends, My name is Tyler and for the past 6 years, I’ve been documenting my experience building a content marketing agency called Optimist. Year 1 - 0 to $500k ARR Year 2 - $500k to $1MM ARR Year 3 - $1MM ARR to $1.5MM(ish) ARR Year 4 - $3,333,686 Revenue Year 5 - $4,539,659 Revenue How Optimist Works First, an overview/recap of the Optimist business model: We operate as a “collective” of full time/professional freelancers Everyone aside from me is a contractor Entirely remote/distributed team Each freelancer earns $65-85/hour Clients pay us a flat monthly fee for full-service content marketing (research, strategy, writing, editing, design/photography, reporting and analytics, targeted linkbuilding, and more) We recently introduced hourly engagements for clients who fit our model but have some existing in-house support Packages range in price from $10-20k/mo We offer profit share to everyone on our core team as a way to give everyone ownership in the company In 2022, we posted $1,434,665 in revenue. It was our highest revenue year to date and brings our lifetime total to $5,974,324. Here’s our monthly revenue from January 2017 to December of 2022. But, like every year, it was a mix of ups and downs. Here’s my dispatch for 2023. — Running a business is like spilling a drink. It starts as a small and simple thing. But, if you don’t clean it up, the spill will spread and grow — taking up more space, seeping into every crack. There’s always something you could be doing. Marketing you could be working on. Pitches you could be making. Networking you could be doing. Client work you could help with. It can be all-consuming. And it will be — if you don’t clean up the spill. I realized this year that I had no containment for the spill that I created. Running an agency was spilling over into nearly every moment of my life. When I wasn’t working, I was thinking about work. When I wasn’t thinking about work, I was dreaming about it. Over the years, I’ve shared about a lot of my personal feelings and experience as an entrepreneur. And I also discussed my reckoning with the limitations of running the business we’ve built. My acceptance that it was an airplane but not a rocket. And my plan to try to compartmentalize the agency to make room in my life for other things — new business ideas, new revenue streams, and maybe some non-income-producing activity. 🤷 What I found in 2022 was that the business wasn’t quite ready for me to make that move. It was still sucking up too much of my time and attention. There were still too many gaps to fill and I was the one who was often filling them. So what do you do? Ultimately you have two choices on the table anytime you run a business and it’s not going the way you want it: Walk away Turn the ship — slowly For a huge number of reasons (personal, professional, financial, etc), walking away from Optimist was not really even an option or the right move for me. But it did feel like things needed to change. I needed to keep turning the ship to get it to the place where it fit into my life — instead of my life fitting around the business. This means 2022 was a year of transition for the agency. (Again?) Refocusing on Profit Some money is better than no money. Right? Oddly, this was one of the questions I found myself asking in 2022. Over the years, we’ve been fortunate to have many clients who have stuck with us a long time. In some cases, we’ve had clients work with us for 2, 3, or even 4 years. (That’s over half of our existence!) But, things have gotten more expensive — we’ve all felt it. We’ve had to increase pay to remain competitive for top talent. Software costs have gone up. It’s eaten into our margin. Because of our increasing costs and evolving scope, many of our best, most loyal clients were our least profitable. In fact, many were barely profitable — if at all. We’ve tried to combat that by increasing rates on new, incoming clients to reflect our new costs and try to make up for shrinking margin on long-term clients. But we didn’t have a good strategy in place for updating pricing for current clients. And it bit us in the ass. Subsidizing lower-profit, long-term clients with new, higher-margin clients ultimately didn’t work out. Our margins continued to dwindle and some months we were barely breaking even while posting six-figures of monthly revenue. 2022 was our highest revenue year but one of our least profitable. It only left one option. We had to raise rates on some of our long-term clients. But, of course, raising rates on a great, long-term client can be delicate. You’ve built a relationship with these people over the years and you’re setting yourself up for an ultimatum — are you more valuable to the client or is the client more valuable to you? Who will blink first? We offered all of these clients the opportunity to move to updated pricing. Unfortunately, some of them weren’t on board. Again, we had 2 options: Keep them at a low/no profit rate Let them churn It seems intuitive that having a low-profit client is better than having no client. But we’ve learned an important lesson many times over the years. Our business doesn’t scale infinitely and we can only handle so many clients at a time. That means that low-profit clients are actually costing us money in some cases. Say our average client generates $2,500 per month in profit — $30,000 per year. If one of our clients is only generating $500/mo in profit, working with them means missing out on bringing on a more profitable client (assuming our team is currently at capacity). Instead of $30,000/year, we’re only making $6,000. Keeping that client costs us $24,000. That’s called opportunity cost. So it’s clear: We had to let these clients churn. We decided to churn about 25% of our existing clients. On paper, the math made sense. And we had a pretty consistent flow of new opportunities coming our way. At the time, it felt like a no-brainer decision. And I felt confident that we could quickly replace these low-profit clients with higher-margin ones. I was wrong. Eating Shit Right after we initiated proactively churning some of our clients, other clients — ones we planned to keep — gave us notice that they were planning to end the engagement. Ouch. Fuck. We went from a 25% planned drop in revenue to a nearly 40% cliff staring us right in the face. Then things got even worse. Around Q3 of this year, talk of recession and layoffs really started to intensify. We work primarily with tech companies and startups. And these were the areas most heavily impacted by the economic news. Venture funding was drying up. Our leads started to slow down. This put us in a tough position. Looking back now, I think it’s clear that I made the wrong decision. We went about this process in the wrong way. The reality sinks in when you consider the imbalance between losing a client and gaining a client. It takes 30 days for someone to fire us. It’s a light switch. But it could take 1-3 months to qualify, close, and onboard a new client. We have lots of upfront work, research, and planning that goes into the process. We have to learn a new brand voice, tone, and style. It’s a marathon. So, for every client we “trade”, there’s a lapse in revenue and work. This means that, in retrospect, I would probably have made this transition using some kind of staggered schedule rather than a cut-and-dry approach. We could have gradually off-boarded clients when we had more definitive work to replace them. I was too confident. But that’s a lesson I had to learn the hard way. Rebuilding & Resetting Most of the voluntary and involuntary churn happened toward the end of 2022. So we’re still dealing with the fall out. Right now, it feels like a period of rebuilding. We didn’t quite lose 50% of our revenue, but we definitely saw a big hit heading into 2023. To be transparent: It sucks. It feels like a gigantic mistake that I made which set us back significantly from our previous high point. I acted rashly and it cost us a lot of money — at least on the surface. But I remind myself of the situation we were in previously. Nearly twice the revenue but struggling to maintain profitability. Would it have been better to try to slowly fix that situation and battle through months of loss or barely-break-even profits? Or was ripping off the bandaid the right move after all? I’m an optimist. (Heh, heh) Plus, I know that spiraling over past decisions won’t change them or help me move forward. So I’m choosing to look at this as an opportunity — to rebuild, reset, and refocus the company. I get to take all of the tough lessons I’ve learned over the last 6 years and apply them to build the company in a way that better aligns with our new and current goals. It’s not quite a fresh, clean start, but by parting ways with some of our oldest clients, we’ve eliminated some of the “debt” that’s accumulated over the years. We get a chance to fully realize the new positioning that we rolled out last year. Many of those long-term clients who churned had a scope of work or engagement structure that didn’t fit with our new positioning and focus. So, by losing them, we’re able to completely close up shop on the SOWs that no longer align with the future version of Optimist. Our smaller roster of clients is a better fit for that future. My job is to protect that positioning by ensuring that while we’re rebuilding our new roster of clients we don’t get desperate. We maintain the qualifications we set out for future clients and only take on work that fits. How’s that for seeing the upside? Some other upside from the situation is that we got an opportunity to ask for candid feedback from clients who were leaving. We asked for insight about their decision, what factors they considered, how they perceived us, and the value of our work. Some of the reasons clients left were obvious and possibly unavoidable. Things like budget cuts, insourcing, and uncertainty about the economy all played at least some part of these decisions. But, reading between the lines, where was one key insight that really struck me. It’s one of those, “oh, yeah — duh — I already knew that,” things that can be difficult to learn and easy to forget…. We’re in the Relationship Business (Plan Accordingly) For all of our focus on things like rankings, keywords, content, conversions, and a buffet of relevant metrics, it can be easy to lose the forest for the trees. Yes, the work itself matters. Yes, the outcomes — the metrics — matter. But sometimes the relationship matters more. When you’re running an agency, you can live or die by someone just liking you. Admittedly, this feels totally unfair. It opens up all kinds of dilemmas, frustration, opportunity for bias and prejudice, and other general messiness. But it’s the real world. If a client doesn’t enjoy working with us — even if for purely personal reasons — they could easily have the power to end of engagement, regardless of how well we did our actual job. We found some evidence of this in the offboarding conversations we had with clients. In some cases, we had clients who we had driven triple- and quadruple-digital growth. Our work was clearly moving the needle and generating positive ROI and we had the data to prove it. But they decided to “take things in another direction” regardless. And when we asked about why they made the decision, it was clear that it was more about the working relationship than anything we could have improved about the service itself. The inverse is also often true. Our best clients have lasting relationships with our team. The work is important — and they want results. But even if things aren’t quite going according to plan, they’re patient and quick to forgive. Those relationships feel solid — unshakeable. Many of these folks move onto new roles or new companies and quickly look for an opportunity to work with us again. On both sides, relationships are often more important than the work itself. We’ve already established that we’re not building a business that will scale in a massive way. Optimist will always be a small, boutique service firm. We don’t need 100 new leads per month We need a small, steady roster of clients who are a great fit for the work we do and the value we create. We want them to stick around. We want to be their long-term partner. I’m not built for churn-and-burn agency life. And neither is the business. When I look at things through this lens, I realize how much I can cut from our overall business strategy. We don’t need an ultra-sophisticated, multi-channel marketing strategy. We just need strong relationships — enough of them to make our business work. There are a few key things we can take away from this as a matter of business strategy: Put most of our effort into building and strengthening relationships with our existing clients Be intentional about establishing a strong relationship with new clients as part of onboarding Focus on relationships as the main driver of future business development Embracing Reality: Theory vs Practice Okay, so with the big learnings out the way, I want to pivot into another key lesson from 2022. It’s the importance of understanding theory vs practice — specifically when it comes to thinking about time, work, and life. It all started when I was considering how to best structure my days and weeks around running Optimist, my other ventures, and my life goals outside of work. Over the years, I’ve dabbled in many different ways to block time and find focus — to compartmentalize all of the things that are spinning and need my attention. As I mapped this out, I realized that I often tried to spread myself too thin throughout the week. Not just that I was trying to do too much but that I was spreading that work into too many small chunks rather than carving out time for focus. In theory, 5 hours is 5 hours. If you have 5 hours of work to get done, you just fit into your schedule whenever you have an open time slot. In reality, a single 5-hour block of work is 10x more productive and satisfying than 10, 30-minute blocks of work spread out across the week. In part, this is because of context switching. Turning your focus from one thing to another thing takes time. Achieving flow and focus takes time. And the more you jump from one project to another, the more time you “lose” to switching. This is insightful for me both in the context of work and planning my day, but also thinking about my life outside of Optimist. One of my personal goals is to put a finite limit on my work time and give myself more freedom. I can structure that in many different ways. Is it better to work 5 days a week but log off 1 hour early each day? Or should I try to fit more hours into each workday so I can take a full day off? Of course, it’s the latter. Both because of the cost of context switching and spreading work into more, smaller chunks — but also because of the remainder that I end up with when I’m done working. A single extra hour in my day probably means nothing. Maybe I can binge-watch one more episode of a new show or do a few extra chores around the house. But it doesn’t significantly improve my life or help me find greater balance. Most things I want to do outside of work can’t fit into a single extra hour. A full day off from work unlocks many more options. I can take the day to go hiking or biking. I can spend the day with my wife, planning or playing a game. Or I can push it up against the weekend and take a 3-day trip. It gives me more of the freedom and balance that I ultimately want. So this has become a guiding principle for how I structure my schedule. I want to: Minimize context switching Maximize focused time for work and for non-work The idea of embracing reality also bleeds into some of the shifts in business strategy that I mentioned above. In theory, any time spent on marketing will have a positive impact on the company. In reality, focusing more on relationships than blasting tweets into the ether is much more likely to drive the kind of growth and stability that we’re seeking. As I think about 2023, I think this is a recurring theme. It manifests in many ways. Companies are making budget cuts and tough decisions about focus and strategy. Most of us are looking for ways to rein in the excess and have greater impact with a bit less time and money. We can’t do everything. We can’t even do most things. So our #1 priority should be to understand the reality of our time and our effort to make the most of every moment (in both work and leisure). That means thinking deeply about our strengths and our limitations. Being practical, even if it feels like sacrifice. Update on Other Businesses Finally, I want to close up by sharing a bit about my ventures outside of Optimist. I shared last year how I planned to shift some of my (finite) time and attention to new ventures and opportunities. And, while I didn’t get to devote as much as I hoped to these new pursuits, they weren’t totally in vain. I made progress across the board on all of the items I laid out in my post. Here’s what happened: Juice: The first Optimist spin-out agency At the end of 2021, we launched our first new service business based on demand from Optimist clients. Focused entirely on building links for SEO, we called the agency Juice. Overall, we made strong progress toward turning this into a legitimate standalone business in 2022. Relying mostly on existing Optimist clients and a few word-of-mouth opportunities (no other marketing), we built a team and set up a decent workflow and operations. There’s still many kinks and challenges that we’re working through on this front. All told, Juice posted almost $100,000 in revenue in our first full year. Monetizing the community I started 2022 with a focus on figuring out how to monetize our free community, Top of the Funnel. Originally, my plan was to sell sponsorships as the main revenue driver. And that option is still on the table. But, this year, I pivoted to selling paid content and subscriptions. We launched a paid tier for content and SEO entrepreneurs where I share more of my lessons, workflows, and ideas for building and running a freelance or agency business. It’s gained some initial traction — we reached \~$1,000 MRR from paid subscriptions. In total, our community revenue for 2022 was about $2,500. In 2023, I’m hoping to turn this into a $30,000 - $50,000 revenue opportunity. Right now, we’re on track for \~$15,000. Agency partnerships and referrals In 2022, we also got more serious about referring leads to other agencies. Any opportunity that was not a fit for Optimist or we didn’t have capacity to take on, we’d try to connect with another partner. Transparently, we struggled to operationalize this as effectively as I would have liked. In part, this was driven by my lack of focus here. With the other challenges throughout the year, I wasn’t able to dedicate as much time as I’d like to setting goals and putting workflows into place. But it wasn’t a total bust. We referred out several dozen potential clients to partner agencies. Of those, a handful ended up converting into sales — and referral commission. In total, we generated about $10,000 in revenue from referrals. I still see this as a huge opportunity for us to unlock in 2023. Affiliate websites Lastly, I mentioned spending some time on my new and existing affiliate sites as another big business opportunity in 2022. This ultimately fell to the bottom of my list and didn’t get nearly the attention I wanted. But I did get a chance to spend a few weeks throughout the year building this income stream. For 2022, I generated just under $2,000 in revenue from affiliate content. My wife has graciously agreed to dedicate some of her time and talent to these projects. So, for 2023, I think this will become a bit of a family venture. I’m hoping to build a solid and consistent workflow, expand the team, and develop a more solid business strategy. Postscript — AI, SEO, OMG As I’m writing this, much of my world is in upheaval. If you’re not in this space (and/or have possibly been living under a rock), the release of ChatGPT in late 2022 has sparked an arms race between Google, Bing, OpenAI, and many other players. The short overview: AI is likely to fundamentally change the way internet search works. This has huge impact on almost all of the work that I do and the businesses that I run. Much of our focus is on SEO and understanding the current Google algorithm, how to generate traffic for clients, and how to drive traffic to our sites and projects. That may all change — very rapidly. This means we’re standing at a very interesting point in time. On the one hand, it’s scary as hell. There’s a non-zero chance that this will fundamentally shift — possibly upturn — our core business model at Optimist. It could dramatically change how we work and/or reduce demand for our core services. No bueno. But it’s also an opportunity (there’s the optimist in me, again). I certainly see a world where we can become leaders in this new frontier. We can pivot, adjust, and capitalize on a now-unknown version of SEO that’s focused on understanding and optimizing for AI-as-search. With that, we may also be able to help others — say, those in our community? — also navigate this tumultuous time. See? It’s an opportunity. I wish I had the answers right now. But, it’s still a time of uncertainty. I just know that there’s a lot of change happening and I want to be in front of it rather than trying to play catch up. Wish me luck. — Alright friends — that's my update for 2023! I’ve always appreciated sharing these updates with the Reddit community, getting feedback, being asked tough questions, and even battling it out with some of my haters (hey!! 👋) As usual, I’m going to pop in throughout the next few days to respond to comments or answer questions. Feel free to share thoughts, ideas, and brutal takedowns in the comments. If you're interested in following the Optimist journey and the other projects I'm working on in 2023, you can follow me on Twitter. Cheers, Tyler P.S. - If you're running or launching a freelance or agency business and looking for help figuring it out, please DM me. Our subscription community, Middle of the Funnel, was created to provide feedback, lessons, and resources for other entrepreneurs in this space.

Beginner to the 1st sale: my journey building an AI for social media marketers
reddit
LLM Vibe Score0
Human Vibe Score1
Current-Payment-5403This week

Beginner to the 1st sale: my journey building an AI for social media marketers

Hey everyone! Here’s my journey building an AI for social media marketers all the way up until my first pre-launch sale, hope that could help some of you: My background: studied maths at uni before dropping out to have some startup experiences. Always been drawn to building new things so I reckoned I would have some proper SaaS experiences and see how VC-funded startups are doing it before launching my own.  I’ve always leaned towards taking more risks in my life so leaving my FT job to launch my company wasn’t a big deal for me (+ I’m 22 so still have time to fail over and over). When I left my job, I started reading a lot about UI/UX, no-code tools, marketing, sales and every tool a worthwhile entrepreneur needs to learn about. Given the complexity of the project I set out to achieve, I asked a more technical friend to join as a cofounder and that's when AirMedia was born. We now use bubble for landing page as I had to learn it and custom-code stack for our platform.  Here's our goal: streamlining social media marketing using AI. I see this technology has only being at the premises of what it will be able to achieve in the near-future. We want to make the experience dynamic i.e. all happens from a discussion and you see the posts being analysed from there as well as the creation process - all from within the chat. Fast forward a few weeks ago, we finished developing the first version of our tool that early users describe as a "neat piece of tech" - just this comment alone can keep me going for months :) Being bootstrapped until now, I decided to sell lifetime deals for the users in the waitlist that want to get the tool in priority as well as secure their spot for life. We've had the first sale the first day we made that public ! Now what you all are looking for: How ?  Here was my process starting to market the platform: I need a high-converting landing page so I reckoned which companies out there have the most data and knows what convert and what doesn’t: Unbounce. Took their landing page and adapted it to my value proposition and my ICP.  The ICP has been defined from day 1 and although I’m no one to provide any advice, I strongly believe the ICP has to be defined from day 1 (even before deciding the name of the company). It helps a lot when the customer is you and you’ve had this work experience that helps you identify the problems your users encounter. Started activating the network, posting on Instagram and LinkedIn about what we've built (I've worked in many SaaS start-ups in the past so I have to admit that's a bit of a cheat code). Cold outreach from Sales NAV to our ICP, been growing the waitlist in parallel of building the tool for months now so email marketings with drip sequences and sharing dev updates to build the trust along the way (after all we're making that tool for our users - they should be the first aware about what we're building). I also came across some Whatsapp groups with an awesome community that welcomed our platform with excitement.) The landing page funnel is the following: Landing page -> register waitlist -> upsell page -> confirmation. I've made several landing pages e.g. for marketing agencies, for real estate agents, for marketing director in several different industries. The goal now is just testing out the profiles and who does it resonate the most with. Another growth hack that got us 40+ people on the waitlist: I identified some Instagram posts from competitors where their CTA was "comment AI" and I'll send you our tool and they got over 2k people commenting. Needless to say, I messaged every single user to check out our tool and see if it could help them. (Now that i think about it, the 2% conversion rate there is not great - especially considering the manual labour and the time put behind it). We’ve now got over 400 people on the waitlist so I guess we’re doing something right but we’ll keep pushing as the goal is to sell these lifetime deals to have a strong community to get started. (Also prevents us from going to VCs and I can keep my time focussing exclusively on our users - I’m not into boardroom politics, just wanna build something useful for marketers). Now I’m still in the process of testing out different marketing strategies while developing and refining our platform to make it next level on launch day. Amongst those:  LinkedIn Sales Nav outreach (first sale came from there) Product Hunt Highly personalised cold emails (there I’m thinking of doing 20 emails a day with a personalised landing page to each of those highly relevant marketers). Never seen that and I think this could impress prospects but not sure it’s worth it time / conversion wise. Make content to could go viral (at least 75 videos) that I’m posting throughout several social media accounts such as airmedia\\, airmedia\reels, airmedia\ai (you get the hack) always redirecting to the main page both in the profile description and tagging the main account. I have no idea how this will work so will certainly update some of you that would like to know the results. Will do the same across Facebook, TikTok, Youtube Shorts etc… I’m just looking for a high potential of virality there. This strategy is mainly used to grow personal brands but never seen it applied to companies. Good old cold calling Reddit (wanna keep it transparent ;) ) I’m alone to execute all these strategies + working in parallel to refine the product upon user’s feedback I’m not sure I can do more than that for now. Let me know if you have any feedback/ideas/ tasks I could implement.  I could also make another post about the proper product building process as this post was about the marketing. No I certainly haven’t accomplished anything that puts me in a position to provide advices but I reckon I’m on my way to learn more and more. Would be glad if this post could help some of you.  And of course as one of these marketing channels is Reddit I’ll post the link below for the entrepreneurs that want to streamline their social media or support us. Hope I was able to provide enough value in this post for you to consider :) https://airmedia.uk/

Recently hit 6,600,000 monthly organic traffic for a B2C SaaS website. Here's the 40 tips that helped me make that happen.
reddit
LLM Vibe Score0
Human Vibe Score1
DrJigsawThis week

Recently hit 6,600,000 monthly organic traffic for a B2C SaaS website. Here's the 40 tips that helped me make that happen.

Hey guys! So as title says, we recently hit 6,600,000 monthly organic traffic / month for a B2C SaaS website (screenshot. Can't give name publicly, but can show testimonial to a mod). Here's 40 tips that "helped" me make this happen. If you get some value of the post, I write an SEO tip every other day on /r/seogrowth. There's around 10 more tips already up there other than the ones I mention here. If you want to give back for all my walls of text, I'd appreciate a sub <3 Also, there are a bunch of free stuff I mention in the article: content outline, writer guidelines, SEO checklist, and other stuff. Here's the Google Doc with all that! Tip #1. Take SEO With a Grain of Salt A lot of the SEO advice and best practices on the internet are based on 2 things: Personal experiences and case studies of companies that managed to make SEO work for them. Google or John Mueller (Google’s Senior Webmaster Trends Analyst). And, unfortunately, neither of these sources are always accurate. Personal SEO accounts are simply about what worked for specific companies. Sometimes, what worked for others, won’t work for you. For example, you might find a company that managed to rank with zero link-building because their website already had a very strong backlink profile. If you’re starting with a fresh website, chances are, you won’t be able to get the same results. At the same time, information from Google or John Mueller is also not 100% accurate. For example, they’ve said that guest posting is against Google’s guidelines and doesn’t work… But practically, guest posting is a very effective link-building strategy. So the takeaway is this: Take all information you read about SEO with a grain of salt. Analyze the information yourself, and make your conclusions. SEO Tip #2. SEO Takes Time You’ve already heard this one before, but considering how many people keep asking, thought I'd include this anyway. On average, it’s going to take you 6 months to 2 years to get SEO results, depending on the following factors: Your backlink profile. The more quality backlinks you have (or build), the faster you’ll rank. Age of your website. If your website is older (or you purchased an aged website), you can expect your content to rank faster. Amount of content published. The more quality content you publish on your website, the more “authoritative” it is in the eyes of Google, and thus more likely to rank faster. SEO work done on the website. If a lot of your pages are already ranking on Google (page 2-3), it’s easier to get them to page #1 than if you just published the content piece. Local VS global SEO. Ranking locally is (sometimes) easier and faster than ranking globally. That said, some marketing agencies can use “SEO takes time” as an excuse for not driving results. Well, fortunately, there is a way to track SEO results from month #2 - #3 of work. Simply check if your new content pieces/pages are getting more and more impressions on Google Search Console month-to-month. While your content won’t be driving traffic for a while after being published, they’ll still have a growing number of impressions from month #2 or #3 since publication. SEO Tip #3. SEO Might Not Be The Best Channel For You In theory, SEO sounds like the best marketing channel ever. You manage to rank on Google and your marketing seemingly goes on auto-pilot - you’re driving new leads every day from existing content without having to lift a finger… And yet, SEO is not for everyone. Avoid SEO as a marketing channel if: You’re just getting started with your business and need to start driving revenue tomorrow (and not in 1-2 years). If this is you, try Google ads, Facebook ads, or organic marketing. Your target audience is pretty small. If you’re selling enterprise B2B software and have around 2,000 prospects in total worldwide, then it’s simply easier to directly reach out to these prospects. Your product type is brand-new. If customers don’t know your product exists, they probably won’t be Googling it. SEO Tip #4. Traffic Can Be a Vanity Metric I've seen hundreds of websites that drive 6-7 digits of traffic but generate only 200-300 USD per month from those numbers. “What’s the deal?” You might be thinking. “How can you fail to monetize that much traffic?” Well, that brings us to today’s tip: traffic can be a vanity metric. See, not all traffic is created equal. Ranking for “hormone balance supplement” is a lot more valuable than ranking for “Madagascar character names.” The person Googling the first keyword is an adult ready to buy your product. Someone Googling the latter, on the other hand, is a child with zero purchasing power. So, when deciding on which keywords to pursue, always keep in mind the buyer intent behind and don’t go after rankings or traffic just because 6-digit traffic numbers look good. SEO Tip #5. Push Content Fast Whenever you publish a piece of content, you can expect it to rank within 6 months to a year (potentially less if you’re an authority in your niche). So, the faster you publish your content, the faster they’re going to age, and, as such, the faster they’ll rank on Google. On average, I recommend you publish a minimum of 10,000 words of content per month and 20,000 to 30,000 optimally. If you’re not doing link-building for your website, then I’d recommend pushing for even more content. Sometimes, content velocity can compensate for the lack of backlinks. SEO Tip #6. Use Backlink Data to Prioritize Content You might be tempted to go for that juicy, 6-digit traffic cornerstone keyword right from the get-go... But I'd recommend doing the opposite. More often than not, to rank for more competitive, cornerstone keywords, you’ll need to have a ton of supporting content, high-quality backlinks, website authority, and so on. Instead, it’s a lot more reasonable to first focus on the less competitive keywords and then, once you’ve covered those, move on to the rest. Now, as for how to check keyword competitiveness, here are 2 options: Use Mozbar to see the number of backlinks for top-ranking pages, as well as their Domain Authority (DA). If all the pages ranking on page #1 have <5 backlinks and DA of 20 - 40, it’s a good opportunity. Use SEMrush or Ahrefs to sort your keywords by difficulty, and focus on the less difficult keywords first. Now, that said, keep in mind that both of these metrics are third-party, and hence not always accurate. SEO Tip #7. Always Start With Competitive Analysis When doing keyword research, the easiest way to get started is via competitive analysis. Chances are, whatever niche you’re in, there’s a competitor that is doing great with SEO. So, instead of having to do all the work from scratch, run their website through SEMrush or Ahrefs and steal their keyword ideas. But don’t just stop there - once you’ve borrowed keyword ideas from all your competitors, run the seed keywords through a keyword research tool such as UberSuggest or SEMrush Keyword Magic Tool. This should give you dozens of new ideas that your competitors might’ve missed. Finally, don’t just stop at borrowing your competitor’s keyword ideas. You can also borrow some inspiration on: The types of graphics and images you can create to supplement your blog content. The tone and style you can use in your articles. The type of information you can include in specific content pieces. SEO Tip #8. Source a LOT of Writers Content writing is one of those professions that has a very low barrier to entry. Anyone can take a writing course, claim to be a writer, and create an UpWork account… This is why 99% of the writers you’ll have to apply for your gigs are going to be, well, horrible. As such, if you want to produce a lot of content on the reg, you’ll need to source a LOT of writers. Let’s do the math: If, by posting a job ad, you source 100 writers, you’ll see that only 5 of them are a good fit. Out of the 5 writers, 1 has a very high rate, so they drop out. Another doesn’t reply back to your communication, which leaves you with 3 writers. You get the 3 writers to do a trial task, and only one turns out to be a good fit for your team. Now, since the writer is freelance, the best they can do is 4 articles per month for a total of 5,000-words (which, for most niches, ain’t all that much). So, what we’re getting at here is, to hire quality writers, you should source a LOT of them. SEO Tip #9. Create a Process for Filtering Writers If you follow the previous tip, you'll end up with a huge database of hundreds of writers. This creates a whole new problem: You now have a database of 500+ writers waiting for you to sift through them and decide which ones are worth the hire. It would take you 2-3 days of intense work to go through all these writers and vet them yourself. Let’s be real - you don’t have time for that. Here’s what you can do instead: When sourcing writers, always get them to fill in a Google form (instead of DMing or emailing you). In this form, make sure to ask for 3 relevant written samples, a link to the writer’s portfolio page, and the writer’s rate per word. Create a SOP for evaluating writers. The criteria for evaluation should be: Level of English. Does the writer’s sample have any English mistakes? If so, they’re not a good fit. Quality of Samples. Are the samples long-form and engaging content or are they boring 500-word copy-pastes? Technical Knowledge. Has the writer written about a hard-to-explain topic before? Anyone can write about simple topics like traveling—you want to look for someone who knows how to research a new topic and explain it in a simple and easy-to-read way. If someone’s written about how to create a perfect cover letter, they can probably write about traveling, but the opposite isn’t true. Get your VA to evaluate the writer’s samples as per the criteria above and short-list writers that seem competent. If you sourced 500 writers, the end result of this process should be around 50 writers. You or your editor goes through the short-list of 50 writers and invites 5-10 for a (paid) trial task. The trial task is very important - you’ll sometimes find that the samples provided by the writer don’t match their writing level. SEO Tip #10. Use the Right Websites to Find Writers Not sure where to source your writers? Here are some ideas: ProBlogger \- Our #1 choice - a lot of quality writers frequent this website. LinkedIn \- You can headhunt content writers in specific locations. Upwork \- If you post a content gig, most writers are going to be awful. Instead, I recommend headhunting top writers instead. WeWorkRemotely \- Good if you’re looking to make a full-time remote hire. Facebook \- There are a ton of quality Facebook groups for writers. Some of our faves are Cult of Copy Job Board and Content Marketing Lounge. SEO Tip #11. Always Use Content Outlines When giving tasks to your writing team, you need to be very specific about the instructions you give them. Don’t just provide a keyword and tell them to “knock themselves out.” The writer isn’t a SEO expert; chances are, they’re going to mess it up big-time and talk about topics that aren’t related to the keyword you’re targeting. Instead, when giving tasks to writers, do it through content outlines. A content outline, in a nutshell, is a skeleton of the article they’re supposed to write. It includes information on: Target word count (aim for the same or 50% more the word count than that of the competition). Article title. Article structure (which sections should be mentioned and in what order). Related topics of keywords that need to be mentioned in the article. Content outline example in the URL in the post intro. SEO Tip #12. Focus on One Niche at a Time I used to work with this one client that had a SaaS consisting of a mixture of CRM, Accounting Software, and HRS. I had to pick whether we were going to focus on topics for one of these 3 niches or focus on all of them at the same time. I decided to do the former. Here’s why: When evaluating what to rank, Google considers the authority of your website. If you have 60 articles about accounting (most of which link to each other), you’re probably an authority in the niche and are more likely to get good rankings. If you have 20 sales, 20 HR, and 20 accounting articles, though, none of these categories are going to rank as well. It always makes more sense to first focus on a single niche (the one that generates the best ROI for your business), and then move on to the rest. This also makes it easier to hire writers - you hire writers specialized in accounting, instead of having to find writers who can pull off 3 unrelated topics. SEO Tip #13. Just Hire a VA Already It’s 2021 already guys—unless you have a virtual assistant, you’re missing out big-time. Since a lot of SEO tasks are very time-consuming, it really helps to have a VA around to take over. As long as you have solid SOPs in place, you can hire a virtual assistant, train them, and use them to free up your time. Some SEO tasks virtual assistants can help with are: Internal linking. Going through all your blog content and ensuring that they link to each other. Backlink prospecting. Going through hundreds of websites daily to find link opportunities. Uploading content on WordPress and ensuring that the content is optimized well for on-page SEO. SEO Tip #14. Use WordPress (And Make Your Life Easier) Not sure which CMS platform to use? 99% of the time, you’re better off with WordPress. It has a TON of plugins that will make your life easier. Want a drag & drop builder? Use Elementor. It’s cheap, efficient, extremely easy to learn, and comes jam-packed with different plugins and features. Wix, SiteGround, and similar drag & drops are pure meh. SEO Tip #15. Use These Nifty WordPress Plugins There are a lot of really cool WordPress plugins that can make your (SEO) life so much easier. Some of our favorites include: RankMath. A more slick alternative to YoastSEO. Useful for on-page SEO. Smush. App that helps you losslessly compress all images on your website, as well as enables lazy loading. WP Rocket. This plugin helps speed up your website pretty significantly. Elementor. Not a techie? This drag & drop plugin makes it significantly easier to manage your website. WP Forms. Very simple form builder. Akismet Spam Protection. Probably the most popular anti-spam WP plugin. Mammoth Docx. A plugin that uploads your content from a Google doc directly to WordPress. SEO Tip #16. No, Voice Search Is Still Not Relevant Voice search is not and will not be relevant (no matter what sensationalist articles might say). Sure, it does have its application (“Alexa, order me toilet paper please”), but it’s pretty niche and not relevant to most SEOs. After all, you wouldn’t use voice search for bigger purchases (“Alexa, order me a new laptop please”) or informational queries (“Alexa, teach me how to do accounting, thanks”). SEO Tip #17. SEO Is Obviously Not Dead I see these articles every year - “SEO is dead because I failed to make it work.” SEO is not dead and as long as there are people looking up for information/things online, it never will be. And no, SEO is not just for large corporations with huge budgets, either. Some niches are hypercompetitive and require a huge link-building budget (CBD, fitness, VPN, etc.), but they’re more of an exception instead of the rule. SEO Tip #18. Doing Local SEO? Focus on Service Pages If you’re doing local SEO, you’re better off focusing on local service pages than blog content. E.g. if you’re an accounting firm based in Boston, you can make a landing page about /accounting-firm-boston/, /tax-accounting-boston/, /cpa-boston/, and so on. Or alternatively, if you’re a personal injury law firm, you’d want to create pages like /car-accident-law-firm/, /truck-accident-law-firm/, /wrongful-death-law-firm/, and the like. Thing is, you don’t really need to rank on global search terms—you just won’t get leads from there. Even if you ranked on the term “financial accounting,” it wouldn’t really matter for your bottom line that much. SEO Tip #19. Engage With the SEO Community The SEO community is (for the most part) composed of extremely helpful and friendly people. There are a lot of online communities (including this sub) where you can ask for help, tips, case studies, and so on. Some of our faves are: This sub :) SEO Signals Lab (FB Group) Fat Graph Content Ops (FB Group) Proper SEO Group (FB Group) BigSEO Subreddit SEO Tip #20. Test Keywords Before Pursuing Them You can use Google ads to test how profitable any given keyword is before you start trying to rank for it. The process here is: Create a Google Ads account. Pick a keyword you want to test. Create a landing page that corresponds to the search intent behind the keyword. Allocate an appropriate budget. E.g. if you assume a conversion rate of 2%, you’d want to buy 100+ clicks. If the CPC is 2 USD, then the right budget would be 200 USD plus. Run the ads! If you don’t have the budget for this, you can still use the average CPC for the keyword to estimate how well it’s going to convert. If someone is willing to bid 10 USD to rank for a certain keyword, it means that the keyword is most probably generating pretty good revenue/conversions. SEO Tip #21. Test & Improve SEO Headlines Sometimes, you’ll see that you’re ranking in the top 3 positions for your search query, but you’re still not driving that much traffic. “What’s the deal?” you might be asking. Chances are, your headline is not clickable enough. Every 3-4 months, go through your Google Search Console and check for articles that are ranking well but not driving enough traffic. Then, create a Google sheet and include the following data: Targeted keyword Page link CTR (for the last 28 days) Date when you implemented the new title Old title New title New CTR (for the month after the CTR change was implemented) From then on, implement the new headline and track changes in the CTR. If you don’t reach your desired result, you can always test another headline. SEO Tip #22. Longer Content Isn’t Always Better Content You’ve probably heard that long-form content is where it’s at in 2021. Well, this isn’t always the case. Rather, this mostly depends on the keyword you’re targeting. If, for example, you’re targeting the keyword “how to tie a tie,” you don’t need a long-ass 5,000-word mega-guide. In such a case, the reader is looking for something that can be explained in 200-300 words and if your article fails to do this, the reader will bounce off and open a different page. On the other hand, if you’re targeting the keyword “how to write a CV,” you’ll need around 4,000 to 5,000 words to adequately explain the topic and, chances are, you won’t rank with less. SEO Tip #23. SEO is Not All About Written Content More often than not, when people talk about SEO they talk about written blog content creation. It’s very important not to forget, though, that blog content is not end-all-be-all for SEO. Certain keywords do significantly better with video content. For example, if the keyword is “how to do a deadlift,” video content is going to perform significantly better than blog content. Or, if the keyword is “CV template,” you’ll see that a big chunk of the rankings are images of the templates. So, the lesson here is, don’t laser-focus on written content—keep other content mediums in mind, too. SEO Tip #24. Write For Your Audience It’s very important that your content resonates well with your target audience. If, for example, you’re covering the keyword “skateboard tricks,” you can be very casual with your language. Heck, it’s even encouraged! Your readers are Googling the keyword in their free time and are most likely teens or in their early 20s. Meaning, you can use informal language, include pop culture references, and avoid complicated language. Now, on the other hand, if you’re writing about high-level investment advice, your audience probably consists of 40-something suit-and-ties. If you include Rick & Morty references in your article, you'll most likely lose credibility and the Googler, who will go to another website. Some of our best tips on writing for your audience include: Define your audience. Who’s the person you’re writing for? Are they reading the content at work or in their free time? Keep your reader’s level of knowledge in mind. If you’re covering an accounting 101 topic, you want to cover the topic’s basics, as the reader is probably a student. If you’re writing about high-level finance, though, you don’t have to teach the reader what a balance sheet is. More often than not, avoid complicated language. The best practice is to write on a 6th-grade level, as it’s understandable for anyone. Plus, no one wants to read Shakespeare when Googling info online (unless they’re looking for Shakespeare's work, of course). SEO Tip #25. Create Compelling Headlines Want to drive clicks to your articles? You’ll need compelling headlines. Compare the following headline: 101 Productivity Tips \[To Get Things Done in 2021\] With this one: Productivity Tips Guide Which one would you click? Data says it’s the first! To create clickable headlines, I recommend you include the following elements: Keyword. This one’s non-negotiable - you need to include the target keyword in the headline. Numbers. If Buzzfeed taught us anything, it’s that people like to click articles with numbers in their titles. Results. If I read your article, what’s going to be the end result? E.g. “X Resume tips (to land the job)”.* Year (If Relevant). Adding a year to your title shows that the article is recent (which is relevant for some specific topics). E.g. If the keyword is “Marketing Trends,” I want to know marketing trends in 2021, not in 2001. So, adding a year in the title makes the headline more clickable. SEO Tip #26. Make Your Content Visual How good your content looks matters, especially if you're in a competitive niche. Here are some tips on how to make your content as visual as possible: Aim for 2-4 sentences per paragraph. Avoid huge blocks of text. Apply a 60-65% content width to your blog pages. Pick a good-looking font. I’d recommend Montserrat, PT Sans, and Roboto. Alternatively, you can also check out your favorite blogs, see which fonts they’re using, and do the same. Use a reasonable font size. Most top blogs use font sizes ranging from 16 pt to 22 pt. Add images when possible. Avoid stock photos, though. No one wants to see random “office people smiling” scattered around your blog posts. Use content boxes to help convey information better. Content boxes example in the URL in the intro of the post. SEO Tip #27. Ditch the Skyscraper Technique Already Brian Dean’s skyscraper technique is awesome and all, but the following bit really got old: “Hey \[name\], I saw you wrote an article. I, too, wrote an article. Please link to you?” The theory here is, if your content is good, the person will be compelled to link to it. In practice, though, the person really, really doesn’t care. At the end of the day, there’s no real incentive for the person to link to your content. They have to take time out of their day to head over to their website, log in to WordPress, find the article you mentioned, and add a link... Just because some stranger on the internet asked them to. Here’s something that works much better: Instead of fake compliments, be very straightforward about what you can offer them in exchange for that link. Some things you can offer are: A free version of your SaaS. Free product delivered to their doorstep. Backlink exchange. A free backlink from your other website. Sharing their content to your social media following. Money. SEO Tip #28. Get the URL Slug Right for Seasonal Content If you want to rank on a seasonal keyword, there are 2 ways to do this. If you want your article to be evergreen (i.e. you update it every year with new information), then your URL should not contain the year. E.g. your URL would be /saas-trends/, and you simply update the article’s contents+headline each year to keep it timely. If you’re planning on publishing a new trends report annually, though, then you can add a year to the URL. E.g. /saas-trends-2020/ instead of /saas-trends/. SEO Tip #29. AI Content Tools Are a Mixed Bag Lots of people are talking about AI content tools these days. Usually, they’re either saying: “AI content tools are garbage and the output is horrible,” Or: “AI content tools are a game-changer!” So which one is it? The truth is somewhere in-between. In 2021, AI content writing tools are pretty bad. The output you’re going to get is far from something you can publish on your website. That said, some SEOs use such tools to get a very, very rough draft of the article written, and then they do intense surgery on it to make it usable. Should you use AI content writing tools? If you ask me, no - it’s easier to hire a proficient content writer than spend hours salvaging AI-written content. That said, I do believe that such tools are going to get much better years down the line. This one was, clearly, more of a personal opinion than a fact. I’d love to hear YOUR opinion on AI content tools! Are they a fad, or are they the future of content creation? Let me know in the comments. SEO Tip #30. Don’t Overdo it With SEO Tools There are a lot of SEO tools out there for pretty much any SEO function. Keyword research, link-building, on-page, outreach, technical SEO, you name it! If you were to buy most of these tools for your business, you’d easily spend 4-figures on SEO tools per month. Luckily, though, you don’t actually need most of them. At the end of the day, the only must-have SEO tools are: An SEO Suite (Paid). Basically SEMrush or Ahrefs. Both of these tools offer an insane number of features - backlink analysis, keyword research, and a ton of other stuff. Yes, 99 USD a month is expensive for a tool. But then again, if you value your time 20 USD/hour and this tool saves you 6 hours, it's obviously worth it, right? On-Page SEO Tool (Free). RankMath or Yoast. Basically, a tool that's going to help you optimize web pages or blog posts as per SEO best practices. Technical SEO Tool (Freemium). You can use ScreamingFrog to crawl your entire website and find technical SEO problems. There are probably other tools that also do this, but ScreamingFrog is the most popular option. The freemium version of the tool only crawls a limited number of pages (500 URLs, to be exact), so if your website is relatively big, you'll need to pay for the tool. Analytics (Free). Obviously, you'll need Google Analytics (to track website traffic) and Google Search Console (to track organic traffic, specifically) set up on your website. Optionally, you can also use Google Track Manager to better track how your website visitors interact with the site. MozBar (Free). Chrome toolbar that lets you simply track the number of backlinks on Google Search Queries, Domain Authority, and a bunch of other stuff. Website Speed Analysis (Free). You can use Google Page Speed Insights to track how fast your website loads, as well as how mobile-friendly it is. Outreach Tool (Paid). Tool for reaching out to prospects for link-building, guest posting, etc. There are about a dozen good options for this. Personally, I like to use Snov for this. Optimized GMB Profile (Free). Not a tool per se, but if you're a local business, you need to have a well-optimized Google My Business profile. Google Keyword Planner (Free). This gives you the most reliable search volume data of all the tools. So, when doing keyword research, grab the search volume from here. Tool for Storing Keyword Research (Free). You can use Google Sheets or AirTable to store your keyword research and, at the same time, use it as a content calendar. Hemingway App (Free). Helps keep your SEO content easy to read. Spots passive voice, complicated words, etc. Email Finder (Freemium). You can use a tool like Hunter to find the email address of basically anyone on the internet (for link-building or guest posting purposes). Most of the tools that don’t fit into these categories are 100% optional. SEO Tip #31. Hiring an SEO? Here’s How to Vet Them Unless you’re an SEO pro yourself, hiring one is going to be far from easy. There’s a reason there are so many “SEO experts” out there - for the layman, it’s very hard to differentiate between someone who knows their salt and a newbie who took an SEO course, like, last week. Here’s how you can vet both freelance and full-time SEOs: Ask for concrete traffic numbers. The SEO pro should give you the exact numbers on how they’ve grown a website in the past - “100% SEO growth in 1 year” doesn’t mean much if the growth is from 10 monthly traffic to 20. “1,000 to 30,000” traffic, on the other hand, is much better. Ask for client names. While some clients ask their SEOs to sign an NDA and not disclose their collaboration, most don’t. If an SEO can’t name a single client they’ve worked with in the past, that’s a red flag. Make sure they have the right experience. Global and local SEO have very different processes. Make sure that the SEO has experience with the type of SEO you need. Make sure you’re looking for the right candidate. SEO pros can be content writers, link-builders, web developers, or all of the above simultaneously. Make sure you understand which one you need before making the hire. If you’re looking for someone to oversee your content ops, you shouldn’t hire a technical SEO expert. Look for SEO pros in the right places. Conventional job boards are overrated. Post your job ads on SEO communities instead. E.g. this sub, bigseo, SEO Signals Facebook group, etc. SEO Tip #32. Blog Post Not Ranking? Follow This Checklist I wanted to format the post natively for Reddit, but it’s just SO much better on Notion. Tl;dr, the checklist covers every reason your post might not be ranking: Search intent mismatch. Inferior content. Lack of internal linking. Lack of backlinks. And the like. Checklist URL at the intro of the post. SEO Tip #33. Avoid BS Link-Building Tactics The only type of link-building that works is building proper, quality links from websites with a good backlink profile and decent organic traffic. Here’s what DOESN’T work: Blog comment links Forum spam links Drive-by Reddit comment/post links Web 2.0 links Fiverr “100 links for 10 bucks” bs If your “SEO agency” says they’re doing any of the above instead of actually trying to build you links from quality websites, you’re being scammed. SEO Tip #34. Know When to Use 301 and 302 Redirects When doing redirects, it’s very important to know the distinction between these two. 301 is a permanent page redirect and passes on link juice. If you’re killing off a page that has backlinks, it’s better to 301 it to your homepage so that you don’t lose the link juice. If you simply delete a page, it’s going to be a 404, and the backlink juice is lost forever. 302 is a temporary page redirect and doesn’t pass on link juice. If the redirect is temporary, you do a 302. E.g. you want to test how well a new page is going to perform w/ your audience. SEO Tip #35. Social Signals Matter (But Not How You Think) Social signals are NOT a ranking factor. And yet, they can help your content rank on Google’s front page. Wondering what the hell am I talking about? Here’s what’s up: As I said, social signals are not a ranking factor. It’s not something Google takes into consideration to decide whether your article should rank or not. That said, social signals CAN lead to your article ranking better. Let’s say your article goes viral and gets around 20k views within a week. A chunk of these viewers are going to forget your domain/link and they’re going to look up the topic on Google via your chosen keyword + your brand name. The amount of people looking for YOUR keyword and exclusively picking your result over others is going to make Google think that your content is satisfying search intent better than the rest, and thus, reward you with better ranking. SEO Tip #36. Run Remarketing Ads to Lift Organic Traffic Conversions Not satisfied with your conversion rates? You can use Facebook ads to help increase them. Facebook allows you to do something called “remarketing.” This means you can target anyone that visited a certain page (or multiple pages) on your website and serve them ads on Facebook. There are a TON of ways you can take advantage of this. For example, you can target anyone that landed on a high buyer intent page and serve them ads pitching your product or a special offer. Alternatively, you can target people who landed on an educational blog post and offer them something to drive them down the funnel. E.g. free e-book or white paper to teach them more about your product or service. SEO Tip #37. Doing Local SEO? Follow These Tips Local SEO is significantly different from global SEO. Here’s how the two differ (and what you need to do to drive local SEO results): You don’t need to publish content. For 95% of local businesses, you only want to rank for keywords related to your services/products, you don’t actually need to create educational content. You need to focus more on reviews and citation-building. One of Google Maps’ biggest ranking factors is the of reviews your business has. Encourage your customers to leave a review if they enjoyed your product/service through email or real-life communication. You need to create service pages for each location. As a local business, your #1 priority is to rank for keywords around your service. E.g. If you're a personal injury law firm, you want to optimize your homepage for “personal injury law firm” and then create separate pages for each service you provide, e.g. “car accident lawyer,” “motorcycle injury law firm,” etc. Focus on building citations. Being listed on business directories makes your business more trustworthy for Google. BrightLocal is a good service for this. You don’t need to focus as much on link-building. As local SEO is less competitive than global, you don’t have to focus nearly as much on building links. You can, in a lot of cases, rank with the right service pages and citations. SEO Tip #38. Stop Ignoring the Outreach Emails You’re Getting (And Use Them to Build Your Own Links) Got a ton of people emailing you asking for links? You might be tempted to just send them all straight to spam, and I don’t blame you. Outreach messages like “Hey Dr Jigsaw, your article is A+++ amazing! ...can I get a backlink?” can get hella annoying. That said, there IS a better way to deal with these emails: Reply and ask for a link back. Most of the time, people who send such outreach emails are also doing heavy guest posting. So, you can ask for a backlink from a 3rd-party website in exchange for you mentioning their link in your article. Win-win! SEO Tip #39. Doing Internal Linking for a Large Website? This’ll Help Internal linking can get super grueling once you have hundreds of articles on your website. Want to make the process easier? Do this: Pick an article you want to interlink on your website. For the sake of the example, let’s say it’s about “business process improvement.” Go on Google and look up variations of this keyword mentioned on your website. For example: Site:\[yourwebsite\] “improve business process” Site:\[yourwebsite\] “improve process” Site:\[yourwebsite\] “process improvement” The above queries will find you the EXACT articles where these keywords are mentioned. Then, all you have to do is go through them and include the links. SEO Tip #40. Got a Competitor Copying Your Content? File a DMCA Notice Fun fact - if your competitors are copying your website, you can file a DMCA notice with Google. That said, keep in mind that there are consequences for filing a fake notice.

 I just sold my startup for $200,000 after 11 months. AMA
reddit
LLM Vibe Score0
Human Vibe Score0
jeannenThis week

I just sold my startup for $200,000 after 11 months. AMA

Last August, I was looking for a startup idea I could grow and made a MVP in a week then launched it. I received the $200,000 wire from the buyer a couple of days ago I found tons of useful info online for free, so I hope this can be my way of giving back :) Here is some background: Idea I got the idea when trying to write a tweet using Google Doc's transcription tool, which was terrible. I was pretty sure I wasn't the only one too lazy to type, I made my own solution using AI to transcribe and reformat voice notes into any kind of content. I called it Talknotes, mainly because it was the only domain available lol Validation: My rule is to only reinvest what the project generates. After listing on startup directories and posting on Twitter, I generated $700 in 10 days. It wasn't much, but enough to show interest and keep me motivated. I added user-requested features, but the launch effect wore off, and daily revenues dropped to $0 after a few weeks. I almost gave up, but friends encouraged me to continue. In October, I launched on ProductHunt and it blew up. It became Product of the Day and reached $1500 MRR thanks to media coverage. I initially built everything using vanilla JS/CSS/HTML + Node for backend. But it's pretty limited for apps with lots of interactivity so, I rebuilt the app using Nuxt.js to make it easier to ship new features. Then, I launched ads on Facebook and I implemented a feedback loop: Get new users Learn about them through onboarding Make more ads based on onboarding data This doubled MRR in about 2 months. Burnout and Sale: In May, I had a bad burnout after emergency bug fixes. This made it hard to work on the app after. At this point MRR was around $7000 and total revenues around $70,0000 I listed it on Acquire.com for $200,000, a very good price for the buyer considering revenues and growth. I could've gotten $300,000 with buyer financing or earn-outs, but I wanted cash, $200,000 today is better than $300,000 in a year. Everything was smooth until we tried using Escrow, which almost fucked up the deal (details here). Long story short, had to threaten them to make a sponsored post on Twitter explaining what they did + legal action. They sent the refund the very next day, and we completed the transfer directly. Now, this isn't an overnight success. It's the result of 7 years of grind. I launched over 40 projects since I started, and most of them failed. I often worked 100 hours per week, and I rarely go out or meet many people. It's not for everyone, but I'm fine with it With the profit from the app + sale, and other projects, I have close to 1/3 of a million dollar. I could retire in Asia if I wanted Just mind blowing to think I wrote funny characters in a code editor and sold it for the price of a house lol Edit 1: A few people got confused. I said it's 7 years of grind and most of my projects failed, not that I was not making money. I also said I OFTEN worked 100h/week, not every week :) Since I learned to code 2 years ago I've made close to $400k from my app's profit + exit (this one + another one for $65k last year). And before that I was making money as a marketing freelancer. Also, I dropped after high-school, so, I had to learn everything from scratch, it takes time! Edit 2: Lots of people asked how/where I learned to code in 2 months. I wrote a blog/journal about it back then with links to resources, you can find it here if you're interested

10 Side Projects in 10 Years: Lessons from Failures and a $700 Exit
reddit
LLM Vibe Score0
Human Vibe Score1
TheValueProviderThis week

10 Side Projects in 10 Years: Lessons from Failures and a $700 Exit

Hey folks, I'm sharing my journey so far in case it can help others. Entrepreneurship can sometimes be demotivating. In my case, I've always been involved in side projects and what I've realized is that every time you crash a project, the next one makes it a bit further. So this is a long-term game and consistency ends up paying off The $1 Android Game (2015, age 18) What Happened: 500 downloads, 1€ in ad revenue Ugly UI, performance issues Key Lessons: Don’t be afraid of launching. Delaying for “perfection” is often a sign that you fear being ignored. I was trying to perfect every aspect of the game. In reality, I was delaying the launch because I feared no one would download the app. Commit to the project or kill it. At some point, this project was no longer fun (it was just about fixing device responsiveness). Most importantly, I wasn't learning anything new so I moved to smth else. The Forex Bot Regret (2016, age 19) What Happened: Lost months identifying inexistent chart patterns Created a Trading bot that was never profitable Key Lessons: Day trading’s real winners are usually brokers. There are plenty of guys selling a bot or systems that are not making money trading, why would they sell a “money-printing machine” otherwise... Develop an unfair advantage. With these projects, I developed a strong coding foundation that gave me an edge when dealing with non-technical business people. Invest countless hours to create a skills gap between you and others, one that becomes increasingly difficult for them to close (coding, public speaking, networking, etc.) The $700 Instagram Exit (2018, age 21) What Happened: Grew a motivational account to 60k followers Sold it for $700 90% of followers were in low-income countries (hard to monetize) Key Lessons: Follower quality > quantity. I focused on growth and ended up with an audience I couldn’t truly define. If brands don’t see value, you won’t generate revenue. Also, if you do not know who you are creating content for, you'll end up demotivated and stop posting. Great 3rd party product + domain authority = Affiliate marketing works. In this case, I could easily promote an IG growing service because my 50k+ followers conveyed trust. Most importantly, the service I was promoting worked amazingly. The Illegal Amazon Review Marketplace (2020, age 23) What Happened: Sellers were reimbursing buyers for positive reviews Built a WordPress marketplace to facilitate “free products for reviews” Realized it violated Amazon’s terms Key Lessons: Check for “red flags” when doing idea assessment. There will always be red and orange flags. It’s about learning to differentiate between them (e.g. illegality, 100% dependence on a platform, etc.) If there’s competition, it’s good, if they are making money it’s even better. I was thrilled when I saw no competition for my “unique idea”. Later, I discovered the obvious reason. Copying a “Proven” Business Model (2020, age 23) What Happened: Tried recreating an Instagram “comment for comment” growth tool Instagram changed the algorithm and killed the growth strategy that the product used. Key Lessons: Do not build a business that depends 100% on another business, it is too risky. Mr. Musk can increase Twitter on API pricing to $42,000 monthly without notice and Tik Tok can be banned in the US. Due to the IG algorithm change, we had built a product that was not useful, and worse, now we had no idea how to grow an IG account. Consider future project synergies before selling. I regret having sold the 60k follower IG account since it could have saved me a lot of time when convincing users to try the service. NFT Marathon Medals (2021, age 24) What Happened: Created NFT race medals Sold 20 for 5€ each, but spent 95% of meetings explaining “what is an NFT?” Key Lessons: Market timing is crucial. As with every new technology, it is only useful as long as society is ready to adopt it. No matter how promising the tech is in the eyes of SV, society will end up dictating its success (blockchain, AI, etc). In this case, the runner community was not ready to adopt blockchain (it is not even prepared today). Race organizers did not know what they were selling, and runners did not know what they were buying. The 30-day rule in Fanatical Prospecting. Do not stop prospecting. I did prospecting and closed deals 3 months after the outbound efforts. Then I was busy executing the projects and had no clients once the projects were finished. AI Portal & Co-Founder Misalignment (2023, age 26) What Happened: Built a portal for SMEs to find AI use cases Co-founders disagreed on vision and execution Platform still gets \~1 new user/day Key Lessons: Define roles and equity clearly. Our biggest strength ended up killing us. Both founders had strong strategic skills and we were constantly arguing about decisions. NextJS + Vercel + Supabase: Great stack to create a SaaS MVP. (but do not use AI with frameworks unless you know how they work conceptually) SEO is king. One of our users creates a use case on “Changing Song Lyrics with AI.” Not being our target use case, it brings 90% of our traffic. Building an AI Tool & Getting Ghosted (2024, age 27) What Happened: SEO agency wanted to automate rewriting product descriptions Built it in 3 weeks, but the client vanished Key Lessons: Validate manually first. Don’t code a full-blown solution for a problem you haven’t tested in real-world workflows. I kept rewriting code only to throw it away. Jumping straight into building a solution ended up costing more time than it saved. Use templates, no-code, and open-source for prototyping. In my case, using a Next.js template saved me about four weeks of development only to hit the same dead end, but much faster. Fall in love with your ICP or walk away. I realized I didn’t enjoy working with SEO agencies. Looking back, I should have been honest with myself and admitted that I wasn’t motivated enough by this type of customer. Ignoring Code Perfection Doubled Traffic (2025, age 28) What Happened: Partnered with an ex-colleague to build an AI agents directory Focused on content & marketing, not endless bug fixes Traffic soared organically Key Lessons: Measure the impact of your actions and double down on what works. We set up an analytics system with PostHog and found wild imbalances (e.g. 1 post about frameworks outperformed 20 promotional posts). You have to start somewhere. For us, the AI agents directory is much more than just a standalone site, it's a strategic project that will allow us to discover new products, gain domain authority, and boost other projects. It builds the path for bigger opportunities. Less coding, more traction. Every day I have to fight against myself not to code “indispensable features”. Surprisingly, the directory keeps gaining consistent traffic despite being far from perfect Quitting My Job & Looking Ahead (2025, age 28) What Happened: Left full-time work to go all-in Plan to build vertical AI agents that handle entire business workflows (support, marketing, sales) Key Lessons: Bet on yourself. The opportunity cost of staying in my full-time job outweighed the benefits. It might be your case too I hope this post helps anyone struggling with their project and inspires those considering quitting their full-time job to take the leap with confidence.

Seeking Feedback: Would a No-Code AI Solution Benefit Your Business?
reddit
LLM Vibe Score0
Human Vibe Score0
chrisparkerofficialThis week

Seeking Feedback: Would a No-Code AI Solution Benefit Your Business?

I'm currently working on an AI startup, with the goal of providing small-medium businesses a seamless and intuitive way to integrate AI into operations without the need for any coding or tech expertise. We're designing an auto machine learning application that's user-friendly and tailored to the unique needs of small businesses. Before we scale, I would really appreciate any insights and feedback. Here are a few questions that would be helpful to get answers to: Pain Points: Are there specific tasks or processes in your operations that you think could be automated or enhanced using AI? This could be anything from customer service chatbots, inventory management, sales forecasting, or anything else you might think of. Features: What features would you want in a no-code AI solution? Perhaps easy integration with existing software? Drag-and-drop model training? Pre-built models for common tasks? Training & Support: How important would training and support be for you in implementing and using an AI solution? Would you prefer video tutorials, live-chat support, or hands-on workshops? Pricing: Would you be willing to invest in such a tool? If so, what would be a reasonable price point for you? We're considering a tiered model based on usage, with a potential starting point of $X/month. Does that sound feasible? Trial Period: Would a free trial period be beneficial for you? How long would you need to assess the tool's impact on your business? Data Concerns: How comfortable are you with sharing data with an AI application? What privacy and security measures would make you feel at ease? Your feedback is really useful. We're building this solution with you in mind, and your insights will guide the next steps. In appreciation for your time and input, we're offering a special discount for early adopters from this community once we launch. Just drop a comment below, and I'll make sure to get in touch when we are ready. Many Thanks, Chris Parker

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰
reddit
LLM Vibe Score0
Human Vibe Score1
benfromwhereThis week

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰

(Monthly income breakdown is in the end) 📌 Introduction Hey everyone! 👋 Before I dive into this month’s breakdown, I just want to be upfront—English isn’t my first language, so I’ve used ChatGPT to refine this post for better readability. That said, everything here is 100% real—my personal experiences, struggles, and earnings as someone running a full-time AI influencer business. Since I get a lot of DMs asking about my AI models, here are their Instagram links: 📷 Emma – https://www.instagram.com/emmalauireal 📷 Jade – https://www.instagram.com/jadelaui (jadecasual is the second account) Also, if you’ve been wondering about the community I run, where I teach others how to build AI influencers from scratch, here’s the link (I got approval from mods for this link): 🔗 AI Winners Now, let’s get into what happened this month. 🚀 \------- First, a huge thank you! 🎉 Three months ago, I shared my journey of building an AI influencer business, and I was blown away by the response. That post got 263K+ views and was shared over 2.7K times—way more than I ever expected. If you’re new here or want to check out the full story of how I started, you can read it here: 🔗 Click Here (Reddit link) \------- 🔹 What I Did in January After the holiday rush in December, I knew January would be a slow month—people had already spent most of their money at the end of the year. So instead of pushing harder on monetization, I shifted my focus to tech development and optimization. Flux Character Loras: I spent a lot of time refining and testing different Flux-based character Loras for my models. This is still a work in progress, but the goal is to improve long-term consistency and make my workflow even more efficient. NSFW Content Expansion: On Emma’s side, I expanded her content library using a real model body double, making her content look more organic and natural. Jade, however, remains 100% AI-generated, keeping her workflow entirely digital. Social Media Wipeout (Thanks, VA 🙃): I had handed off both Twitter accounts to a virtual assistant to help with engagement and DMs. Big mistake. He ended up spamming DMs, which got both accounts banned—Emma (80K followers) and Jade (20K followers). 🤦‍♂️ Right now, I’m rebuilding Emma’s account from scratch and taking a much more cautious approach. Jade’s account is still offline for now. New Platform: Threads – I hadn’t touched Threads before, but since engagement on Instagram can be unpredictable, I decided to start accounts for both models. So far, they’re performing well, and I’ll continue experimenting. Launched AI Winners Community: After getting flooded with DMs (both here and on Instagram), I realized there was a massive demand for structured learning around AI influencers. So, I launched AI Winners, a paid community where I break down everything I’ve learned. It’s still early, but I see it turning into a solid, long-term community. Investment & Acquisition Talks: I’m still evaluating potential investors and acquisition offers for my AI models. There’s growing interest in buying or investing in Emma & Jade, so I’ve been having conversations to explore different options. Overall, January was about tech, rebuilding, and long-term planning—not immediate revenue. But that’s what keeps this business sustainable. 🚀 \------- ⚠️ Biggest Challenges This Month Lost Both Twitter Accounts (Massive Traffic Hit) 🚨 The biggest blow this month was losing my models’ Twitter accounts. Twitter was responsible for about 40% of my total traffic, meaning both free and paid subs took a direct hit. While Emma’s revenue took a slight dip, Jade’s income dropped significantly—partly due to the account loss and partly because January is naturally slow. (Full revenue breakdown at the end of the post.) Jade’s Instagram Tanked (Possible Shadow Ban?) 🤔 Jade’s Instagram completely lost momentum in early January. Engagement and reach dropped by over 80%, and I still haven’t figured out why. It feels like a shadow ban, but I have no clear confirmation. To counter this, I launched a second backup account, and things are starting to recover. \------- 🚀 Potential Improvements & What’s Next Locking in a Stable Workflow 🔄 Right now, Emma & Jade’s workflow is still evolving, but I’m aiming to fully stabilize it. As I’m writing this, content is generating on my second monitor—a sign that I’m close to achieving full automation without compromising quality. Boosting Jade’s Fanvue Revenue 💰 Jade’s income took a hit this month, and it’s 100% a traffic issue. The solution? More content, more reach. I’ll be increasing social media output to drive consistent traffic back to Fanvue and restore her earnings. Patreon is Done. All Focus on Fanvue 🚫 I shut down both Emma & Jade’s Patreon accounts. The goal is not to split revenue—I want everything funneled into Fanvue for higher engagement and bigger paydays. \------- 💰 January 2025 Earnings Breakdown Despite January being one of the slowest months for online creators, Emma and Jade still brought in over $29K in revenue, with a net profit exceeding $20K after all expenses. Emma Laui generated $20,206.77, with around $6,000 in expenses (chatter payments, NSFW designer fees, and other operational costs). Jade Laui earned $8,939.05, with $2,000 in expenses. Considering Twitter account losses, Instagram setbacks, and the usual January spending slump, this is still a solid outcome. The focus now is on scaling traffic and maximizing Fanvue revenue heading into February. 🚀🔥 That’s the full breakdown for January! If you have questions, feel free to drop a comment, and I’ll answer when I can. Happy to help, just like others helped me when I was starting out! 🚀🔥

Seeking Feedback on Business Idea: AI-Powered Business Partner Matching Platform
reddit
LLM Vibe Score0
Human Vibe Score1
torrentialdownpour34This week

Seeking Feedback on Business Idea: AI-Powered Business Partner Matching Platform

Hey everyone, I've been toying with an idea for a new business venture and I'd love to get some feedback and insights from this community. The Idea: I'm considering building a business platform that utilizes AI to match potential business partners. Whether you're a startup looking for a co-founder, a company seeking strategic partnerships, or an investor searching for promising ventures, this platform would help connect you with compatible partners based on your specific needs, goals, and preferences. How It Works: Users would create detailed profiles outlining their business objectives, industry expertise, skills, and what they're looking for in a partner. The AI algorithm would then analyze this data to identify compatible matches, taking into account factors like complementary skill sets, shared values, and mutual goals. The platform would provide users with a curated list of potential partners, along with insights and recommendations to facilitate meaningful connections. Key Features: Comprehensive Profiles: Users can create detailed profiles highlighting their background, experience, and what they bring to the table. AI Matching Algorithm: The platform's AI algorithm would use advanced data analysis techniques to generate accurate partner recommendations. Communication Tools: Built-in messaging and video conferencing tools would enable seamless communication between potential partners. Feedback and Ratings: Users can leave feedback and ratings for their matches, helping to build trust and credibility within the community. Resource Hub: Access to resources, articles, and guides on partnership development, negotiation strategies, and other relevant topics. Why It's Needed: Finding the right business partner can be a daunting task, often relying on personal networks or serendipitous encounters. By harnessing the power of AI, this platform aims to streamline the partner matching process, saving time and increasing the likelihood of finding compatible collaborators. Looking for Feedback: Before diving headfirst into this venture, I wanted to reach out to this community to gather some feedback: Does this idea resonate with you? Why or why not? Are there any existing platforms or services that offer similar functionalities? What features would be essential for you as a user? Any potential challenges or concerns you foresee with this concept? I'm eager to hear your thoughts and suggestions. Thanks in advance for your input!

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey
reddit
LLM Vibe Score0
Human Vibe Score0.778
benfromwhereThis week

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey

Update on February 22th: I changed my AI influencer's names because it caused some problems on my business. One year, two AI-powered influencers, and $250K in revenue. Sounds unreal? It’s not. Today, I’m pulling back the curtain on the strategies, tools, and hard-won lessons that took me from concept to a six-figure success story in the AI influencer space. Hey, I'm Ben—a 32-year-old designer who spent the past year navigating the world of AI influencers. Let me clear up any confusion right from the start: I’m not here to sell you anything. This is purely a case study to share what worked, what didn’t, and what I’ve learned along the way. I’ll also make sure to answer all your questions in the comments for free whenever I can, so don’t hesitate to ask. Links to Past Topics: If you're curious about some of the groundwork I covered, check out a few of my earlier posts here: How I Make $10,000 Monthly | AI Influencer Management How I Earned $7000+ in 15 Days | AI Influencer Business Update These earlier posts cover a lot of the backstory, so feel free to explore them before diving into this one. So if you're ready, here is the full story: \---- The idea of creating an AI influencer was one of those “what if” moments that wouldn’t leave my mind. At first, it sounded futuristic—even a bit too ambitious. It all started when I stumbled upon an AI influencer on Instagram with the handle AnnaMaes2000. Her content blew me away—the quality, the detail, and just how real everything looked. I was instantly hooked and ended up going through every post, just trying to figure out how she was pulling this off. That’s when I knew I had to learn how this was done. The next step? YouTube. I dived into videos on Stable Diffusion, soaking up everything I could about creating AI-generated images. Those tutorials taught me the basics and got me up to speed. Then, I created my first AI influencer, let's call her Mel for now. Right after that, to complete the storyline and boost engagement, I introduced Mel's “mother,” Jess. Adding Jess gave the whole project depth and a narrative that drew people in, creating a unique family dynamic that instantly elevated traffic and interest. After thousands of bad photos, hundreds of deleted posts, and months of trial and error, you can now see the quality that defines my current accounts. Here’s a rundown of the tools and checkpoints I’ve used from day one, in order: Fooocus on RunDiffusion — Juggernaut V8 Fooocus on RunDiffusion — Juggernaut V9 Fooocus on PC (locally) — Juggernaut V9 Fooocus on PC (locally) —Lyuyang Mix + Juggernaut V9 Flux on PC (couple of photos only since it's so slow even on RTX 4090) Flux on Fal.ai. \---- There’s no magic Instagram hack that guarantees success, despite what everyone thinks and keeps asking me. Quality content, consistent uploads, and solid craftsmanship are what actually help your photos hit trends and show up on the Explore page. Unlike 95% of low-quality AI accounts out there, I don’t rely on faceswap videos, spam Reels, or go around liking comments on other accounts. My approach is fully organic, focused solely on creating my own unique content. By following Instagram's guidelines to the letter, I've managed to direct some of Mel and Jess' fans over to Patreon and Fanvue. There, for a small subscription fee, fans can access exclusive lingerie content. For those looking for more, higher-tier subscriptions give access to even more premium content. Some possible questions and their answers: No, you can't share hardcore NSFW content on Patreon. You can do that on Fanvue. Yes, you can create AI creators on Fanvue — OnlyFans doesn't allow it. Yes, you can use your own ID to get KYC. Yes, we're telling both Mel and Jess is (or use) AI to generate content. And yes, some people leave and some people still have fun with chatting, having a good time and get perfect content for their needs. And yes, we have a chatter team to work on these accounts. \---- This journey wasn’t all smooth sailing. I faced unexpected roadblocks, like platform restrictions that limited certain types of content, and managing fan expectations was more challenging than anticipated. Staying within guidelines while keeping fans engaged required constant adaptation. These hurdles forced me to get creative, adjust my approach, and learn fast. Once I saw Mel and Jess gaining traction, I knew it was time to scale up. Expanding meant finding new ways to keep content fresh, creating deeper narratives, and considering how to bring even more followers into the fold. My focus turned to building a sustainable model that could grow without sacrificing quality or authenticity. If you’re thinking about diving into AI content creation, here’s my advice: patience, consistency, and a focus on quality are key. Don’t cut corners or rely on quick-fix hacks. Invest time in learning the right tools, creating engaging stories, and building an audience that values what you bring to the table. This approach took me from zero to six figures, and it’s what makes the journey worth it. \---- And finally, here’s the income breakdown that everyone’s curious about: Mel on Fanvue: $82,331.58 (Gross earnings because we have chatter cuts like 15%) Mel on Patreon: $50,865.98 (Net earnings) Jess on Fanvue: $89,068.26 (Gross earnings because we have chatter cuts like 15%) Jess on Patreon: $39,040.70 And thanks to Reddit and my old posts, I got a perfect investor like after 5 months, so this is a "payback" for that. Like I said, I'll answer every question in the comments — take care and let me know.

Idea feedback: AI-native self-improvement & wellness
reddit
LLM Vibe Score0
Human Vibe Score1
thewhitelynxThis week

Idea feedback: AI-native self-improvement & wellness

Hello redditors! Thesis: We're all trying to live our best lives and many of us try to leverage technology to become better faster and easier. I’m trying to build a company that builds an AI-native solution for self-improvement. My thesis is that AI is an incredibly powerful tool for solving problems, particularly in programming and generally life - but ChatGPT isn't really designed to be your long-term 'coach'. It's great for handling specific tasks, answering questions, doing research, etc. - but it's memory and UX isn't optimized around things like behavior change, mental health support, and long-term personal life planning I believe my core problems (which I think are shared by many) are: 1) Staying motivated - it's easy to lose motivation when progress isn't immediately apparent, there are setbacks, etc. 2) Self-doubt - it makes me question myself and waste time wondering if I'm the right person to be doing this, if the idea is too broad, etc. Some of this is good - but a lot of it just makes me less effective 3) Staying on Track - I start a thing, but then gradually pivot a million different directions. This may be a touch of ADHD. I find that I'll have a long-term goal (e.g. launching a successful business), but I'll tend to wonder a lot in the process of executing over weeks and months. Staying on track just feels suprisingly difficult. I do create TODO lists and have a Kanban board I’m considering a bunch of features and have built a version focused more specifically towards mental health which implements a few: \----- • Guided Journaling Guided journaling prompts to facilitate deeper reflection • Specialist AI Coaches Personalized, expert AI coaching for your specific area of focus and goals For startup, marketing, life, fashion, whatever you want. • Goal Tracking Define, track, and achieve your goals • Behavior Change & Habit Formation Leverage the science of behavior change to help you make lasting changes in your life • Mood tracking Track and improve your mood leveraging science-backed techniques • Areas for growth Identify and develop your strengths and manage your weaknesses • Insight reports Get personalized insights into your cognitive and behavioral patterns • Inspirational Quotes Stay motivated with curated daily quotes relevant to your journey • Gamification of Growth & Mood Turn your mental health journey into a game and earn rewards for your progress \---- Would love thoughts on the idea, and feedback - and if anyone is interested in being a design partner / early user, I'd love to chat in greater depth 1:1!

Learn Coding while building your dream idea (and pay for the lessons).
reddit
LLM Vibe Score0
Human Vibe Score1
ekim2077This week

Learn Coding while building your dream idea (and pay for the lessons).

I have a business idea and would welcome feedback. As a coder with experience running a successful outsourcing shop and teaching coding skills to employees, I want to create a unique class concept where students are entrepreneurs who have an idea they want to develop and sell while learning how to code it themselves. The course would span 6–12 months, focusing on building an MVP (Minimum Viable Product) for each student's project. I would teach coding and guide students on effectively using AI tools like GPT4 and Claude to streamline the coding process by 50-75%. We would start by creating a comprehensive PRD (Product Requirements Document) and estimating the time to completion and the necessary tech stack. The class size would be limited to 4–5 students to ensure proper management and support. The monthly fee would be around $1,000 or more, considering the personalized attention and the potential for students to launch their own products by the end of the course. Students would need to commit significant effort and time (at least 6 months) to the program. Upon completion, students would be equipped with the skills to market their product, add features using AI tools, and manage other coders if needed. They would also gain a solid understanding of the time and resources required to implement new features. As an added benefit, the total cost of this program would likely be comparable to outsourcing the project development. What are your thoughts on this business idea? Do you think there is a market for this type of learning experience?

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey
reddit
LLM Vibe Score0
Human Vibe Score0.778
benfromwhereThis week

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey

Update on February 22th: I changed my AI influencer's names because it caused some problems on my business. One year, two AI-powered influencers, and $250K in revenue. Sounds unreal? It’s not. Today, I’m pulling back the curtain on the strategies, tools, and hard-won lessons that took me from concept to a six-figure success story in the AI influencer space. Hey, I'm Ben—a 32-year-old designer who spent the past year navigating the world of AI influencers. Let me clear up any confusion right from the start: I’m not here to sell you anything. This is purely a case study to share what worked, what didn’t, and what I’ve learned along the way. I’ll also make sure to answer all your questions in the comments for free whenever I can, so don’t hesitate to ask. Links to Past Topics: If you're curious about some of the groundwork I covered, check out a few of my earlier posts here: How I Make $10,000 Monthly | AI Influencer Management How I Earned $7000+ in 15 Days | AI Influencer Business Update These earlier posts cover a lot of the backstory, so feel free to explore them before diving into this one. So if you're ready, here is the full story: \---- The idea of creating an AI influencer was one of those “what if” moments that wouldn’t leave my mind. At first, it sounded futuristic—even a bit too ambitious. It all started when I stumbled upon an AI influencer on Instagram with the handle AnnaMaes2000. Her content blew me away—the quality, the detail, and just how real everything looked. I was instantly hooked and ended up going through every post, just trying to figure out how she was pulling this off. That’s when I knew I had to learn how this was done. The next step? YouTube. I dived into videos on Stable Diffusion, soaking up everything I could about creating AI-generated images. Those tutorials taught me the basics and got me up to speed. Then, I created my first AI influencer, let's call her Mel for now. Right after that, to complete the storyline and boost engagement, I introduced Mel's “mother,” Jess. Adding Jess gave the whole project depth and a narrative that drew people in, creating a unique family dynamic that instantly elevated traffic and interest. After thousands of bad photos, hundreds of deleted posts, and months of trial and error, you can now see the quality that defines my current accounts. Here’s a rundown of the tools and checkpoints I’ve used from day one, in order: Fooocus on RunDiffusion — Juggernaut V8 Fooocus on RunDiffusion — Juggernaut V9 Fooocus on PC (locally) — Juggernaut V9 Fooocus on PC (locally) —Lyuyang Mix + Juggernaut V9 Flux on PC (couple of photos only since it's so slow even on RTX 4090) Flux on Fal.ai. \---- There’s no magic Instagram hack that guarantees success, despite what everyone thinks and keeps asking me. Quality content, consistent uploads, and solid craftsmanship are what actually help your photos hit trends and show up on the Explore page. Unlike 95% of low-quality AI accounts out there, I don’t rely on faceswap videos, spam Reels, or go around liking comments on other accounts. My approach is fully organic, focused solely on creating my own unique content. By following Instagram's guidelines to the letter, I've managed to direct some of Mel and Jess' fans over to Patreon and Fanvue. There, for a small subscription fee, fans can access exclusive lingerie content. For those looking for more, higher-tier subscriptions give access to even more premium content. Some possible questions and their answers: No, you can't share hardcore NSFW content on Patreon. You can do that on Fanvue. Yes, you can create AI creators on Fanvue — OnlyFans doesn't allow it. Yes, you can use your own ID to get KYC. Yes, we're telling both Mel and Jess is (or use) AI to generate content. And yes, some people leave and some people still have fun with chatting, having a good time and get perfect content for their needs. And yes, we have a chatter team to work on these accounts. \---- This journey wasn’t all smooth sailing. I faced unexpected roadblocks, like platform restrictions that limited certain types of content, and managing fan expectations was more challenging than anticipated. Staying within guidelines while keeping fans engaged required constant adaptation. These hurdles forced me to get creative, adjust my approach, and learn fast. Once I saw Mel and Jess gaining traction, I knew it was time to scale up. Expanding meant finding new ways to keep content fresh, creating deeper narratives, and considering how to bring even more followers into the fold. My focus turned to building a sustainable model that could grow without sacrificing quality or authenticity. If you’re thinking about diving into AI content creation, here’s my advice: patience, consistency, and a focus on quality are key. Don’t cut corners or rely on quick-fix hacks. Invest time in learning the right tools, creating engaging stories, and building an audience that values what you bring to the table. This approach took me from zero to six figures, and it’s what makes the journey worth it. \---- And finally, here’s the income breakdown that everyone’s curious about: Mel on Fanvue: $82,331.58 (Gross earnings because we have chatter cuts like 15%) Mel on Patreon: $50,865.98 (Net earnings) Jess on Fanvue: $89,068.26 (Gross earnings because we have chatter cuts like 15%) Jess on Patreon: $39,040.70 And thanks to Reddit and my old posts, I got a perfect investor like after 5 months, so this is a "payback" for that. Like I said, I'll answer every question in the comments — take care and let me know.

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰
reddit
LLM Vibe Score0
Human Vibe Score1
benfromwhereThis week

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰

(Monthly income breakdown is in the end) 📌 Introduction Hey everyone! 👋 Before I dive into this month’s breakdown, I just want to be upfront—English isn’t my first language, so I’ve used ChatGPT to refine this post for better readability. That said, everything here is 100% real—my personal experiences, struggles, and earnings as someone running a full-time AI influencer business. Since I get a lot of DMs asking about my AI models, here are their Instagram links: 📷 Emma – https://www.instagram.com/emmalauireal 📷 Jade – https://www.instagram.com/jadelaui (jadecasual is the second account) Also, if you’ve been wondering about the community I run, where I teach others how to build AI influencers from scratch, here’s the link (I got approval from mods for this link): 🔗 AI Winners Now, let’s get into what happened this month. 🚀 \------- First, a huge thank you! 🎉 Three months ago, I shared my journey of building an AI influencer business, and I was blown away by the response. That post got 263K+ views and was shared over 2.7K times—way more than I ever expected. If you’re new here or want to check out the full story of how I started, you can read it here: 🔗 Click Here (Reddit link) \------- 🔹 What I Did in January After the holiday rush in December, I knew January would be a slow month—people had already spent most of their money at the end of the year. So instead of pushing harder on monetization, I shifted my focus to tech development and optimization. Flux Character Loras: I spent a lot of time refining and testing different Flux-based character Loras for my models. This is still a work in progress, but the goal is to improve long-term consistency and make my workflow even more efficient. NSFW Content Expansion: On Emma’s side, I expanded her content library using a real model body double, making her content look more organic and natural. Jade, however, remains 100% AI-generated, keeping her workflow entirely digital. Social Media Wipeout (Thanks, VA 🙃): I had handed off both Twitter accounts to a virtual assistant to help with engagement and DMs. Big mistake. He ended up spamming DMs, which got both accounts banned—Emma (80K followers) and Jade (20K followers). 🤦‍♂️ Right now, I’m rebuilding Emma’s account from scratch and taking a much more cautious approach. Jade’s account is still offline for now. New Platform: Threads – I hadn’t touched Threads before, but since engagement on Instagram can be unpredictable, I decided to start accounts for both models. So far, they’re performing well, and I’ll continue experimenting. Launched AI Winners Community: After getting flooded with DMs (both here and on Instagram), I realized there was a massive demand for structured learning around AI influencers. So, I launched AI Winners, a paid community where I break down everything I’ve learned. It’s still early, but I see it turning into a solid, long-term community. Investment & Acquisition Talks: I’m still evaluating potential investors and acquisition offers for my AI models. There’s growing interest in buying or investing in Emma & Jade, so I’ve been having conversations to explore different options. Overall, January was about tech, rebuilding, and long-term planning—not immediate revenue. But that’s what keeps this business sustainable. 🚀 \------- ⚠️ Biggest Challenges This Month Lost Both Twitter Accounts (Massive Traffic Hit) 🚨 The biggest blow this month was losing my models’ Twitter accounts. Twitter was responsible for about 40% of my total traffic, meaning both free and paid subs took a direct hit. While Emma’s revenue took a slight dip, Jade’s income dropped significantly—partly due to the account loss and partly because January is naturally slow. (Full revenue breakdown at the end of the post.) Jade’s Instagram Tanked (Possible Shadow Ban?) 🤔 Jade’s Instagram completely lost momentum in early January. Engagement and reach dropped by over 80%, and I still haven’t figured out why. It feels like a shadow ban, but I have no clear confirmation. To counter this, I launched a second backup account, and things are starting to recover. \------- 🚀 Potential Improvements & What’s Next Locking in a Stable Workflow 🔄 Right now, Emma & Jade’s workflow is still evolving, but I’m aiming to fully stabilize it. As I’m writing this, content is generating on my second monitor—a sign that I’m close to achieving full automation without compromising quality. Boosting Jade’s Fanvue Revenue 💰 Jade’s income took a hit this month, and it’s 100% a traffic issue. The solution? More content, more reach. I’ll be increasing social media output to drive consistent traffic back to Fanvue and restore her earnings. Patreon is Done. All Focus on Fanvue 🚫 I shut down both Emma & Jade’s Patreon accounts. The goal is not to split revenue—I want everything funneled into Fanvue for higher engagement and bigger paydays. \------- 💰 January 2025 Earnings Breakdown Despite January being one of the slowest months for online creators, Emma and Jade still brought in over $29K in revenue, with a net profit exceeding $20K after all expenses. Emma Laui generated $20,206.77, with around $6,000 in expenses (chatter payments, NSFW designer fees, and other operational costs). Jade Laui earned $8,939.05, with $2,000 in expenses. Considering Twitter account losses, Instagram setbacks, and the usual January spending slump, this is still a solid outcome. The focus now is on scaling traffic and maximizing Fanvue revenue heading into February. 🚀🔥 That’s the full breakdown for January! If you have questions, feel free to drop a comment, and I’ll answer when I can. Happy to help, just like others helped me when I was starting out! 🚀🔥

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰
reddit
LLM Vibe Score0
Human Vibe Score1
benfromwhereThis week

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰

(Monthly income breakdown is in the end) 📌 Introduction Hey everyone! 👋 Before I dive into this month’s breakdown, I just want to be upfront—English isn’t my first language, so I’ve used ChatGPT to refine this post for better readability. That said, everything here is 100% real—my personal experiences, struggles, and earnings as someone running a full-time AI influencer business. Since I get a lot of DMs asking about my AI models, here are their Instagram links: 📷 Emma – https://www.instagram.com/emmalauireal 📷 Jade – https://www.instagram.com/jadelaui (jadecasual is the second account) Also, if you’ve been wondering about the community I run, where I teach others how to build AI influencers from scratch, here’s the link (I got approval from mods for this link): 🔗 AI Winners Now, let’s get into what happened this month. 🚀 \------- First, a huge thank you! 🎉 Three months ago, I shared my journey of building an AI influencer business, and I was blown away by the response. That post got 263K+ views and was shared over 2.7K times—way more than I ever expected. If you’re new here or want to check out the full story of how I started, you can read it here: 🔗 Click Here (Reddit link) \------- 🔹 What I Did in January After the holiday rush in December, I knew January would be a slow month—people had already spent most of their money at the end of the year. So instead of pushing harder on monetization, I shifted my focus to tech development and optimization. Flux Character Loras: I spent a lot of time refining and testing different Flux-based character Loras for my models. This is still a work in progress, but the goal is to improve long-term consistency and make my workflow even more efficient. NSFW Content Expansion: On Emma’s side, I expanded her content library using a real model body double, making her content look more organic and natural. Jade, however, remains 100% AI-generated, keeping her workflow entirely digital. Social Media Wipeout (Thanks, VA 🙃): I had handed off both Twitter accounts to a virtual assistant to help with engagement and DMs. Big mistake. He ended up spamming DMs, which got both accounts banned—Emma (80K followers) and Jade (20K followers). 🤦‍♂️ Right now, I’m rebuilding Emma’s account from scratch and taking a much more cautious approach. Jade’s account is still offline for now. New Platform: Threads – I hadn’t touched Threads before, but since engagement on Instagram can be unpredictable, I decided to start accounts for both models. So far, they’re performing well, and I’ll continue experimenting. Launched AI Winners Community: After getting flooded with DMs (both here and on Instagram), I realized there was a massive demand for structured learning around AI influencers. So, I launched AI Winners, a paid community where I break down everything I’ve learned. It’s still early, but I see it turning into a solid, long-term community. Investment & Acquisition Talks: I’m still evaluating potential investors and acquisition offers for my AI models. There’s growing interest in buying or investing in Emma & Jade, so I’ve been having conversations to explore different options. Overall, January was about tech, rebuilding, and long-term planning—not immediate revenue. But that’s what keeps this business sustainable. 🚀 \------- ⚠️ Biggest Challenges This Month Lost Both Twitter Accounts (Massive Traffic Hit) 🚨 The biggest blow this month was losing my models’ Twitter accounts. Twitter was responsible for about 40% of my total traffic, meaning both free and paid subs took a direct hit. While Emma’s revenue took a slight dip, Jade’s income dropped significantly—partly due to the account loss and partly because January is naturally slow. (Full revenue breakdown at the end of the post.) Jade’s Instagram Tanked (Possible Shadow Ban?) 🤔 Jade’s Instagram completely lost momentum in early January. Engagement and reach dropped by over 80%, and I still haven’t figured out why. It feels like a shadow ban, but I have no clear confirmation. To counter this, I launched a second backup account, and things are starting to recover. \------- 🚀 Potential Improvements & What’s Next Locking in a Stable Workflow 🔄 Right now, Emma & Jade’s workflow is still evolving, but I’m aiming to fully stabilize it. As I’m writing this, content is generating on my second monitor—a sign that I’m close to achieving full automation without compromising quality. Boosting Jade’s Fanvue Revenue 💰 Jade’s income took a hit this month, and it’s 100% a traffic issue. The solution? More content, more reach. I’ll be increasing social media output to drive consistent traffic back to Fanvue and restore her earnings. Patreon is Done. All Focus on Fanvue 🚫 I shut down both Emma & Jade’s Patreon accounts. The goal is not to split revenue—I want everything funneled into Fanvue for higher engagement and bigger paydays. \------- 💰 January 2025 Earnings Breakdown Despite January being one of the slowest months for online creators, Emma and Jade still brought in over $29K in revenue, with a net profit exceeding $20K after all expenses. Emma Laui generated $20,206.77, with around $6,000 in expenses (chatter payments, NSFW designer fees, and other operational costs). Jade Laui earned $8,939.05, with $2,000 in expenses. Considering Twitter account losses, Instagram setbacks, and the usual January spending slump, this is still a solid outcome. The focus now is on scaling traffic and maximizing Fanvue revenue heading into February. 🚀🔥 That’s the full breakdown for January! If you have questions, feel free to drop a comment, and I’ll answer when I can. Happy to help, just like others helped me when I was starting out! 🚀🔥

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰
reddit
LLM Vibe Score0
Human Vibe Score1
benfromwhereThis week

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰

(Monthly income breakdown is in the end) 📌 Introduction Hey everyone! 👋 Before I dive into this month’s breakdown, I just want to be upfront—English isn’t my first language, so I’ve used ChatGPT to refine this post for better readability. That said, everything here is 100% real—my personal experiences, struggles, and earnings as someone running a full-time AI influencer business. Since I get a lot of DMs asking about my AI models, here are their Instagram links: 📷 Emma – https://www.instagram.com/emmalauireal 📷 Jade – https://www.instagram.com/jadelaui (jadecasual is the second account) Also, if you’ve been wondering about the community I run, where I teach others how to build AI influencers from scratch, here’s the link (I got approval from mods for this link): 🔗 AI Winners Now, let’s get into what happened this month. 🚀 \------- First, a huge thank you! 🎉 Three months ago, I shared my journey of building an AI influencer business, and I was blown away by the response. That post got 263K+ views and was shared over 2.7K times—way more than I ever expected. If you’re new here or want to check out the full story of how I started, you can read it here: 🔗 Click Here (Reddit link) \------- 🔹 What I Did in January After the holiday rush in December, I knew January would be a slow month—people had already spent most of their money at the end of the year. So instead of pushing harder on monetization, I shifted my focus to tech development and optimization. Flux Character Loras: I spent a lot of time refining and testing different Flux-based character Loras for my models. This is still a work in progress, but the goal is to improve long-term consistency and make my workflow even more efficient. NSFW Content Expansion: On Emma’s side, I expanded her content library using a real model body double, making her content look more organic and natural. Jade, however, remains 100% AI-generated, keeping her workflow entirely digital. Social Media Wipeout (Thanks, VA 🙃): I had handed off both Twitter accounts to a virtual assistant to help with engagement and DMs. Big mistake. He ended up spamming DMs, which got both accounts banned—Emma (80K followers) and Jade (20K followers). 🤦‍♂️ Right now, I’m rebuilding Emma’s account from scratch and taking a much more cautious approach. Jade’s account is still offline for now. New Platform: Threads – I hadn’t touched Threads before, but since engagement on Instagram can be unpredictable, I decided to start accounts for both models. So far, they’re performing well, and I’ll continue experimenting. Launched AI Winners Community: After getting flooded with DMs (both here and on Instagram), I realized there was a massive demand for structured learning around AI influencers. So, I launched AI Winners, a paid community where I break down everything I’ve learned. It’s still early, but I see it turning into a solid, long-term community. Investment & Acquisition Talks: I’m still evaluating potential investors and acquisition offers for my AI models. There’s growing interest in buying or investing in Emma & Jade, so I’ve been having conversations to explore different options. Overall, January was about tech, rebuilding, and long-term planning—not immediate revenue. But that’s what keeps this business sustainable. 🚀 \------- ⚠️ Biggest Challenges This Month Lost Both Twitter Accounts (Massive Traffic Hit) 🚨 The biggest blow this month was losing my models’ Twitter accounts. Twitter was responsible for about 40% of my total traffic, meaning both free and paid subs took a direct hit. While Emma’s revenue took a slight dip, Jade’s income dropped significantly—partly due to the account loss and partly because January is naturally slow. (Full revenue breakdown at the end of the post.) Jade’s Instagram Tanked (Possible Shadow Ban?) 🤔 Jade’s Instagram completely lost momentum in early January. Engagement and reach dropped by over 80%, and I still haven’t figured out why. It feels like a shadow ban, but I have no clear confirmation. To counter this, I launched a second backup account, and things are starting to recover. \------- 🚀 Potential Improvements & What’s Next Locking in a Stable Workflow 🔄 Right now, Emma & Jade’s workflow is still evolving, but I’m aiming to fully stabilize it. As I’m writing this, content is generating on my second monitor—a sign that I’m close to achieving full automation without compromising quality. Boosting Jade’s Fanvue Revenue 💰 Jade’s income took a hit this month, and it’s 100% a traffic issue. The solution? More content, more reach. I’ll be increasing social media output to drive consistent traffic back to Fanvue and restore her earnings. Patreon is Done. All Focus on Fanvue 🚫 I shut down both Emma & Jade’s Patreon accounts. The goal is not to split revenue—I want everything funneled into Fanvue for higher engagement and bigger paydays. \------- 💰 January 2025 Earnings Breakdown Despite January being one of the slowest months for online creators, Emma and Jade still brought in over $29K in revenue, with a net profit exceeding $20K after all expenses. Emma Laui generated $20,206.77, with around $6,000 in expenses (chatter payments, NSFW designer fees, and other operational costs). Jade Laui earned $8,939.05, with $2,000 in expenses. Considering Twitter account losses, Instagram setbacks, and the usual January spending slump, this is still a solid outcome. The focus now is on scaling traffic and maximizing Fanvue revenue heading into February. 🚀🔥 That’s the full breakdown for January! If you have questions, feel free to drop a comment, and I’ll answer when I can. Happy to help, just like others helped me when I was starting out! 🚀🔥

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey
reddit
LLM Vibe Score0
Human Vibe Score0.778
benfromwhereThis week

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey

Update on February 22th: I changed my AI influencer's names because it caused some problems on my business. One year, two AI-powered influencers, and $250K in revenue. Sounds unreal? It’s not. Today, I’m pulling back the curtain on the strategies, tools, and hard-won lessons that took me from concept to a six-figure success story in the AI influencer space. Hey, I'm Ben—a 32-year-old designer who spent the past year navigating the world of AI influencers. Let me clear up any confusion right from the start: I’m not here to sell you anything. This is purely a case study to share what worked, what didn’t, and what I’ve learned along the way. I’ll also make sure to answer all your questions in the comments for free whenever I can, so don’t hesitate to ask. Links to Past Topics: If you're curious about some of the groundwork I covered, check out a few of my earlier posts here: How I Make $10,000 Monthly | AI Influencer Management How I Earned $7000+ in 15 Days | AI Influencer Business Update These earlier posts cover a lot of the backstory, so feel free to explore them before diving into this one. So if you're ready, here is the full story: \---- The idea of creating an AI influencer was one of those “what if” moments that wouldn’t leave my mind. At first, it sounded futuristic—even a bit too ambitious. It all started when I stumbled upon an AI influencer on Instagram with the handle AnnaMaes2000. Her content blew me away—the quality, the detail, and just how real everything looked. I was instantly hooked and ended up going through every post, just trying to figure out how she was pulling this off. That’s when I knew I had to learn how this was done. The next step? YouTube. I dived into videos on Stable Diffusion, soaking up everything I could about creating AI-generated images. Those tutorials taught me the basics and got me up to speed. Then, I created my first AI influencer, let's call her Mel for now. Right after that, to complete the storyline and boost engagement, I introduced Mel's “mother,” Jess. Adding Jess gave the whole project depth and a narrative that drew people in, creating a unique family dynamic that instantly elevated traffic and interest. After thousands of bad photos, hundreds of deleted posts, and months of trial and error, you can now see the quality that defines my current accounts. Here’s a rundown of the tools and checkpoints I’ve used from day one, in order: Fooocus on RunDiffusion — Juggernaut V8 Fooocus on RunDiffusion — Juggernaut V9 Fooocus on PC (locally) — Juggernaut V9 Fooocus on PC (locally) —Lyuyang Mix + Juggernaut V9 Flux on PC (couple of photos only since it's so slow even on RTX 4090) Flux on Fal.ai. \---- There’s no magic Instagram hack that guarantees success, despite what everyone thinks and keeps asking me. Quality content, consistent uploads, and solid craftsmanship are what actually help your photos hit trends and show up on the Explore page. Unlike 95% of low-quality AI accounts out there, I don’t rely on faceswap videos, spam Reels, or go around liking comments on other accounts. My approach is fully organic, focused solely on creating my own unique content. By following Instagram's guidelines to the letter, I've managed to direct some of Mel and Jess' fans over to Patreon and Fanvue. There, for a small subscription fee, fans can access exclusive lingerie content. For those looking for more, higher-tier subscriptions give access to even more premium content. Some possible questions and their answers: No, you can't share hardcore NSFW content on Patreon. You can do that on Fanvue. Yes, you can create AI creators on Fanvue — OnlyFans doesn't allow it. Yes, you can use your own ID to get KYC. Yes, we're telling both Mel and Jess is (or use) AI to generate content. And yes, some people leave and some people still have fun with chatting, having a good time and get perfect content for their needs. And yes, we have a chatter team to work on these accounts. \---- This journey wasn’t all smooth sailing. I faced unexpected roadblocks, like platform restrictions that limited certain types of content, and managing fan expectations was more challenging than anticipated. Staying within guidelines while keeping fans engaged required constant adaptation. These hurdles forced me to get creative, adjust my approach, and learn fast. Once I saw Mel and Jess gaining traction, I knew it was time to scale up. Expanding meant finding new ways to keep content fresh, creating deeper narratives, and considering how to bring even more followers into the fold. My focus turned to building a sustainable model that could grow without sacrificing quality or authenticity. If you’re thinking about diving into AI content creation, here’s my advice: patience, consistency, and a focus on quality are key. Don’t cut corners or rely on quick-fix hacks. Invest time in learning the right tools, creating engaging stories, and building an audience that values what you bring to the table. This approach took me from zero to six figures, and it’s what makes the journey worth it. \---- And finally, here’s the income breakdown that everyone’s curious about: Mel on Fanvue: $82,331.58 (Gross earnings because we have chatter cuts like 15%) Mel on Patreon: $50,865.98 (Net earnings) Jess on Fanvue: $89,068.26 (Gross earnings because we have chatter cuts like 15%) Jess on Patreon: $39,040.70 And thanks to Reddit and my old posts, I got a perfect investor like after 5 months, so this is a "payback" for that. Like I said, I'll answer every question in the comments — take care and let me know.

AITreasureBox
github
LLM Vibe Score0.447
Human Vibe Score0.1014145151561518
superiorluMar 28, 2025

AITreasureBox

AI TreasureBox English | 中文 Collect practical AI repos, tools, websites, papers and tutorials on AI. Translated from ChatGPT, picture from Midjourney. Catalog Repos Tools Websites Report&Paper Tutorials Repos updated repos and stars every 2 hours and re-ranking automatically. | No. | Repos | Description | | ----:|:-----------------------------------------|:------------------------------------------------------------------------------------------------------| | 1|🔥codecrafters-io/build-your-own-x !2025-03-28364681428|Master programming by recreating your favorite technologies from scratch.| | 2|sindresorhus/awesome !2025-03-28353614145|😎 Awesome lists about all kinds of interesting topics| | 3|public-apis/public-apis !2025-03-28334299125|A collective list of free APIs| | 4|kamranahmedse/developer-roadmap !2025-03-2831269540|Interactive roadmaps, guides and other educational content to help developers grow in their careers.| | 5|vinta/awesome-python !2025-03-28238581114|A curated list of awesome Python frameworks, libraries, software and resources| | 6|practical-tutorials/project-based-learning !2025-03-28222661124|Curated list of project-based tutorials| | 7|tensorflow/tensorflow !2025-03-281888714|An Open Source Machine Learning Framework for Everyone| | 8|Significant-Gravitas/AutoGPT !2025-03-2817391338|An experimental open-source attempt to make GPT-4 fully autonomous.| | 9|jackfrued/Python-100-Days !2025-03-2816305141|Python - 100天从新手到大师| | 10|AUTOMATIC1111/stable-diffusion-webui !2025-03-2815011553|Stable Diffusion web UI| | 11|huggingface/transformers !2025-03-2814207850|🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.| | 12|ollama/ollama !2025-03-28135166151|Get up and running with Llama 2, Mistral, Gemma, and other large language models.| | 13|f/awesome-chatgpt-prompts !2025-03-2812212738 |This repo includes ChatGPT prompt curation to use ChatGPT better.| | 14|justjavac/free-programming-books-zhCN !2025-03-2811316119|📚 免费的计算机编程类中文书籍,欢迎投稿| | 15|krahets/hello-algo !2025-03-2811107930|《Hello 算法》:动画图解、一键运行的数据结构与算法教程。支持 Python, Java, C++, C, C#, JS, Go, Swift, Rust, Ruby, Kotlin, TS, Dart 代码。简体版和繁体版同步更新,English version ongoing| | 16|yt-dlp/yt-dlp !2025-03-28105801114|A feature-rich command-line audio/video downloader| | 17|langchain-ai/langchain !2025-03-2810449479|⚡ Building applications with LLMs through composability ⚡| | 18|goldbergyoni/nodebestpractices !2025-03-281021629|✅ The Node.js best practices list (July 2024)| | 19|puppeteer/puppeteer !2025-03-289018212|JavaScript API for Chrome and Firefox| | 20|pytorch/pytorch !2025-03-288833938|Tensors and Dynamic neural networks in Python with strong GPU acceleration| | 21|neovim/neovim !2025-03-288781482|Vim-fork focused on extensibility and usability| | 22|🔥🔥langgenius/dify !2025-03-2887342639 |One API for plugins and datasets, one interface for prompt engineering and visual operation, all for creating powerful AI applications.| | 23|mtdvio/every-programmer-should-know !2025-03-28867069|A collection of (mostly) technical things every software developer should know about| | 24|open-webui/open-webui !2025-03-2886025159|User-friendly WebUI for LLMs (Formerly Ollama WebUI)| | 25|ChatGPTNextWeb/NextChat !2025-03-288231521|✨ Light and Fast AI Assistant. Support: Web | | 26|supabase/supabase !2025-03-287990956|The open source Firebase alternative.| | 27|openai/whisper !2025-03-287905542|Robust Speech Recognition via Large-Scale Weak Supervision| | 28|home-assistant/core !2025-03-287773219|🏡 Open source home automation that puts local control and privacy first.| | 29|tensorflow/models !2025-03-28774694|Models and examples built with TensorFlow| | 30| ggerganov/llama.cpp !2025-03-287731836 | Port of Facebook's LLaMA model in C/C++ | | 31|3b1b/manim !2025-03-287641918|Animation engine for explanatory math videos| | 32|microsoft/generative-ai-for-beginners !2025-03-287623860|12 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/| | 33|nomic-ai/gpt4all !2025-03-28729285 |gpt4all: an ecosystem of open-source chatbots trained on a massive collection of clean assistant data including code, stories and dialogue| | 34|comfyanonymous/ComfyUI !2025-03-2872635111|The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.| | 35|bregman-arie/devops-exercises !2025-03-2872225209|Linux, Jenkins, AWS, SRE, Prometheus, Docker, Python, Ansible, Git, Kubernetes, Terraform, OpenStack, SQL, NoSQL, Azure, GCP, DNS, Elastic, Network, Virtualization. DevOps Interview Questions| | 36|elastic/elasticsearch !2025-03-28721419|Free and Open, Distributed, RESTful Search Engine| | 37|🔥n8n-io/n8n !2025-03-2872093495|Free and source-available fair-code licensed workflow automation tool. Easily automate tasks across different services.| | 38|fighting41love/funNLP !2025-03-287200422|The Most Powerful NLP-Weapon Arsenal| | 39|hoppscotch/hoppscotch !2025-03-287060134|Open source API development ecosystem - https://hoppscotch.io (open-source alternative to Postman, Insomnia)| | 40|abi/screenshot-to-code !2025-03-286932817|Drop in a screenshot and convert it to clean HTML/Tailwind/JS code| | 41|binary-husky/gptacademic !2025-03-28680374|Academic Optimization of GPT| | 42|d2l-ai/d2l-zh !2025-03-286774142|Targeting Chinese readers, functional and open for discussion. The Chinese and English versions are used for teaching in over 400 universities across more than 60 countries| | 43|josephmisiti/awesome-machine-learning !2025-03-286739215|A curated list of awesome Machine Learning frameworks, libraries and software.| | 44|grafana/grafana !2025-03-286725414|The open and composable observability and data visualization platform. Visualize metrics, logs, and traces from multiple sources like Prometheus, Loki, Elasticsearch, InfluxDB, Postgres and many more.| | 45|python/cpython !2025-03-286602218|The Python programming language| | 46|apache/superset !2025-03-286519020|Apache Superset is a Data Visualization and Data Exploration Platform| | 47|xtekky/gpt4free !2025-03-28639391 |decentralizing the Ai Industry, free gpt-4/3.5 scripts through several reverse engineered API's ( poe.com, phind.com, chat.openai.com etc...)| | 48|sherlock-project/sherlock !2025-03-286332536|Hunt down social media accounts by username across social networks| | 49|twitter/the-algorithm !2025-03-28630586 |Source code for Twitter's Recommendation Algorithm| | 50|keras-team/keras !2025-03-28627835|Deep Learning for humans| | 51|openai/openai-cookbook !2025-03-28625136 |Examples and guides for using the OpenAI API| | 52|immich-app/immich !2025-03-286238670|High performance self-hosted photo and video management solution.| | 53|AppFlowy-IO/AppFlowy !2025-03-286173528|Bring projects, wikis, and teams together with AI. AppFlowy is an AI collaborative workspace where you achieve more without losing control of your data. The best open source alternative to Notion.| | 54|scikit-learn/scikit-learn !2025-03-286158212|scikit-learn: machine learning in Python| | 55|binhnguyennus/awesome-scalability !2025-03-286117021|The Patterns of Scalable, Reliable, and Performant Large-Scale Systems| | 56|labmlai/annotateddeeplearningpaperimplementations !2025-03-285951726|🧑‍🏫 59 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠| | 57|OpenInterpreter/open-interpreter !2025-03-285894710|A natural language interface for computers| | 58|lobehub/lobe-chat !2025-03-285832054|🤖 Lobe Chat - an open-source, extensible (Function Calling), high-performance chatbot framework. It supports one-click free deployment of your private ChatGPT/LLM web application.| | 59|meta-llama/llama !2025-03-28579536|Inference code for Llama models| | 60|nuxt/nuxt !2025-03-28566437|The Intuitive Vue Framework.| | 61|imartinez/privateGPT !2025-03-28555192|Interact with your documents using the power of GPT, 100% privately, no data leaks| | 62|Stirling-Tools/Stirling-PDF !2025-03-285500846|#1 Locally hosted web application that allows you to perform various operations on PDF files| | 63|PlexPt/awesome-chatgpt-prompts-zh !2025-03-285459720|ChatGPT Chinese Training Guide. Guidelines for various scenarios. Learn how to make it listen to you| | 64|dair-ai/Prompt-Engineering-Guide !2025-03-285451025 |🐙 Guides, papers, lecture, notebooks and resources for prompt engineering| | 65|ageitgey/facerecognition !2025-03-28544382|The world's simplest facial recognition api for Python and the command line| | 66|CorentinJ/Real-Time-Voice-Cloning !2025-03-285384814|Clone a voice in 5 seconds to generate arbitrary speech in real-time| | 67|geekan/MetaGPT !2025-03-285375376|The Multi-Agent Meta Programming Framework: Given one line Requirement, return PRD, Design, Tasks, Repo | | 68|gpt-engineer-org/gpt-engineer !2025-03-285367419|Specify what you want it to build, the AI asks for clarification, and then builds it.| | 69|lencx/ChatGPT !2025-03-2853653-3|🔮 ChatGPT Desktop Application (Mac, Windows and Linux)| | 70|deepfakes/faceswap !2025-03-28535672|Deepfakes Software For All| | 71|langflow-ai/langflow !2025-03-285319584|Langflow is a low-code app builder for RAG and multi-agent AI applications. It’s Python-based and agnostic to any model, API, or database.| | 72|commaai/openpilot !2025-03-28529759|openpilot is an operating system for robotics. Currently, it upgrades the driver assistance system on 275+ supported cars.| | 73|clash-verge-rev/clash-verge-rev !2025-03-2852848124|Continuation of Clash Verge - A Clash Meta GUI based on Tauri (Windows, MacOS, Linux)| | 74|All-Hands-AI/OpenHands !2025-03-285150675|🙌 OpenHands: Code Less, Make More| | 75|xai-org/grok-1 !2025-03-28502504|Grok open release| | 76|meilisearch/meilisearch !2025-03-284999122|A lightning-fast search API that fits effortlessly into your apps, websites, and workflow| | 77|🔥browser-use/browser-use !2025-03-2849910294|Make websites accessible for AI agents| | 78|jgthms/bulma !2025-03-28496783|Modern CSS framework based on Flexbox| | 79|facebookresearch/segment-anything !2025-03-284947116|The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.| |!green-up-arrow.svg 80|hacksider/Deep-Live-Cam !2025-03-2848612146|real time face swap and one-click video deepfake with only a single image (uncensored)| |!red-down-arrow 81|mlabonne/llm-course !2025-03-284860934|Course with a roadmap and notebooks to get into Large Language Models (LLMs).| | 82|PaddlePaddle/PaddleOCR !2025-03-284785530|Awesome multilingual OCR toolkits based on PaddlePaddle (practical ultra lightweight OCR system, support 80+ languages recognition, provide data annotation and synthesis tools, support training and deployment among server, mobile, embedded and IoT devices)| | 83|alist-org/alist !2025-03-284732618|🗂️A file list/WebDAV program that supports multiple storages, powered by Gin and Solidjs. / 一个支持多存储的文件列表/WebDAV程序,使用 Gin 和 Solidjs。| | 84|infiniflow/ragflow !2025-03-2847027129|RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.| | 85|Avik-Jain/100-Days-Of-ML-Code !2025-03-284679312|100 Days of ML Coding| | 86|v2ray/v2ray-core !2025-03-28458706|A platform for building proxies to bypass network restrictions.| | 87|hiyouga/LLaMA-Factory !2025-03-284555881|Easy-to-use LLM fine-tuning framework (LLaMA, BLOOM, Mistral, Baichuan, Qwen, ChatGLM)| | 88|Asabeneh/30-Days-Of-Python !2025-03-284544930|30 days of Python programming challenge is a step-by-step guide to learn the Python programming language in 30 days. This challenge may take more than100 days, follow your own pace. These videos may help too: https://www.youtube.com/channel/UC7PNRuno1rzYPb1xLa4yktw| | 89|type-challenges/type-challenges !2025-03-284488511|Collection of TypeScript type challenges with online judge| | 90|lllyasviel/Fooocus !2025-03-284402716|Focus on prompting and generating| | 91|RVC-Boss/GPT-SoVITS !2025-03-284327738|1 min voice data can also be used to train a good TTS model! (few shot voice cloning)| | 92|rasbt/LLMs-from-scratch !2025-03-284320667|Implementing a ChatGPT-like LLM from scratch, step by step| | 93|oobabooga/text-generation-webui !2025-03-284302012 |A gradio web UI for running Large Language Models like LLaMA, llama.cpp, GPT-J, OPT, and GALACTICA.| | 94|vllm-project/vllm !2025-03-2842982102|A high-throughput and memory-efficient inference and serving engine for LLMs| | 95|dani-garcia/vaultwarden !2025-03-284297121|Unofficial Bitwarden compatible server written in Rust, formerly known as bitwarden_rs| | 96|microsoft/autogen !2025-03-284233049|Enable Next-Gen Large Language Model Applications. Join our Discord: https://discord.gg/pAbnFJrkgZ| | 97|jeecgboot/JeecgBoot !2025-03-284205920|🔥「企业级低代码平台」前后端分离架构SpringBoot 2.x/3.x,SpringCloud,Ant Design&Vue3,Mybatis,Shiro,JWT。强大的代码生成器让前后端代码一键生成,无需写任何代码! 引领新的开发模式OnlineCoding->代码生成->手工MERGE,帮助Java项目解决70%重复工作,让开发更关注业务,既能快速提高效率,帮助公司节省成本,同时又不失灵活性。| | 98|Mintplex-Labs/anything-llm !2025-03-284186955|A full-stack application that turns any documents into an intelligent chatbot with a sleek UI and easier way to manage your workspaces.| | 99|THUDM/ChatGLM-6B !2025-03-28410192 |ChatGLM-6B: An Open Bilingual Dialogue Language Model| | 100|hpcaitech/ColossalAI !2025-03-28406902|Making large AI models cheaper, faster and more accessible| | 101|Stability-AI/stablediffusion !2025-03-28406337|High-Resolution Image Synthesis with Latent Diffusion Models| | 102|mingrammer/diagrams !2025-03-28405063|🎨 Diagram as Code for prototyping cloud system architectures| | 103|Kong/kong !2025-03-28404616|🦍 The Cloud-Native API Gateway and AI Gateway.| | 104|getsentry/sentry !2025-03-284040913|Developer-first error tracking and performance monitoring| | 105| karpathy/nanoGPT !2025-03-284034613 |The simplest, fastest repository for training/finetuning medium-sized GPTs| | 106|fastlane/fastlane !2025-03-2840014-1|🚀 The easiest way to automate building and releasing your iOS and Android apps| | 107|psf/black !2025-03-28399765|The uncompromising Python code formatter| | 108|OpenBB-finance/OpenBBTerminal !2025-03-283972074 |Investment Research for Everyone, Anywhere.| | 109|2dust/v2rayNG !2025-03-283943415|A V2Ray client for Android, support Xray core and v2fly core| | 110|apache/airflow !2025-03-283937314|Apache Airflow - A platform to programmatically author, schedule, and monitor workflows| | 111|KRTirtho/spotube !2025-03-283902746|🎧 Open source Spotify client that doesn't require Premium nor uses Electron! Available for both desktop & mobile!| | 112|coqui-ai/TTS !2025-03-283889719 |🐸💬 - a deep learning toolkit for Text-to-Speech, battle-tested in research and production| | 113|ggerganov/whisper.cpp !2025-03-283882116|Port of OpenAI's Whisper model in C/C++| | 114|ultralytics/ultralytics !2025-03-283866951|NEW - YOLOv8 🚀 in PyTorch > ONNX > OpenVINO > CoreML > TFLite| | 115|typst/typst !2025-03-283863914|A new markup-based typesetting system that is powerful and easy to learn.| | 116|streamlit/streamlit !2025-03-283845828|Streamlit — A faster way to build and share data apps.| | 117|LC044/WeChatMsg !2025-03-283836931|提取微信聊天记录,将其导出成HTML、Word、Excel文档永久保存,对聊天记录进行分析生成年度聊天报告,用聊天数据训练专属于个人的AI聊天助手| | 118|lm-sys/FastChat !2025-03-283822112 |An open platform for training, serving, and evaluating large languages. Release repo for Vicuna and FastChat-T5.| | 119|NaiboWang/EasySpider !2025-03-283819013|A visual no-code/code-free web crawler/spider易采集:一个可视化浏览器自动化测试/数据采集/爬虫软件,可以无代码图形化的设计和执行爬虫任务。别名:ServiceWrapper面向Web应用的智能化服务封装系统。| | 120|microsoft/DeepSpeed !2025-03-283765816 |A deep learning optimization library that makes distributed training and inference easy, efficient, and effective| | 121|QuivrHQ/quivr !2025-03-28376067|Your GenAI Second Brain 🧠 A personal productivity assistant (RAG) ⚡️🤖 Chat with your docs (PDF, CSV, ...) & apps using Langchain, GPT 3.5 / 4 turbo, Private, Anthropic, VertexAI, Ollama, LLMs, that you can share with users ! Local & Private alternative to OpenAI GPTs & ChatGPT powered by retrieval-augmented generation.| | 122|freqtrade/freqtrade !2025-03-283757817 |Free, open source crypto trading bot| | 123|suno-ai/bark !2025-03-28373178 |🔊 Text-Prompted Generative Audio Model| | 124|🔥cline/cline !2025-03-2837307282|Autonomous coding agent right in your IDE, capable of creating/editing files, executing commands, and more with your permission every step of the way.| | 125|LAION-AI/Open-Assistant !2025-03-28372712 |OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.| | 126|penpot/penpot !2025-03-283716217|Penpot: The open-source design tool for design and code collaboration| | 127|gradio-app/gradio !2025-03-283713320|Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work!| | 128|FlowiseAI/Flowise !2025-03-283667135 |Drag & drop UI to build your customized LLM flow using LangchainJS| | 129|SimplifyJobs/Summer2025-Internships !2025-03-28366506|Collection of Summer 2025 tech internships!| | 130|TencentARC/GFPGAN !2025-03-28365027 |GFPGAN aims at developing Practical Algorithms for Real-world Face Restoration.| | 131|ray-project/ray !2025-03-283626819|Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a toolkit of libraries (Ray AIR) for accelerating ML workloads.| | 132|babysor/MockingBird !2025-03-28360498|🚀AI拟声: 5秒内克隆您的声音并生成任意语音内容 Clone a voice in 5 seconds to generate arbitrary speech in real-time| | 133|unslothai/unsloth !2025-03-283603691|5X faster 50% less memory LLM finetuning| | 134|zhayujie/chatgpt-on-wechat !2025-03-283600124 |Wechat robot based on ChatGPT, which uses OpenAI api and itchat library| | 135|upscayl/upscayl !2025-03-283599824|🆙 Upscayl - Free and Open Source AI Image Upscaler for Linux, MacOS and Windows built with Linux-First philosophy.| | 136|freeCodeCamp/devdocs !2025-03-28359738|API Documentation Browser| | 137|XingangPan/DragGAN !2025-03-28359043 |Code for DragGAN (SIGGRAPH 2023)| | 138|2noise/ChatTTS !2025-03-283543922|ChatTTS is a generative speech model for daily dialogue.| | 139|google-research/google-research !2025-03-28352207 |Google Research| | 140|karanpratapsingh/system-design !2025-03-28351003|Learn how to design systems at scale and prepare for system design interviews| | 141|lapce/lapce !2025-03-28350855|Lightning-fast and Powerful Code Editor written in Rust| | 142| microsoft/TaskMatrix !2025-03-2834500-3 | Talking, Drawing and Editing with Visual Foundation Models| | 143|chatchat-space/Langchain-Chatchat !2025-03-283442020|Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM) QA app with langchain| | 144|unclecode/crawl4ai !2025-03-283434163|🔥🕷️ Crawl4AI: Open-source LLM Friendly Web Crawler & Scrapper| | 145|Bin-Huang/chatbox !2025-03-283374733 |A desktop app for GPT-4 / GPT-3.5 (OpenAI API) that supports Windows, Mac & Linux| | 146|milvus-io/milvus !2025-03-283366525 |A cloud-native vector database, storage for next generation AI applications| | 147|mendableai/firecrawl !2025-03-2833297128|🔥 Turn entire websites into LLM-ready markdown| | 148|pola-rs/polars !2025-03-283269320|Fast multi-threaded, hybrid-out-of-core query engine focussing on DataFrame front-ends| | 149|Pythagora-io/gpt-pilot !2025-03-28325321|PoC for a scalable dev tool that writes entire apps from scratch while the developer oversees the implementation| | 150|hashicorp/vault !2025-03-28320797|A tool for secrets management, encryption as a service, and privileged access management| | 151|shardeum/shardeum !2025-03-28319580|Shardeum is an EVM based autoscaling blockchain| | 152|Chanzhaoyu/chatgpt-web !2025-03-28319242 |A demonstration website built with Express and Vue3 called ChatGPT| | 153|lllyasviel/ControlNet !2025-03-283186413 |Let us control diffusion models!| | 154|google/jax !2025-03-28317727|Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more| | 155|facebookresearch/detectron2 !2025-03-28315987|Detectron2 is a platform for object detection, segmentation and other visual recognition tasks.| | 156|myshell-ai/OpenVoice !2025-03-28315233|Instant voice cloning by MyShell| | 157|TheAlgorithms/C-Plus-Plus !2025-03-283151411|Collection of various algorithms in mathematics, machine learning, computer science and physics implemented in C++ for educational purposes.| | 158|hiroi-sora/Umi-OCR !2025-03-283138129|OCR图片转文字识别软件,完全离线。截屏/批量导入图片,支持多国语言、合并段落、竖排文字。可排除水印区域,提取干净的文本。基于 PaddleOCR 。| | 159|mudler/LocalAI !2025-03-283127815|🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.| | 160|facebookresearch/fairseq !2025-03-28312124 |Facebook AI Research Sequence-to-Sequence Toolkit written in Python.| | 161|alibaba/nacos !2025-03-28310559|an easy-to-use dynamic service discovery, configuration and service management platform for building cloud native applications.| | 162|yunjey/pytorch-tutorial !2025-03-28310326|PyTorch Tutorial for Deep Learning Researchers| | 163|v2fly/v2ray-core !2025-03-28307448|A platform for building proxies to bypass network restrictions.| | 164|mckaywrigley/chatbot-ui !2025-03-283067714|The open-source AI chat interface for everyone.| | 165|TabbyML/tabby !2025-03-28305949 |Self-hosted AI coding assistant| | 166|deepseek-ai/awesome-deepseek-integration !2025-03-283053193|| | 167|danielmiessler/fabric !2025-03-283028914|fabric is an open-source framework for augmenting humans using AI.| | 168|xinntao/Real-ESRGAN !2025-03-283026623 |Real-ESRGAN aims at developing Practical Algorithms for General Image/Video Restoration.| | 169|paul-gauthier/aider !2025-03-283014642|aider is GPT powered coding in your terminal| | 170|tatsu-lab/stanfordalpaca !2025-03-28299022 |Code and documentation to train Stanford's Alpaca models, and generate the data.| | 171|DataTalksClub/data-engineering-zoomcamp !2025-03-282971817|Free Data Engineering course!| | 172|HeyPuter/puter !2025-03-282967014|🌐 The Internet OS! Free, Open-Source, and Self-Hostable.| | 173|mli/paper-reading !2025-03-282962314|Classic Deep Learning and In-Depth Reading of New Papers Paragraph by Paragraph| | 174|linexjlin/GPTs !2025-03-28295568|leaked prompts of GPTs| | 175|s0md3v/roop !2025-03-28295286 |one-click deepfake (face swap)| | 176|JushBJJ/Mr.-Ranedeer-AI-Tutor !2025-03-2829465-1 |A GPT-4 AI Tutor Prompt for customizable personalized learning experiences.| | 177|opendatalab/MinerU !2025-03-282927074|A one-stop, open-source, high-quality data extraction tool, supports PDF/webpage/e-book extraction.一站式开源高质量数据提取工具,支持PDF/网页/多格式电子书提取。| | 178|mouredev/Hello-Python !2025-03-282920720|Curso para aprender el lenguaje de programación Python desde cero y para principiantes. 75 clases, 37 horas en vídeo, código, proyectos y grupo de chat. Fundamentos, frontend, backend, testing, IA...| | 179|Lightning-AI/pytorch-lightning !2025-03-28292039|Pretrain, finetune and deploy AI models on multiple GPUs, TPUs with zero code changes.| | 180|crewAIInc/crewAI !2025-03-282919344|Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.| | 181|facebook/folly !2025-03-282916612|An open-source C++ library developed and used at Facebook.| | 182|google-ai-edge/mediapipe !2025-03-28291519|Cross-platform, customizable ML solutions for live and streaming media.| | 183| getcursor/cursor !2025-03-282892025 | An editor made for programming with AI| | 184|chatanywhere/GPTAPIfree !2025-03-282856424|Free ChatGPT API Key, Free ChatGPT API, supports GPT-4 API (free), ChatGPT offers a free domestic forwarding API that allows direct connections without the need for a proxy. It can be used in conjunction with software/plugins like ChatBox, significantly reducing interface usage costs. Enjoy unlimited and unrestricted chatting within China| | 185|meta-llama/llama3 !2025-03-28285552|The official Meta Llama 3 GitHub site| | 186|tinygrad/tinygrad !2025-03-282845811|You like pytorch? You like micrograd? You love tinygrad! ❤️| | 187|google-research/tuningplaybook !2025-03-282841514|A playbook for systematically maximizing the performance of deep learning models.| | 188|huggingface/diffusers !2025-03-282830222|🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.| | 189|tokio-rs/tokio !2025-03-28282408|A runtime for writing reliable asynchronous applications with Rust. Provides I/O, networking, scheduling, timers, ...| | 190|RVC-Project/Retrieval-based-Voice-Conversion-WebUI !2025-03-282823817|Voice data !2025-03-282822612|Jan is an open source alternative to ChatGPT that runs 100% offline on your computer| | 192|openai/CLIP !2025-03-282814720|CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image| | 193|🔥khoj-ai/khoj !2025-03-2828112313|Your AI second brain. A copilot to get answers to your questions, whether they be from your own notes or from the internet. Use powerful, online (e.g gpt4) or private, local (e.g mistral) LLMs. Self-host locally or use our web app. Access from Obsidian, Emacs, Desktop app, Web or Whatsapp.| | 194| acheong08/ChatGPT !2025-03-2828054-2 | Reverse engineered ChatGPT API | | 195|iperov/DeepFaceLive !2025-03-28279345 |Real-time face swap for PC streaming or video calls| | 196|eugeneyan/applied-ml !2025-03-28278471|📚 Papers & tech blogs by companies sharing their work on data science & machine learning in production.| | 197|XTLS/Xray-core !2025-03-282778213|Xray, Penetrates Everything. Also the best v2ray-core, with XTLS support. Fully compatible configuration.| | 198|feder-cr/JobsApplierAIAgent !2025-03-282776410|AutoJobsApplierAI_Agent aims to easy job hunt process by automating the job application process. Utilizing artificial intelligence, it enables users to apply for multiple jobs in an automated and personalized way.| | 199|mindsdb/mindsdb !2025-03-282750631|The platform for customizing AI from enterprise data| | 200|DataExpert-io/data-engineer-handbook !2025-03-282721611|This is a repo with links to everything you'd ever want to learn about data engineering| | 201|exo-explore/exo !2025-03-282721633|Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚| | 202|taichi-dev/taichi !2025-03-2826926-1|Productive, portable, and performant GPU programming in Python.| | 203|mem0ai/mem0 !2025-03-282689134|The memory layer for Personalized AI| | 204|svc-develop-team/so-vits-svc !2025-03-28268096 |SoftVC VITS Singing Voice Conversion| | 205|OpenBMB/ChatDev !2025-03-28265624|Create Customized Software using Natural Language Idea (through Multi-Agent Collaboration)| | 206|roboflow/supervision !2025-03-282632010|We write your reusable computer vision tools. 💜| | 207|drawdb-io/drawdb !2025-03-282626913|Free, simple, and intuitive online database design tool and SQL generator.| | 208|karpathy/llm.c !2025-03-28261633|LLM training in simple, raw C/CUDA| | 209|airbnb/lottie-ios !2025-03-28261431|An iOS library to natively render After Effects vector animations| | 210|openai/openai-python !2025-03-282607713|The OpenAI Python library provides convenient access to the OpenAI API from applications written in the Python language.| | 211|academic/awesome-datascience !2025-03-28259876|📝 An awesome Data Science repository to learn and apply for real world problems.| | 212|harry0703/MoneyPrinterTurbo !2025-03-282576618|Generate short videos with one click using a large model| | 213|gabime/spdlog !2025-03-282571511|Fast C++ logging library.| | 214|ocrmypdf/OCRmyPDF !2025-03-2825674217|OCRmyPDF adds an OCR text layer to scanned PDF files, allowing them to be searched| | 215|Vision-CAIR/MiniGPT-4 !2025-03-28256170 |Enhancing Vision-language Understanding with Advanced Large Language Models| | 216|Stability-AI/generative-models !2025-03-28255936|Generative Models by Stability AI| | 217|DS4SD/docling !2025-03-282555662|Get your docs ready for gen AI| | 218|PostHog/posthog !2025-03-282533227|🦔 PostHog provides open-source product analytics, session recording, feature flagging and A/B testing that you can self-host.| | 219|nrwl/nx !2025-03-282509612|Smart Monorepos · Fast CI| | 220|continuedev/continue !2025-03-282500737|⏩ the open-source copilot chat for software development—bring the power of ChatGPT to VS Code| | 221|opentofu/opentofu !2025-03-28247968|OpenTofu lets you declaratively manage your cloud infrastructure.| | 222|invoke-ai/InvokeAI !2025-03-28247293|InvokeAI is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The solution offers an industry leading WebUI, supports terminal use through a CLI, and serves as the foundation for multiple commercial products.| | 223|deepinsight/insightface !2025-03-282471615 |State-of-the-art 2D and 3D Face Analysis Project| | 224|apache/flink !2025-03-28246865|Apache Flink| | 225|ComposioHQ/composio !2025-03-28246436|Composio equips agents with well-crafted tools empowering them to tackle complex tasks| | 226|Genesis-Embodied-AI/Genesis !2025-03-282458314|A generative world for general-purpose robotics & embodied AI learning.| | 227|stretchr/testify !2025-03-28243184|A toolkit with common assertions and mocks that plays nicely with the standard library| | 228| yetone/openai-translator !2025-03-28242921 | Browser extension and cross-platform desktop application for translation based on ChatGPT API | | 229|frappe/erpnext !2025-03-282425211|Free and Open Source Enterprise Resource Planning (ERP)| | 230|songquanpeng/one-api !2025-03-282410034|OpenAI 接口管理 & 分发系统,支持 Azure、Anthropic Claude、Google PaLM 2 & Gemini、智谱 ChatGLM、百度文心一言、讯飞星火认知、阿里通义千问、360 智脑以及腾讯混元,可用于二次分发管理 key,仅单可执行文件,已打包好 Docker 镜像,一键部署,开箱即用. OpenAI key management & redistribution system, using a single API for all LLMs, and features an English UI.| | 231| microsoft/JARVIS !2025-03-28240604 | a system to connect LLMs with ML community | | 232|google/flatbuffers !2025-03-28239965|FlatBuffers: Memory Efficient Serialization Library| | 233|microsoft/graphrag !2025-03-282398928|A modular graph-based Retrieval-Augmented Generation (RAG) system| | 234|rancher/rancher !2025-03-28239675|Complete container management platform| | 235|bazelbuild/bazel !2025-03-282384618|a fast, scalable, multi-language and extensible build system| | 236|modularml/mojo !2025-03-28238236 |The Mojo Programming Language| | 237|danny-avila/LibreChat !2025-03-282378753|Enhanced ChatGPT Clone: Features OpenAI, GPT-4 Vision, Bing, Anthropic, OpenRouter, Google Gemini, AI model switching, message search, langchain, DALL-E-3, ChatGPT Plugins, OpenAI Functions, Secure Multi-User System, Presets, completely open-source for self-hosting. More features in development| |!green-up-arrow.svg 238|🔥🔥🔥Shubhamsaboo/awesome-llm-apps !2025-03-28237391211|Collection of awesome LLM apps with RAG using OpenAI, Anthropic, Gemini and opensource models.| |!red-down-arrow 239|microsoft/semantic-kernel !2025-03-282373611|Integrate cutting-edge LLM technology quickly and easily into your apps| |!red-down-arrow 240|TheAlgorithms/Rust !2025-03-28236995|All Algorithms implemented in Rust| | 241|stanford-oval/storm !2025-03-28236326|An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.| | 242|openai/gpt-2 !2025-03-28232483|Code for the paper "Language Models are Unsupervised Multitask Learners"| | 243|labring/FastGPT !2025-03-282319445|A platform that uses the OpenAI API to quickly build an AI knowledge base, supporting many-to-many relationships.| | 244|pathwaycom/llm-app !2025-03-2822928-10|Ready-to-run cloud templates for RAG, AI pipelines, and enterprise search with live data. 🐳Docker-friendly.⚡Always in sync with Sharepoint, Google Drive, S3, Kafka, PostgreSQL, real-time data APIs, and more.| | 245|warpdotdev/Warp !2025-03-282286825|Warp is a modern, Rust-based terminal with AI built in so you and your team can build great software, faster.| | 246|🔥agno-agi/agno !2025-03-2822833298|Agno is a lightweight library for building Multimodal Agents. It exposes LLMs as a unified API and gives them superpowers like memory, knowledge, tools and reasoning.| | 247|qdrant/qdrant !2025-03-282275214 |Qdrant - Vector Database for the next generation of AI applications. Also available in the cloud https://cloud.qdrant.io/| | 248|ashishpatel26/500-AI-Machine-learning-Deep-learning-Computer-vision-NLP-Projects-with-code !2025-03-282271815|500 AI Machine learning Deep learning Computer vision NLP Projects with code| | 249|stanfordnlp/dspy !2025-03-282268321|Stanford DSPy: The framework for programming—not prompting—foundation models| | 250|PaddlePaddle/Paddle !2025-03-28226246|PArallel Distributed Deep LEarning: Machine Learning Framework from Industrial Practice (『飞桨』核心框架,深度学习&机器学习高性能单机、分布式训练和跨平台部署)| | 251|zulip/zulip !2025-03-28225464|Zulip server and web application. Open-source team chat that helps teams stay productive and focused.| | 252|Hannibal046/Awesome-LLM !2025-03-282240721|Awesome-LLM: a curated list of Large Language Model| | 253|facefusion/facefusion !2025-03-282218812|Next generation face swapper and enhancer| | 254|Mozilla-Ocho/llamafile !2025-03-28220624|Distribute and run LLMs with a single file.| | 255|yuliskov/SmartTube !2025-03-282201614|SmartTube - an advanced player for set-top boxes and tvs running Android OS| | 256|haotian-liu/LLaVA !2025-03-282201316 |Large Language-and-Vision Assistant built towards multimodal GPT-4 level capabilities.| | 257|ashishps1/awesome-system-design-resources !2025-03-282189367|This repository contains System Design resources which are useful while preparing for interviews and learning Distributed Systems| | 258|Cinnamon/kotaemon !2025-03-28218248|An open-source RAG-based tool for chatting with your documents.| | 259|CodePhiliaX/Chat2DB !2025-03-282179757|🔥🔥🔥AI-driven database tool and SQL client, The hottest GUI client, supporting MySQL, Oracle, PostgreSQL, DB2, SQL Server, DB2, SQLite, H2, ClickHouse, and more.| | 260|blakeblackshear/frigate !2025-03-282177113|NVR with realtime local object detection for IP cameras| | 261|facebookresearch/audiocraft !2025-03-28217111|Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable music generation LM with textual and melodic conditioning.| | 262|karpathy/minGPT !2025-03-28216567|A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training| | 263|grpc/grpc-go !2025-03-282159510|The Go language implementation of gRPC. HTTP/2 based RPC| | 264|HumanSignal/label-studio !2025-03-282137618|Label Studio is a multi-type data labeling and annotation tool with standardized output format| | 265|yoheinakajima/babyagi !2025-03-28212764 |uses OpenAI and Pinecone APIs to create, prioritize, and execute tasks, This is a pared-down version of the original Task-Driven Autonomous Agent| | 266|deepseek-ai/DeepSeek-Coder !2025-03-282118210|DeepSeek Coder: Let the Code Write Itself| | 267|BuilderIO/gpt-crawler !2025-03-282118010|Crawl a site to generate knowledge files to create your own custom GPT from a URL| | 268| openai/chatgpt-retrieval-plugin !2025-03-2821152-1 | Plugins are chat extensions designed specifically for language models like ChatGPT, enabling them to access up-to-date information, run computations, or interact with third-party services in response to a user's request.| | 269|microsoft/OmniParser !2025-03-282113123|A simple screen parsing tool towards pure vision based GUI agent| | 270|black-forest-labs/flux !2025-03-282107219|Official inference repo for FLUX.1 models| | 271|ItzCrazyKns/Perplexica !2025-03-282099154|Perplexica is an AI-powered search engine. It is an Open source alternative to Perplexity AI| | 272|microsoft/unilm !2025-03-28209876|Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities| | 273|Sanster/lama-cleaner !2025-03-282077614|Image inpainting tool powered by SOTA AI Model. Remove any unwanted object, defect, people from your pictures or erase and replace(powered by stable diffusion) any thing on your pictures.| | 274|assafelovic/gpt-researcher !2025-03-282057222|GPT based autonomous agent that does online comprehensive research on any given topic| | 275|PromtEngineer/localGPT !2025-03-28204230 |Chat with your documents on your local device using GPT models. No data leaves your device and 100% private.| | 276|elastic/kibana !2025-03-28203482|Your window into the Elastic Stack| | 277|fishaudio/fish-speech !2025-03-282033222|Brand new TTS solution| | 278|mlc-ai/mlc-llm !2025-03-282028110 |Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.| | 279|deepset-ai/haystack !2025-03-282005320|🔍 Haystack is an open source NLP framework to interact with your data using Transformer models and LLMs (GPT-4, ChatGPT and alike). Haystack offers production-ready tools to quickly build complex question answering, semantic search, text generation applications, and more.| | 280|tree-sitter/tree-sitter !2025-03-28200487|An incremental parsing system for programming tools| | 281|Anjok07/ultimatevocalremovergui !2025-03-281999811|GUI for a Vocal Remover that uses Deep Neural Networks.| | 282|guidance-ai/guidance !2025-03-28199622|A guidance language for controlling large language models.| | 283|ml-explore/mlx !2025-03-28199619|MLX: An array framework for Apple silicon| | 284|mlflow/mlflow !2025-03-281995314|Open source platform for the machine learning lifecycle| | 285|ml-tooling/best-of-ml-python !2025-03-28198631|🏆 A ranked list of awesome machine learning Python libraries. Updated weekly.| | 286|BerriAI/litellm !2025-03-281981862|Call all LLM APIs using the OpenAI format. Use Bedrock, Azure, OpenAI, Cohere, Anthropic, Ollama, Sagemaker, HuggingFace, Replicate (100+ LLMs)| | 287|LazyVim/LazyVim !2025-03-281981320|Neovim config for the lazy| | 288|wez/wezterm !2025-03-281976018|A GPU-accelerated cross-platform terminal emulator and multiplexer written by @wez and implemented in Rust| | 289|valkey-io/valkey !2025-03-281970416|A flexible distributed key-value datastore that supports both caching and beyond caching workloads.| | 290|LiLittleCat/awesome-free-chatgpt !2025-03-28196185|🆓免费的 ChatGPT 镜像网站列表,持续更新。List of free ChatGPT mirror sites, continuously updated.| | 291|Byaidu/PDFMathTranslate !2025-03-281947645|PDF scientific paper translation with preserved formats - 基于 AI 完整保留排版的 PDF 文档全文双语翻译,支持 Google/DeepL/Ollama/OpenAI 等服务,提供 CLI/GUI/Docker| | 292|openai/swarm !2025-03-281947111|Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.| | 293|HqWu-HITCS/Awesome-Chinese-LLM !2025-03-281921423|Organizing smaller, cost-effective, privately deployable open-source Chinese language models, including related datasets and tutorials| | 294|stitionai/devika !2025-03-28190903|Devika is an Agentic AI Software Engineer that can understand high-level human instructions, break them down into steps, research relevant information, and write code to achieve the given objective. Devika aims to be a competitive open-source alternative to Devin by Cognition AI.| | 295|OpenBMB/MiniCPM-o !2025-03-28190887|MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone| | 296|samber/lo !2025-03-281904815|💥 A Lodash-style Go library based on Go 1.18+ Generics (map, filter, contains, find...)| | 297|chroma-core/chroma !2025-03-281895221 |the AI-native open-source embedding database| | 298|DarkFlippers/unleashed-firmware !2025-03-28189278|Flipper Zero Unleashed Firmware| | 299|brave/brave-browser !2025-03-281892710|Brave browser for Android, iOS, Linux, macOS, Windows.| | 300| tloen/alpaca-lora !2025-03-28188641 | Instruct-tune LLaMA on consumer hardware| | 301|VinciGit00/Scrapegraph-ai !2025-03-281884618|Python scraper based on AI| | 302|gitroomhq/postiz-app !2025-03-281879110|📨 Schedule social posts, measure them, exchange with other members and get a lot of help from AI 🚀| | 303|PrefectHQ/prefect !2025-03-281878715|Prefect is a workflow orchestration tool empowering developers to build, observe, and react to data pipelines| | 304|ymcui/Chinese-LLaMA-Alpaca !2025-03-28187723 |Chinese LLaMA & Alpaca LLMs| | 305|kenjihiranabe/The-Art-of-Linear-Algebra !2025-03-28187335|Graphic notes on Gilbert Strang's "Linear Algebra for Everyone"| | 306|joonspk-research/generativeagents !2025-03-28187288|Generative Agents: Interactive Simulacra of Human Behavior| | 307|renovatebot/renovate !2025-03-28186820|Universal dependency update tool that fits into your workflows.| | 308|gventuri/pandas-ai !2025-03-28186109 |Pandas AI is a Python library that integrates generative artificial intelligence capabilities into Pandas, making dataframes conversational| | 309|thingsboard/thingsboard !2025-03-28185184|Open-source IoT Platform - Device management, data collection, processing and visualization.| | 310|ente-io/ente !2025-03-28184722|Fully open source, End to End Encrypted alternative to Google Photos and Apple Photos| | 311|serengil/deepface !2025-03-281840113|A Lightweight Face Recognition and Facial Attribute Analysis (Age, Gender, Emotion and Race) Library for Python| | 312|Raphire/Win11Debloat !2025-03-281840132|A simple, easy to use PowerShell script to remove pre-installed apps from windows, disable telemetry, remove Bing from windows search as well as perform various other changes to declutter and improve your windows experience. This script works for both windows 10 and windows 11.| | 313|Avaiga/taipy !2025-03-28179235|Turns Data and AI algorithms into production-ready web applications in no time.| | 314|microsoft/qlib !2025-03-281784231|Qlib is an AI-oriented quantitative investment platform that aims to realize the potential, empower research, and create value using AI technologies in quantitative investment, from exploring ideas to implementing productions. Qlib supports diverse machine learning modeling paradigms. including supervised learning, market dynamics modeling, and RL.| | 315|CopilotKit/CopilotKit !2025-03-281778571|Build in-app AI chatbots 🤖, and AI-powered Textareas ✨, into react web apps.| | 316|QwenLM/Qwen-7B !2025-03-281766017|The official repo of Qwen-7B (通义千问-7B) chat & pretrained large language model proposed by Alibaba Cloud.| | 317|w-okada/voice-changer !2025-03-28176078 |リアルタイムボイスチェンジャー Realtime Voice Changer| | 318|rlabbe/Kalman-and-Bayesian-Filters-in-Python !2025-03-281756011|Kalman Filter book using Jupyter Notebook. Focuses on building intuition and experience, not formal proofs. Includes Kalman filters,extended Kalman filters, unscented Kalman filters, particle filters, and more. All exercises include solutions.| | 319|Mikubill/sd-webui-controlnet !2025-03-28174794 |WebUI extension for ControlNet| | 320|jingyaogong/minimind !2025-03-2817380116|「大模型」3小时完全从0训练26M的小参数GPT,个人显卡即可推理训练!| | 321|apify/crawlee !2025-03-28172696|Crawlee—A web scraping and browser automation library for Node.js to build reliable crawlers. In JavaScript and TypeScript. Extract data for AI, LLMs, RAG, or GPTs. Download HTML, PDF, JPG, PNG, and other files from websites. Works with Puppeteer, Playwright, Cheerio, JSDOM, and raw HTTP. Both headful and headless mode. With proxy rotation.| | 322|apple/ml-stable-diffusion !2025-03-28172395|Stable Diffusion with Core ML on Apple Silicon| | 323| transitive-bullshit/chatgpt-api !2025-03-28172095 | Node.js client for the official ChatGPT API. | | 324|teableio/teable !2025-03-281719222|✨ The Next Gen Airtable Alternative: No-Code Postgres| | 325| xx025/carrot !2025-03-28170900 | Free ChatGPT Site List | | 326|microsoft/LightGBM !2025-03-28170723|A fast, distributed, high-performance gradient boosting (GBT, GBDT, GBRT, GBM or MART) framework based on decision tree algorithms, used for ranking, classification and many other machine learning tasks.| | 327|VikParuchuri/surya !2025-03-28169827|Accurate line-level text detection and recognition (OCR) in any language| | 328|deepseek-ai/Janus !2025-03-281692825|Janus-Series: Unified Multimodal Understanding and Generation Models| | 329|ardalis/CleanArchitecture !2025-03-28168823|Clean Architecture Solution Template: A starting point for Clean Architecture with ASP.NET Core| | 330|neondatabase/neon !2025-03-28166466|Neon: Serverless Postgres. We separated storage and compute to offer autoscaling, code-like database branching, and scale to zero.| | 331|kestra-io/kestra !2025-03-281661313|⚡ Workflow Automation Platform. Orchestrate & Schedule code in any language, run anywhere, 500+ plugins. Alternative to Zapier, Rundeck, Camunda, Airflow...| | 332|Dao-AILab/flash-attention !2025-03-281659720|Fast and memory-efficient exact attention| | 333|RPCS3/rpcs3 !2025-03-281655712|PS3 emulator/debugger| | 334|meta-llama/llama-recipes !2025-03-28165486|Scripts for fine-tuning Llama2 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization & question answering. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment.Demo apps to showcase Llama2 for WhatsApp & Messenger| | 335|emilwallner/Screenshot-to-code !2025-03-28165180|A neural network that transforms a design mock-up into a static website.| | 336|datawhalechina/llm-cookbook !2025-03-281650922|面向开发者的 LLM 入门教程,吴恩达大模型系列课程中文版| | 337|e2b-dev/awesome-ai-agents !2025-03-281643923|A list of AI autonomous agents| | 338|QwenLM/Qwen2.5 !2025-03-281641114|Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.| | 339|dair-ai/ML-YouTube-Courses !2025-03-28164114|📺 Discover the latest machine learning / AI courses on YouTube.| | 340|pybind/pybind11 !2025-03-28163620|Seamless operability between C++11 and Python| | 341|graphdeco-inria/gaussian-splatting !2025-03-281627116|Original reference implementation of "3D Gaussian Splatting for Real-Time Radiance Field Rendering"| | 342|meta-llama/codellama !2025-03-28162531|Inference code for CodeLlama models| | 343|TransformerOptimus/SuperAGI !2025-03-28161292 | SuperAGI - A dev-first open source autonomous AI agent framework. Enabling developers to build, manage & run useful autonomous agents quickly and reliably.| | 344|microsoft/onnxruntime !2025-03-28161169|ONNX Runtime: cross-platform, high-performance ML inferencing and training accelerator| | 345|IDEA-Research/Grounded-Segment-Anything !2025-03-281601411 |Marrying Grounding DINO with Segment Anything & Stable Diffusion & BLIP - Automatically Detect, Segment and Generate Anything with Image and Text Inputs| | 346|ddbourgin/numpy-ml !2025-03-28160054|Machine learning, in numpy| | 347|eosphoros-ai/DB-GPT !2025-03-281585225|Revolutionizing Database Interactions with Private LLM Technology| | 348|Stability-AI/StableLM !2025-03-28158310 |Stability AI Language Models| | 349|openai/evals !2025-03-28157935 |Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.| | 350|THUDM/ChatGLM2-6B !2025-03-28157500|ChatGLM2-6B: An Open Bilingual Chat LLM | | 351|sunner/ChatALL !2025-03-28156761 |Concurrently chat with ChatGPT, Bing Chat, Bard, Alpaca, Vincuna, Claude, ChatGLM, MOSS, iFlytek Spark, ERNIE and more, discover the best answers| | 352|abseil/abseil-cpp !2025-03-28156656|Abseil Common Libraries (C++)| | 353|NVIDIA/open-gpu-kernel-modules !2025-03-28156531|NVIDIA Linux open GPU kernel module source| | 354|letta-ai/letta !2025-03-281563718|Letta (formerly MemGPT) is a framework for creating LLM services with memory.| | 355|typescript-eslint/typescript-eslint !2025-03-28156211|✨ Monorepo for all the tooling which enables ESLint to support TypeScript| | 356|umijs/umi !2025-03-28156211|A framework in react community ✨| | 357|AI4Finance-Foundation/FinGPT !2025-03-281561215|Data-Centric FinGPT. Open-source for open finance! Revolutionize 🔥 We'll soon release the trained model.| | 358|amplication/amplication !2025-03-28156022|🔥🔥🔥 The Only Production-Ready AI-Powered Backend Code Generation| | 359|KindXiaoming/pykan !2025-03-28155477|Kolmogorov Arnold Networks| | 360|arc53/DocsGPT !2025-03-28154900|GPT-powered chat for documentation, chat with your documents| | 361|influxdata/telegraf !2025-03-28154502|Agent for collecting, processing, aggregating, and writing metrics, logs, and other arbitrary data.| | 362|microsoft/Bringing-Old-Photos-Back-to-Life !2025-03-28154084|Bringing Old Photo Back to Life (CVPR 2020 oral)| | 363|GaiZhenbiao/ChuanhuChatGPT !2025-03-2815394-2|GUI for ChatGPT API and many LLMs. Supports agents, file-based QA, GPT finetuning and query with web search. All with a neat UI.| | 364|Zeyi-Lin/HivisionIDPhotos !2025-03-281529710|⚡️HivisionIDPhotos: a lightweight and efficient AI ID photos tools. 一个轻量级的AI证件照制作算法。| | 365| mayooear/gpt4-pdf-chatbot-langchain !2025-03-281529518 | GPT4 & LangChain Chatbot for large PDF docs | | 366|1Panel-dev/MaxKB !2025-03-2815277148|? Based on LLM large language model knowledge base Q&A system. Ready to use out of the box, supports quick integration into third-party business systems. Officially produced by 1Panel| | 367|ai16z/eliza !2025-03-281526811|Conversational Agent for Twitter and Discord| | 368|apache/arrow !2025-03-28151684|Apache Arrow is a multi-language toolbox for accelerated data interchange and in-memory processing| | 369|princeton-nlp/SWE-agent !2025-03-281516119|SWE-agent: Agent Computer Interfaces Enable Software Engineering Language Models| | 370|mlc-ai/web-llm !2025-03-281509311 |Bringing large-language models and chat to web browsers. Everything runs inside the browser with no server support.| | 371|guillaumekln/faster-whisper !2025-03-281507117 |Faster Whisper transcription with CTranslate2| | 372|overleaf/overleaf !2025-03-28150316|A web-based collaborative LaTeX editor| | 373|triton-lang/triton !2025-03-28150169|Development repository for the Triton language and compiler| | 374|soxoj/maigret !2025-03-281500410|🕵️‍♂️ Collect a dossier on a person by username from thousands of sites| | 375|alibaba/lowcode-engine !2025-03-28149841|An enterprise-class low-code technology stack with scale-out design / 一套面向扩展设计的企业级低代码技术体系| | 376|espressif/esp-idf !2025-03-28148545|Espressif IoT Development Framework. Official development framework for Espressif SoCs.| | 377|pgvector/pgvector !2025-03-281484913|Open-source vector similarity search for Postgres| | 378|datawhalechina/leedl-tutorial !2025-03-28148246|《李宏毅深度学习教程》(李宏毅老师推荐👍),PDF下载地址:https://github.com/datawhalechina/leedl-tutorial/releases| | 379|xcanwin/KeepChatGPT !2025-03-28147972 |Using ChatGPT is more efficient and smoother, perfectly solving ChatGPT network errors. No longer do you need to frequently refresh the webpage, saving over 10 unnecessary steps| | 380|m-bain/whisperX !2025-03-281471313|WhisperX: Automatic Speech Recognition with Word-level Timestamps (& Diarization)| | 381|HumanAIGC/AnimateAnyone !2025-03-2814706-1|Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation| |!green-up-arrow.svg 382|naklecha/llama3-from-scratch !2025-03-281469024|llama3 implementation one matrix multiplication at a time| |!red-down-arrow 383| fauxpilot/fauxpilot !2025-03-28146871 | An open-source GitHub Copilot server | | 384|LlamaFamily/Llama-Chinese !2025-03-28145111|Llama Chinese Community, the best Chinese Llama large model, fully open source and commercially available| | 385|BradyFU/Awesome-Multimodal-Large-Language-Models !2025-03-281450121|Latest Papers and Datasets on Multimodal Large Language Models| | 386|vanna-ai/vanna !2025-03-281449819|🤖 Chat with your SQL database 📊. Accurate Text-to-SQL Generation via LLMs using RAG 🔄.| | 387|bleedline/aimoneyhunter !2025-03-28144845|AI Side Hustle Money Mega Collection: Teaching You How to Utilize AI for Various Side Projects to Earn Extra Income.| | 388|stefan-jansen/machine-learning-for-trading !2025-03-28144629|Code for Machine Learning for Algorithmic Trading, 2nd edition.| | 389|state-spaces/mamba !2025-03-28144139|Mamba: Linear-Time Sequence Modeling with Selective State Spaces| | 390|vercel/ai-chatbot !2025-03-281434614|A full-featured, hackable Next.js AI chatbot built by Vercel| | 391|steven-tey/novel !2025-03-281428410|Notion-style WYSIWYG editor with AI-powered autocompletions| | 392|unifyai/ivy !2025-03-281409348|Unified AI| | 393|chidiwilliams/buzz !2025-03-281402411 |Buzz transcribes and translates audio offline on your personal computer. Powered by OpenAI's Whisper.| | 394|lukas-blecher/LaTeX-OCR !2025-03-28139769|pix2tex: Using a ViT to convert images of equations into LaTeX code.| | 395|openai/tiktoken !2025-03-28139599|tiktoken is a fast BPE tokeniser for use with OpenAI's models.| | 396|nocobase/nocobase !2025-03-281391522|NocoBase is a scalability-first, open-source no-code/low-code platform for building business applications and enterprise solutions.| | 397|neonbjb/tortoise-tts !2025-03-28139010 |A multi-voice TTS system trained with an emphasis on quality| | 398|yamadashy/repomix !2025-03-281382036|📦 Repomix (formerly Repopack) is a powerful tool that packs your entire repository into a single, AI-friendly file. Perfect for when you need to feed your codebase to Large Language Models (LLMs) or other AI tools like Claude, ChatGPT, and Gemini.| | 399|adobe/react-spectrum !2025-03-28136766|A collection of libraries and tools that help you build adaptive, accessible, and robust user experiences.| | 400|THUDM/ChatGLM3 !2025-03-28136684|ChatGLM3 series: Open Bilingual Chat LLMs | | 401|NVIDIA/NeMo !2025-03-28134837|A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)| | 402|BlinkDL/RWKV-LM !2025-03-28134346 |RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it combines the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.| | 403| fuergaosi233/wechat-chatgpt !2025-03-28133330 | Use ChatGPT On Wechat via wechaty | | 404|udecode/plate !2025-03-28133325|A rich-text editor powered by AI| | 405|xenova/transformers.js !2025-03-281331219|State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!| | 406|stas00/ml-engineering !2025-03-281325615|Machine Learning Engineering Guides and Tools| | 407| wong2/chatgpt-google-extension !2025-03-2813241-1 | A browser extension that enhances search engines with ChatGPT, this repos will not be updated from 2023-02-20| | 408|mrdbourke/pytorch-deep-learning !2025-03-281317520|Materials for the Learn PyTorch for Deep Learning: Zero to Mastery course.| | 409|Koenkk/zigbee2mqtt !2025-03-28131544|Zigbee 🐝 to MQTT bridge 🌉, get rid of your proprietary Zigbee bridges 🔨| | 410|vercel-labs/ai !2025-03-281298528|Build AI-powered applications with React, Svelte, and Vue| | 411|netease-youdao/QAnything !2025-03-28129318|Question and Answer based on Anything.| | 412|huggingface/trl !2025-03-281289622|Train transformer language models with reinforcement learning.| | 413|microsoft/BitNet !2025-03-28128503|Official inference framework for 1-bit LLMs| | 414|mediar-ai/screenpipe !2025-03-281283915|24/7 local AI screen & mic recording. Build AI apps that have the full context. Works with Ollama. Alternative to Rewind.ai. Open. Secure. You own your data. Rust.| | 415|Skyvern-AI/skyvern !2025-03-281277612|Automate browser-based workflows with LLMs and Computer Vision| | 416|pytube/pytube !2025-03-28126591|A lightweight, dependency-free Python library (and command-line utility) for downloading YouTube Videos.| | 417|official-stockfish/Stockfish !2025-03-28126574|UCI chess engine| | 418|sgl-project/sglang !2025-03-281260143|SGLang is a structured generation language designed for large language models (LLMs). It makes your interaction with LLMs faster and more controllable.| | 419|plasma-umass/scalene !2025-03-28125535|Scalene: a high-performance, high-precision CPU, GPU, and memory profiler for Python with AI-powered optimization proposals| | 420|danswer-ai/danswer !2025-03-28125503|Ask Questions in natural language and get Answers backed by private sources. Connects to tools like Slack, GitHub, Confluence, etc.| | 421|OpenTalker/SadTalker !2025-03-28125226|[CVPR 2023] SadTalker:Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation| | 422|facebookresearch/AnimatedDrawings !2025-03-28123693 |Code to accompany "A Method for Animating Children's Drawings of the Human Figure"| | 423|activepieces/activepieces !2025-03-28123609|Your friendliest open source all-in-one automation tool ✨ Workflow automation tool 100+ integration / Enterprise automation tool / Zapier Alternative| | 424|ggerganov/ggml !2025-03-28121992 |Tensor library for machine learning| | 425|bytebase/bytebase !2025-03-28121694|World's most advanced database DevOps and CI/CD for Developer, DBA and Platform Engineering teams. The GitLab/GitHub for database DevOps.| | 426| willwulfken/MidJourney-Styles-and-Keywords-Reference !2025-03-28120971 | A reference containing Styles and Keywords that you can use with MidJourney AI| | 427|Huanshere/VideoLingo !2025-03-281207013|Netflix-level subtitle cutting, translation, alignment, and even dubbing - one-click fully automated AI video subtitle team | | 428|OpenLMLab/MOSS !2025-03-28120330 |An open-source tool-augmented conversational language model from Fudan University| | 429|llmware-ai/llmware !2025-03-281200727|Providing enterprise-grade LLM-based development framework, tools, and fine-tuned models.| | 430|PKU-YuanGroup/Open-Sora-Plan !2025-03-28119362|This project aim to reproduce Sora (Open AI T2V model), but we only have limited resource. We deeply wish the all open source community can contribute to this project.| | 431|ShishirPatil/gorilla !2025-03-28119332 |Gorilla: An API store for LLMs| | 432|NVIDIA/Megatron-LM !2025-03-281192716|Ongoing research training transformer models at scale| | 433|illacloud/illa-builder !2025-03-28119192|Create AI-Driven Apps like Assembling Blocks| | 434|marimo-team/marimo !2025-03-281191521|A reactive notebook for Python — run reproducible experiments, execute as a script, deploy as an app, and version with git.| | 435|smol-ai/developer !2025-03-28119111 | With 100k context windows on the way, it's now feasible for every dev to have their own smol developer| | 436|Lightning-AI/litgpt !2025-03-28118878|Pretrain, finetune, deploy 20+ LLMs on your own data. Uses state-of-the-art techniques: flash attention, FSDP, 4-bit, LoRA, and more.| | 437|openai/shap-e !2025-03-28118474 |Generate 3D objects conditioned on text or images| | 438|eugeneyan/open-llms !2025-03-28118451 |A list of open LLMs available for commercial use.| | 439|andrewyng/aisuite !2025-03-28118124|Simple, unified interface to multiple Generative AI providers| | 440|hajimehoshi/ebiten !2025-03-28117816|Ebitengine - A dead simple 2D game engine for Go| | 441|kgrzybek/modular-monolith-with-ddd !2025-03-28117493|Full Modular Monolith application with Domain-Driven Design approach.| | 442|h2oai/h2ogpt !2025-03-2811736-1 |Come join the movement to make the world's best open source GPT led by H2O.ai - 100% private chat and document search, no data leaks, Apache 2.0| | 443|owainlewis/awesome-artificial-intelligence !2025-03-28117332|A curated list of Artificial Intelligence (AI) courses, books, video lectures and papers.| | 444|DataTalksClub/mlops-zoomcamp !2025-03-28116643|Free MLOps course from DataTalks.Club| | 445|Rudrabha/Wav2Lip !2025-03-281163410|This repository contains the codes of "A Lip Sync Expert Is All You Need for Speech to Lip Generation In the Wild", published at ACM Multimedia 2020.| | 446|aishwaryanr/awesome-generative-ai-guide !2025-03-281152810|A one stop repository for generative AI research updates, interview resources, notebooks and much more!| | 447|karpathy/micrograd !2025-03-28115146|A tiny scalar-valued autograd engine and a neural net library on top of it with PyTorch-like API| | 448|InstantID/InstantID !2025-03-28115111|InstantID : Zero-shot Identity-Preserving Generation in Seconds 🔥| | 449|facebookresearch/seamlesscommunication !2025-03-28114434|Foundational Models for State-of-the-Art Speech and Text Translation| | 450|anthropics/anthropic-cookbook !2025-03-281140112|A collection of notebooks/recipes showcasing some fun and effective ways of using Claude.| | 451|mastra-ai/mastra !2025-03-281139240|the TypeScript AI agent framework| | 452|NVIDIA/TensorRT !2025-03-28113864|NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.| | 453|plandex-ai/plandex !2025-03-28113645|An AI coding engine for complex tasks| | 454|RUCAIBox/LLMSurvey !2025-03-28112735 |A collection of papers and resources related to Large Language Models.| | 455|kubeshark/kubeshark !2025-03-28112711|The API traffic analyzer for Kubernetes providing real-time K8s protocol-level visibility, capturing and monitoring all traffic and payloads going in, out and across containers, pods, nodes and clusters. Inspired by Wireshark, purposely built for Kubernetes| | 456|electric-sql/pglite !2025-03-28112617|Lightweight Postgres packaged as WASM into a TypeScript library for the browser, Node.js, Bun and Deno from https://electric-sql.com| | 457|lightaime/camel !2025-03-281124441 |🐫 CAMEL: Communicative Agents for “Mind” Exploration of Large Scale Language Model Society| | 458|huggingface/lerobot !2025-03-281120184|🤗 LeRobot: State-of-the-art Machine Learning for Real-World Robotics in Pytorch| | 459|normal-computing/outlines !2025-03-28111657|Generative Model Programming| | 460|libretro/RetroArch !2025-03-28110701|Cross-platform, sophisticated frontend for the libretro API. Licensed GPLv3.| | 461|THUDM/CogVideo !2025-03-28110599|Text-to-video generation: CogVideoX (2024) and CogVideo (ICLR 2023)| | 462|bentoml/OpenLLM !2025-03-28110495|An open platform for operating large language models (LLMs) in production. Fine-tune, serve, deploy, and monitor any LLMs with ease.| | 463|vosen/ZLUDA !2025-03-28110429|CUDA on AMD GPUs| | 464|dair-ai/ML-Papers-of-the-Week !2025-03-28110304 |🔥Highlighting the top ML papers every week.| | 465|WordPress/gutenberg !2025-03-28110212|The Block Editor project for WordPress and beyond. Plugin is available from the official repository.| | 466|microsoft/data-formulator !2025-03-281099827|🪄 Create rich visualizations with AI| | 467|LibreTranslate/LibreTranslate !2025-03-28109887|Free and Open Source Machine Translation API. Self-hosted, offline capable and easy to setup.| | 468|block/goose !2025-03-281097737|an open-source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM| | 469|getumbrel/llama-gpt !2025-03-28109553|A self-hosted, offline, ChatGPT-like chatbot. Powered by Llama 2. 100% private, with no data leaving your device.| | 470|HigherOrderCO/HVM !2025-03-28109182|A massively parallel, optimal functional runtime in Rust| | 471|databrickslabs/dolly !2025-03-2810812-3 | A large language model trained on the Databricks Machine Learning Platform| | 472|srush/GPU-Puzzles !2025-03-28108014|Solve puzzles. Learn CUDA.| | 473|Z3Prover/z3 !2025-03-28107952|The Z3 Theorem Prover| | 474|UFund-Me/Qbot !2025-03-281079313 |Qbot is an AI-oriented quantitative investment platform, which aims to realize the potential, empower AI technologies in quantitative investment| | 475|langchain-ai/langgraph !2025-03-281077336|| | 476|lz4/lz4 !2025-03-28107647|Extremely Fast Compression algorithm| | 477|magic-research/magic-animate !2025-03-28107160|MagicAnimate: Temporally Consistent Human Image Animation using Diffusion Model| | 478|PaperMC/Paper !2025-03-281071410|The most widely used, high performance Minecraft server that aims to fix gameplay and mechanics inconsistencies| | 479|getomni-ai/zerox !2025-03-281071015|Zero shot pdf OCR with gpt-4o-mini| |!green-up-arrow.svg 480|🔥NirDiamant/GenAIAgents !2025-03-2810693318|This repository provides tutorials and implementations for various Generative AI Agent techniques, from basic to advanced. It serves as a comprehensive guide for building intelligent, interactive AI systems.| |!red-down-arrow 481|Unstructured-IO/unstructured !2025-03-28106889|Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.| | 482|apache/thrift !2025-03-28106610|Apache Thrift| | 483| TheR1D/shellgpt !2025-03-28106097 | A command-line productivity tool powered by ChatGPT, will help you accomplish your tasks faster and more efficiently | | 484|TheRamU/Fay !2025-03-281060312 |Fay is a complete open source project that includes Fay controller and numeral models, which can be used in different applications such as virtual hosts, live promotion, numeral human interaction and so on| | 485|zyronon/douyin !2025-03-28105566|Vue3 + Pinia + Vite5 仿抖音,Vue 在移动端的最佳实践 . Imitate TikTok ,Vue Best practices on Mobile| | 486|THU-MIG/yolov10 !2025-03-28105485|YOLOv10: Real-Time End-to-End Object Detection| | 487|idootop/mi-gpt !2025-03-281052522|? Transform XiaoAi speaker into a personal voice assistant with ChatGPT and DouBao integration.| | 488|SakanaAI/AI-Scientist !2025-03-281051310|The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery 🧑‍🔬| | 489|szimek/sharedrop !2025-03-28105101|Easy P2P file transfer powered by WebRTC - inspired by Apple AirDrop| | 490|salesforce/LAVIS !2025-03-28103942 |LAVIS - A One-stop Library for Language-Vision Intelligence| | 491|aws/amazon-sagemaker-examples !2025-03-28103654|Example 📓 Jupyter notebooks that demonstrate how to build, train, and deploy machine learning models using 🧠 Amazon SageMaker.| | 492|artidoro/qlora !2025-03-28103402 |QLoRA: Efficient Finetuning of Quantized LLMs| | 493|lllyasviel/stable-diffusion-webui-forge !2025-03-281029314| a platform on top of Stable Diffusion WebUI (based on Gradio) to make development easier, optimize resource management, and speed up inference| | 494|NielsRogge/Transformers-Tutorials !2025-03-28102487|This repository contains demos I made with the Transformers library by HuggingFace.| | 495|kedro-org/kedro !2025-03-28102371|Kedro is a toolbox for production-ready data science. It uses software engineering best practices to help you create data engineering and data science pipelines that are reproducible, maintainable, and modular.| | 496| chathub-dev/chathub !2025-03-28102301 | All-in-one chatbot client | | 497|microsoft/promptflow !2025-03-28101612|Build high-quality LLM apps - from prototyping, testing to production deployment and monitoring.| | 498|mistralai/mistral-src !2025-03-28101372|Reference implementation of Mistral AI 7B v0.1 model.| | 499|burn-rs/burn !2025-03-28101183|Burn - A Flexible and Comprehensive Deep Learning Framework in Rust| | 500|AIGC-Audio/AudioGPT !2025-03-28101150 |AudioGPT: Understanding and Generating Speech, Music, Sound, and Talking Head| | 501|facebookresearch/dinov2 !2025-03-281011210 |PyTorch code and models for the DINOv2 self-supervised learning method.| | 502|RockChinQ/LangBot !2025-03-281008455|😎丰富生态、🧩支持扩展、🦄多模态 - 大模型原生即时通信机器人平台 🤖 | | 503|78/xiaozhi-esp32 !2025-03-281008180|Build your own AI friend| | 504|cumulo-autumn/StreamDiffusion !2025-03-28100761|StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation| | 505|DataTalksClub/machine-learning-zoomcamp !2025-03-28100664|The code from the Machine Learning Bookcamp book and a free course based on the book| | 506|nerfstudio-project/nerfstudio !2025-03-28100343|A collaboration friendly studio for NeRFs| | 507|cupy/cupy !2025-03-28100344|NumPy & SciPy for GPU| | 508|NVIDIA/TensorRT-LLM !2025-03-281000823|TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.| | 509|wasp-lang/open-saas !2025-03-2899665|A free, open-source SaaS app starter for React & Node.js with superpowers. Production-ready. Community-driven.| | 510|huggingface/text-generation-inference !2025-03-2899383|Large Language Model Text Generation Inference| | 511|jxnl/instructor !2025-03-2899224|structured outputs for llms| | 512|GoogleCloudPlatform/generative-ai !2025-03-2899086|Sample code and notebooks for Generative AI on Google Cloud| | 513|manticoresoftware/manticoresearch !2025-03-2898799|Easy to use open source fast database for search | | 514|langfuse/langfuse !2025-03-28985134|🪢 Open source LLM engineering platform. Observability, metrics, evals, prompt management, testing, prompt playground, datasets, LLM evaluations -- 🍊YC W23 🤖 integrate via Typescript, Python / Decorators, OpenAI, Langchain, LlamaIndex, Litellm, Instructor, Mistral, Perplexity, Claude, Gemini, Vertex| | 515|keephq/keep !2025-03-2897949|The open-source alert management and AIOps platform| | 516|sashabaranov/go-openai !2025-03-2897843|OpenAI ChatGPT, GPT-3, GPT-4, DALL·E, Whisper API wrapper for Go| | 517|autowarefoundation/autoware !2025-03-2897766|Autoware - the world's leading open-source software project for autonomous driving| | 518|anthropics/courses !2025-03-2897269|Anthropic's educational courses| | 519|popcorn-official/popcorn-desktop !2025-03-2896853|Popcorn Time is a multi-platform, free software BitTorrent client that includes an integrated media player ( Windows / Mac / Linux ) A Butter-Project Fork| | 520|getmaxun/maxun !2025-03-28968515|🔥 Open-source no-code web data extraction platform. Turn websites to APIs and spreadsheets with no-code robots in minutes! [In Beta]| | 521|wandb/wandb !2025-03-2896763|🔥 A tool for visualizing and tracking your machine learning experiments. This repo contains the CLI and Python API.| | 522|karpathy/minbpe !2025-03-2895353|Minimal, clean, code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.| | 523|bigscience-workshop/petals !2025-03-2895142|🌸 Run large language models at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading| | 524|OthersideAI/self-operating-computer !2025-03-2894931|A framework to enable multimodal models to operate a computer.| | 525|mshumer/gpt-prompt-engineer !2025-03-2894911|| | 526| BloopAI/bloop !2025-03-2894710 | A fast code search engine written in Rust| | 527|BlinkDL/ChatRWKV !2025-03-289467-1 |ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.| | 528|timlrx/tailwind-nextjs-starter-blog !2025-03-2894677|This is a Next.js, Tailwind CSS blogging starter template. Comes out of the box configured with the latest technologies to make technical writing a breeze. Easily configurable and customizable. Perfect as a replacement to existing Jekyll and Hugo individual blogs.| | 529|google/benchmark !2025-03-2893634|A microbenchmark support library| | 530|facebookresearch/nougat !2025-03-2893603|Implementation of Nougat Neural Optical Understanding for Academic Documents| | 531|modelscope/facechain !2025-03-2893536|FaceChain is a deep-learning toolchain for generating your Digital-Twin.| | 532|DrewThomasson/ebook2audiobook !2025-03-2893388|Convert ebooks to audiobooks with chapters and metadata using dynamic AI models and voice cloning. Supports 1,107+ languages!| | 533|RayTracing/raytracing.github.io !2025-03-2893035|Main Web Site (Online Books)| | 534|QwenLM/Qwen2.5-VL !2025-03-28930249|Qwen2.5-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.| | 535|WongKinYiu/yolov9 !2025-03-2892201|Implementation of paper - YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information| | 536|alibaba-damo-academy/FunASR !2025-03-28920222|A Fundamental End-to-End Speech Recognition Toolkit and Open Source SOTA Pretrained Models.| | 537|Visualize-ML/Book4Power-of-Matrix !2025-03-2891931|Book4 'Power of Matrix' | | 538|dice2o/BingGPT !2025-03-289185-1 |Desktop application of new Bing's AI-powered chat (Windows, macOS and Linux)| | 539|browserbase/stagehand !2025-03-28917621|An AI web browsing framework focused on simplicity and extensibility.| | 540|FlagOpen/FlagEmbedding !2025-03-28914111|Dense Retrieval and Retrieval-augmented LLMs| | 541|Const-me/Whisper !2025-03-2890979|High-performance GPGPU inference of OpenAI's Whisper automatic speech recognition (ASR) model| | 542|lucidrains/denoising-diffusion-pytorch !2025-03-2890942|Implementation of Denoising Diffusion Probabilistic Model in Pytorch| | 543|Chainlit/chainlit !2025-03-28904422|Build Conversational AI in minutes ⚡️| | 544|togethercomputer/OpenChatKit !2025-03-2890160 |OpenChatKit provides a powerful, open-source base to create both specialized and general purpose chatbots for various applications| | 545|Stability-AI/StableStudio !2025-03-2889631 |Community interface for generative AI| | 546|voicepaw/so-vits-svc-fork !2025-03-2889482 |so-vits-svc fork with realtime support, improved interface and more features.| | 547|pymc-devs/pymc !2025-03-2889413|Bayesian Modeling and Probabilistic Programming in Python| | 548|espnet/espnet !2025-03-2889302|End-to-End Speech Processing Toolkit| | 549|kedacore/keda !2025-03-2888991|KEDA is a Kubernetes-based Event Driven Autoscaling component. It provides event driven scale for any container running in Kubernetes| | 550|open-mmlab/Amphion !2025-03-28886911|Amphion (/æmˈfaɪən/) is a toolkit for Audio, Music, and Speech Generation. Its purpose is to support reproducible research and help junior researchers and engineers get started in the field of audio, music, and speech generation research and development.| | 551|gorse-io/gorse !2025-03-2888451|Gorse open source recommender system engine| | 552|adams549659584/go-proxy-bingai !2025-03-288768-1 |A Microsoft New Bing demo site built with Vue3 and Go, providing a consistent UI experience, supporting ChatGPT prompts, and accessible within China| | 553|open-mmlab/mmsegmentation !2025-03-2887513|OpenMMLab Semantic Segmentation Toolbox and Benchmark.| | 554|bytedance/monolith !2025-03-2887223|ByteDance's Recommendation System| | 555|LouisShark/chatgptsystemprompt !2025-03-2887216|store all agent's system prompt| | 556|brexhq/prompt-engineering !2025-03-2887080 |Tips and tricks for working with Large Language Models like OpenAI's GPT-4.| | 557|erincatto/box2d !2025-03-2886841|Box2D is a 2D physics engine for games| | 558|🔥microsoft/ai-agents-for-beginners !2025-03-288669323|10 Lessons to Get Started Building AI Agents| | 559|nashsu/FreeAskInternet !2025-03-2886102|FreeAskInternet is a completely free, private and locally running search aggregator & answer generate using LLM, without GPU needed. The user can ask a question and the system will make a multi engine search and combine the search result to the ChatGPT3.5 LLM and generate the answer based on search results.| | 560|goldmansachs/gs-quant !2025-03-2885981|Python toolkit for quantitative finance| | 561|srbhr/Resume-Matcher !2025-03-2885800|Open Source Free ATS Tool to compare Resumes with Job Descriptions and create a score to rank them.| | 562|facebookresearch/ImageBind !2025-03-2885681 |ImageBind One Embedding Space to Bind Them All| | 563|ashawkey/stable-dreamfusion !2025-03-2885481 |A pytorch implementation of text-to-3D dreamfusion, powered by stable diffusion.| | 564|meetecho/janus-gateway !2025-03-2885232|Janus WebRTC Server| | 565|google/magika !2025-03-2885003|Detect file content types with deep learning| | 566|huggingface/chat-ui !2025-03-2884871 |Open source codebase powering the HuggingChat app| | 567|EleutherAI/lm-evaluation-harness !2025-03-28843012|A framework for few-shot evaluation of autoregressive language models.| | 568|jina-ai/reader !2025-03-2884089|Convert any URL to an LLM-friendly input with a simple prefix https://r.jina.ai/| | 569|microsoft/TypeChat !2025-03-288406-1|TypeChat is a library that makes it easy to build natural language interfaces using types.| | 570|thuml/Time-Series-Library !2025-03-28839715|A Library for Advanced Deep Time Series Models.| | 571|OptimalScale/LMFlow !2025-03-2883882|An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Model for All.| | 572|baptisteArno/typebot.io !2025-03-2883845|💬 Typebot is a powerful chatbot builder that you can self-host.| | 573|jzhang38/TinyLlama !2025-03-2883504|The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.| | 574|fishaudio/Bert-VITS2 !2025-03-2883472|vits2 backbone with multilingual-bert| | 575|OpenBMB/XAgent !2025-03-2882683|An Autonomous LLM Agent for Complex Task Solving| | 576|Acly/krita-ai-diffusion !2025-03-2882387|Streamlined interface for generating images with AI in Krita. Inpaint and outpaint with optional text prompt, no tweaking required.| | 577|jasonppy/VoiceCraft !2025-03-2882151|Zero-Shot Speech Editing and Text-to-Speech in the Wild| | 578|SJTU-IPADS/PowerInfer !2025-03-2881693|High-speed Large Language Model Serving on PCs with Consumer-grade GPUs| | 579|modelscope/DiffSynth-Studio !2025-03-28814713|Enjoy the magic of Diffusion models!| | 580|o3de/o3de !2025-03-2881443|Open 3D Engine (O3DE) is an Apache 2.0-licensed multi-platform 3D engine that enables developers and content creators to build AAA games, cinema-quality 3D worlds, and high-fidelity simulations without any fees or commercial obligations.| | 581|zmh-program/chatnio !2025-03-2881325|🚀 Next Generation AI One-Stop Internationalization Solution. 🚀 下一代 AI 一站式 B/C 端解决方案,支持 OpenAI,Midjourney,Claude,讯飞星火,Stable Diffusion,DALL·E,ChatGLM,通义千问,腾讯混元,360 智脑,百川 AI,火山方舟,新必应,Gemini,Moonshot 等模型,支持对话分享,自定义预设,云端同步,模型市场,支持弹性计费和订阅计划模式,支持图片解析,支持联网搜索,支持模型缓存,丰富美观的后台管理与仪表盘数据统计。| | 582|leptonai/searchwithlepton !2025-03-2880632|Building a quick conversation-based search demo with Lepton AI.| | 583|sebastianstarke/AI4Animation !2025-03-2880620|Bringing Characters to Life with Computer Brains in Unity| | 584|wangrongding/wechat-bot !2025-03-2880528|🤖一个基于 WeChaty 结合 DeepSeek / ChatGPT / Kimi / 讯飞等Ai服务实现的微信机器人 ,可以用来帮助你自动回复微信消息,或者管理微信群/好友,检测僵尸粉等...| | 585|openvinotoolkit/openvino !2025-03-2880528|OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference| | 586|steven2358/awesome-generative-ai !2025-03-28802610|A curated list of modern Generative Artificial Intelligence projects and services| | 587|adam-maj/tiny-gpu !2025-03-2880234|A minimal GPU design in Verilog to learn how GPUs work from the ground up| | 588| anse-app/chatgpt-demo !2025-03-2880180 | A demo repo based on OpenAI API (gpt-3.5-turbo) | | 589| acheong08/EdgeGPT !2025-03-288015-1 |Reverse engineered API of Microsoft's Bing Chat | | 590|ai-collection/ai-collection !2025-03-2879994 |The Generative AI Landscape - A Collection of Awesome Generative AI Applications| | 591|GreyDGL/PentestGPT !2025-03-2879953 |A GPT-empowered penetration testing tool| | 592|delta-io/delta !2025-03-2879112|An open-source storage framework that enables building a Lakehouse architecture with compute engines including Spark, PrestoDB, Flink, Trino, and Hive and APIs| | 593|dataelement/bisheng !2025-03-2879085|Bisheng is an open LLM devops platform for next generation AI applications.| | 594|e2b-dev/e2b !2025-03-2878447 |Vercel for AI agents. We help developers to build, deploy, and monitor AI agents. Focusing on specialized AI agents that build software for you - your personal software developers.| | 595|01-ai/Yi !2025-03-2878311|A series of large language models trained from scratch by developers @01-ai| | 596|Plachtaa/VALL-E-X !2025-03-287830-1|An open source implementation of Microsoft's VALL-E X zero-shot TTS model. The demo is available at https://plachtaa.github.io| | 597|abhishekkrthakur/approachingalmost !2025-03-2878204|Approaching (Almost) Any Machine Learning Problem| | 598|pydantic/pydantic-ai !2025-03-28781041|Agent Framework / shim to use Pydantic with LLMs| | 599|rany2/edge-tts !2025-03-2877901|Use Microsoft Edge's online text-to-speech service from Python WITHOUT needing Microsoft Edge or Windows or an API key| | 600|CASIA-IVA-Lab/FastSAM !2025-03-2877881|Fast Segment Anything| | 601|netease-youdao/EmotiVoice !2025-03-2877817|EmotiVoice 😊: a Multi-Voice and Prompt-Controlled TTS Engine| | 602|lllyasviel/IC-Light !2025-03-2877804|More relighting!| | 603|kroma-network/tachyon !2025-03-287774-1|Modular ZK(Zero Knowledge) backend accelerated by GPU| | 604|deep-floyd/IF !2025-03-2877731 |A novel state-of-the-art open-source text-to-image model with a high degree of photorealism and language understanding| | 605|oumi-ai/oumi !2025-03-2877705|Everything you need to build state-of-the-art foundation models, end-to-end.| | 606|reorproject/reor !2025-03-2877681|AI note-taking app that runs models locally.| | 607|lightpanda-io/browser !2025-03-28775813|Lightpanda: the headless browser designed for AI and automation| | 608|xiangsx/gpt4free-ts !2025-03-287755-1|Providing a free OpenAI GPT-4 API ! This is a replication project for the typescript version of xtekky/gpt4free| | 609|IDEA-Research/GroundingDINO !2025-03-28773311|Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"| | 610|bunkerity/bunkerweb !2025-03-2877326|🛡️ Make your web services secure by default !| | 611|vikhyat/moondream !2025-03-2877057|tiny vision language model| | 612|firmai/financial-machine-learning !2025-03-287703-1|A curated list of practical financial machine learning tools and applications.| | 613|n8n-io/self-hosted-ai-starter-kit !2025-03-28765121|The Self-hosted AI Starter Kit is an open-source template that quickly sets up a local AI environment. Curated by n8n, it provides essential tools for creating secure, self-hosted AI workflows.| | 614|intel-analytics/ipex-llm !2025-03-2876507|Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with llama.cpp, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, ModelScope, etc.| | 615|jrouwe/JoltPhysics !2025-03-28764510|A multi core friendly rigid body physics and collision detection library. Written in C++. Suitable for games and VR applications. Used by Horizon Forbidden West.| | 616|THUDM/CodeGeeX2 !2025-03-2876270|CodeGeeX2: A More Powerful Multilingual Code Generation Model| | 617|meta-llama/llama-stack !2025-03-2875866|Composable building blocks to build Llama Apps| | 618|sweepai/sweep !2025-03-287530-1|Sweep is an AI junior developer| | 619|lllyasviel/Omost !2025-03-2875301|Your image is almost there!| | 620|ahmedbahaaeldin/From-0-to-Research-Scientist-resources-guide !2025-03-2875050|Detailed and tailored guide for undergraduate students or anybody want to dig deep into the field of AI with solid foundation.| | 621|dair-ai/ML-Papers-Explained !2025-03-2875050|Explanation to key concepts in ML| | 622|zaidmukaddam/scira !2025-03-28750110|Scira (Formerly MiniPerplx) is a minimalistic AI-powered search engine that helps you find information on the internet. Powered by Vercel AI SDK! Search with models like Grok 2.0.| | 623|Portkey-AI/gateway !2025-03-28749416|A Blazing Fast AI Gateway. Route to 100+ LLMs with 1 fast & friendly API.| | 624|web-infra-dev/midscene !2025-03-28748729|An AI-powered automation SDK can control the page, perform assertions, and extract data in JSON format using natural language.| | 625|zilliztech/GPTCache !2025-03-2874801 |GPTCache is a library for creating semantic cache to store responses from LLM queries.| | 626|niedev/RTranslator !2025-03-2874742|RTranslator is the world's first open source real-time translation app.| |!green-up-arrow.svg 627|roboflow/notebooks !2025-03-2874666|Examples and tutorials on using SOTA computer vision models and techniques. Learn everything from old-school ResNet, through YOLO and object-detection transformers like DETR, to the latest models like Grounding DINO and SAM.| |!red-down-arrow 628|openlm-research/openllama !2025-03-2874652|OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset| | 629|LiheYoung/Depth-Anything !2025-03-2874155|Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data| | 630|enso-org/enso !2025-03-2874040|Hybrid visual and textual functional programming.| | 631|bigcode-project/starcoder !2025-03-287401-1 |Home of StarCoder: fine-tuning & inference!| | 632|git-ecosystem/git-credential-manager !2025-03-2873975|Secure, cross-platform Git credential storage with authentication to GitHub, Azure Repos, and other popular Git hosting services.| | 633|OpenGVLab/InternVL !2025-03-2873634|[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4V. 接近GPT-4V表现的可商用开源模型| | 634|WooooDyy/LLM-Agent-Paper-List !2025-03-2873551|The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.| | 635|lencx/Noi !2025-03-2873157|🦄 AI + Tools + Plugins + Community| | 636|udlbook/udlbook !2025-03-2873075|Understanding Deep Learning - Simon J.D. Prince| | 637|OpenBMB/MiniCPM !2025-03-2872841|MiniCPM-2B: An end-side LLM outperforms Llama2-13B.| | 638|jaywalnut310/vits !2025-03-2872815 |VITS: Conditional Variational Autoencoder with Adversarial Learning for End-to-End Text-to-Speech| | 639|xorbitsai/inference !2025-03-28727528|Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.| | 640|PWhiddy/PokemonRedExperiments !2025-03-2872492|Playing Pokemon Red with Reinforcement Learning| | 641|Canner/WrenAI !2025-03-28723213|🤖 Open-source AI Agent that empowers data-driven teams to chat with their data to generate Text-to-SQL, charts, spreadsheets, reports, and BI. 📈📊📋🧑‍💻| | 642|miurla/morphic !2025-03-2872258|An AI-powered answer engine with a generative UI| | 643|ml-explore/mlx-examples !2025-03-2872168|Examples in the MLX framework| | 644|PKU-YuanGroup/ChatLaw !2025-03-2872010|Chinese Legal Large Model| | 645|NVIDIA/cutlass !2025-03-2871883|CUDA Templates for Linear Algebra Subroutines| | 646|FoundationVision/VAR !2025-03-28717444|[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction"| | 647|ymcui/Chinese-LLaMA-Alpaca-2 !2025-03-2871561|Chinese LLaMA-2 & Alpaca-2 LLMs| | 648|nadermx/backgroundremover !2025-03-2871514 |Background Remover lets you Remove Background from images and video using AI with a simple command line interface that is free and open source.| | 649|onuratakan/gpt-computer-assistant !2025-03-28714514|gpt-4o for windows, macos and ubuntu| | 650|graviraja/MLOps-Basics !2025-03-2871326|| | 651|Future-House/paper-qa !2025-03-287118-1|High accuracy RAG for answering questions from scientific documents with citations| | 652|open-mmlab/mmagic !2025-03-2871102 |OpenMMLab Multimodal Advanced, Generative, and Intelligent Creation Toolbox| | 653|bhaskatripathi/pdfGPT !2025-03-2870941 |PDF GPT allows you to chat with the contents of your PDF file by using GPT capabilities. The only open source solution to turn your pdf files in a chatbot!| | 654|ollama/ollama-python !2025-03-28709117|Ollama Python library| | 655|facebookresearch/DiT !2025-03-2870376|Official PyTorch Implementation of "Scalable Diffusion Models with Transformers"| | 656|geekyutao/Inpaint-Anything !2025-03-2870262 |Inpaint anything using Segment Anything and inpainting models.| | 657|AbdullahAlfaraj/Auto-Photoshop-StableDiffusion-Plugin !2025-03-2870160 |A user-friendly plug-in that makes it easy to generate stable diffusion images inside Photoshop using Automatic1111-sd-webui as a backend.| | 658|apple/corenet !2025-03-2869990|CoreNet: A library for training deep neural networks| | 659|openstatusHQ/openstatus !2025-03-2869926|🏓 The open-source synthetic monitoring platform 🏓| | 660|weaviate/Verba !2025-03-2869772|Retrieval Augmented Generation (RAG) chatbot powered by Weaviate| | 661|meshery/meshery !2025-03-2869630|Meshery, the cloud native manager| | 662|OpenTalker/video-retalking !2025-03-2869530|[SIGGRAPH Asia 2022] VideoReTalking: Audio-based Lip Synchronization for Talking Head Video Editing In the Wild| | 663|digitalinnovationone/dio-lab-open-source !2025-03-28689013|Repositório do lab "Contribuindo em um Projeto Open Source no GitHub" da Digital Innovation One.| | 664|jianchang512/ChatTTS-ui !2025-03-2868842|一个简单的本地网页界面,直接使用ChatTTS将文字合成为语音,同时支持对外提供API接口。| | 665|patchy631/ai-engineering-hub !2025-03-28686434|In-depth tutorials on LLMs, RAGs and real-world AI agent applications.| | 666|gunnarmorling/1brc !2025-03-2868512|1️⃣🐝🏎️ The One Billion Row Challenge -- A fun exploration of how quickly 1B rows from a text file can be aggregated with Java| | 667|Azure-Samples/azure-search-openai-demo !2025-03-2868482 |A sample app for the Retrieval-Augmented Generation pattern running in Azure, using Azure Cognitive Search for retrieval and Azure OpenAI large language models to power ChatGPT-style and Q&A experiences.| | 668|mit-han-lab/streaming-llm !2025-03-2868382|Efficient Streaming Language Models with Attention Sinks| | 669|InternLM/InternLM !2025-03-2868352|InternLM has open-sourced a 7 billion parameter base model, a chat model tailored for practical scenarios and the training system.| | 670|dependency-check/DependencyCheck !2025-03-2868191|OWASP dependency-check is a software composition analysis utility that detects publicly disclosed vulnerabilities in application dependencies.| | 671|Soulter/AstrBot !2025-03-28678643|✨易上手的多平台 LLM 聊天机器人及开发框架✨。支持 QQ、QQ频道、Telegram、微信平台(Gewechat, 企业微信)、内置 Web Chat,OpenAI GPT、DeepSeek、Ollama、Llama、GLM、Gemini、OneAPI、LLMTuner,支持 LLM Agent 插件开发,可视化面板。一键部署。支持 Dify 工作流、代码执行器、Whisper 语音转文字。| | 672|react-native-webview/react-native-webview !2025-03-2867792|React Native Cross-Platform WebView| | 673|modelscope/agentscope !2025-03-28676916|Start building LLM-empowered multi-agent applications in an easier way.| | 674|mylxsw/aidea !2025-03-2867381|AIdea is a versatile app that supports GPT and domestic large language models,also supports "Stable Diffusion" text-to-image generation, image-to-image generation, SDXL 1.0, super-resolution, and image colorization| | 675|langchain-ai/ollama-deep-researcher !2025-03-28668635|Fully local web research and report writing assistant| | 676|threestudio-project/threestudio !2025-03-2866653|A unified framework for 3D content generation.| | 677|gaomingqi/Track-Anything !2025-03-2866631 |A flexible and interactive tool for video object tracking and segmentation, based on Segment Anything, XMem, and E2FGVI.| | 678|spdustin/ChatGPT-AutoExpert !2025-03-2866570|🚀🧠💬 Supercharged Custom Instructions for ChatGPT (non-coding) and ChatGPT Advanced Data Analysis (coding).| | 679|HariSekhon/DevOps-Bash-tools !2025-03-2866463|1000+ DevOps Bash Scripts - AWS, GCP, Kubernetes, Docker, CI/CD, APIs, SQL, PostgreSQL, MySQL, Hive, Impala, Kafka, Hadoop, Jenkins, GitHub, GitLab, BitBucket, Azure DevOps, TeamCity, Spotify, MP3, LDAP, Code/Build Linting, pkg mgmt for Linux, Mac, Python, Perl, Ruby, NodeJS, Golang, Advanced dotfiles: .bashrc, .vimrc, .gitconfig, .screenrc, tmux..| | 680|modelscope/swift !2025-03-28661530|ms-swift: Use PEFT or Full-parameter to finetune 200+ LLMs or 15+ MLLMs| | 681|langchain-ai/opengpts !2025-03-2866080|This is an open source effort to create a similar experience to OpenAI's GPTs and Assistants API| | 682| yihong0618/xiaogpt !2025-03-2865131 | Play ChatGPT with xiaomi ai speaker | | 683| civitai/civitai !2025-03-2865111 | Build a platform where people can share their stable diffusion models | | 684|KoljaB/RealtimeSTT !2025-03-28649513|A robust, efficient, low-latency speech-to-text library with advanced voice activity detection, wake word activation and instant transcription.| | 685|qunash/chatgpt-advanced !2025-03-2864910 | A browser extension that augments your ChatGPT prompts with web results.| | 686|Licoy/ChatGPT-Midjourney !2025-03-2864850|🎨 Own your own ChatGPT+Midjourney web service with one click| | 687|friuns2/BlackFriday-GPTs-Prompts !2025-03-2864744|List of free GPTs that doesn't require plus subscription| | 688|PixarAnimationStudios/OpenUSD !2025-03-2864700|Universal Scene Description| | 689|linyiLYi/street-fighter-ai !2025-03-2864630 |This is an AI agent for Street Fighter II Champion Edition.| | 690|run-llama/rags !2025-03-2864380|Build ChatGPT over your data, all with natural language| | 691|frdel/agent-zero !2025-03-2864154|Agent Zero AI framework| | 692|microsoft/DeepSpeedExamples !2025-03-2863911 |Example models using DeepSpeed| | 693|k8sgpt-ai/k8sgpt !2025-03-2863882|Giving Kubernetes Superpowers to everyone| | 694|open-metadata/OpenMetadata !2025-03-2863514|OpenMetadata is a unified platform for discovery, observability, and governance powered by a central metadata repository, in-depth lineage, and seamless team collaboration.| | 695|google/gemma.cpp !2025-03-2863163|lightweight, standalone C++ inference engine for Google's Gemma models.| | 696|RayVentura/ShortGPT !2025-03-286314-1|🚀🎬 ShortGPT - An experimental AI framework for automated short/video content creation. Enables creators to rapidly produce, manage, and deliver content using AI and automation.| | 697|openai/consistencymodels !2025-03-2862940 |Official repo for consistency models.| | 698|yangjianxin1/Firefly !2025-03-2862924|Firefly: Chinese conversational large language model (full-scale fine-tuning + QLoRA), supporting fine-tuning of Llma2, Llama, Baichuan, InternLM, Ziya, Bloom, and other large models| | 699|enricoros/big-AGI !2025-03-2862665|Generative AI suite powered by state-of-the-art models and providing advanced AI/AGI functions. It features AI personas, AGI functions, multi-model chats, text-to-image, voice, response streaming, code highlighting and execution, PDF import, presets for developers, much more. Deploy on-prem or in the cloud.| | 700|aptos-labs/aptos-core !2025-03-2862633|Aptos is a layer 1 blockchain built to support the widespread use of blockchain through better technology and user experience.| | 701|wenda-LLM/wenda !2025-03-286262-1 |Wenda: An LLM invocation platform. Its objective is to achieve efficient content generation tailored to specific environments while considering the limited computing resources of individuals and small businesses, as well as knowledge security and privacy concerns| | 702|Project-MONAI/MONAI !2025-03-2862603|AI Toolkit for Healthcare Imaging| | 703|HVision-NKU/StoryDiffusion !2025-03-2862470|Create Magic Story!| | 704|deepseek-ai/DeepSeek-LLM !2025-03-2862463|DeepSeek LLM: Let there be answers| | 705|Tohrusky/Final2x !2025-03-2862393|2^x Image Super-Resolution| | 706|OpenSPG/KAG !2025-03-28619611|KAG is a logical form-guided reasoning and retrieval framework based on OpenSPG engine and LLMs. It is used to build logical reasoning and factual Q&A solutions for professional domain knowledge bases. It can effectively overcome the shortcomings of the traditional RAG vector similarity calculation model.| | 707|Moonvy/OpenPromptStudio !2025-03-2861861 |AIGC Hint Word Visualization Editor| | 708|levihsu/OOTDiffusion !2025-03-2861761|Official implementation of OOTDiffusion| | 709|tmc/langchaingo !2025-03-2861729|LangChain for Go, the easiest way to write LLM-based programs in Go| | 710|vladmandic/automatic !2025-03-2861374|SD.Next: Advanced Implementation of Stable Diffusion and other Diffusion-based generative image models| | 711|clovaai/donut !2025-03-2861231 |Official Implementation of OCR-free Document Understanding Transformer (Donut) and Synthetic Document Generator (SynthDoG), ECCV 2022| | 712|Shaunwei/RealChar !2025-03-286121-1|🎙️🤖Create, Customize and Talk to your AI Character/Companion in Realtime(All in One Codebase!). Have a natural seamless conversation with AI everywhere(mobile, web and terminal) using LLM OpenAI GPT3.5/4, Anthropic Claude2, Chroma Vector DB, Whisper Speech2Text, ElevenLabs Text2Speech🎙️🤖| | 713|microsoft/TinyTroupe !2025-03-2861142|LLM-powered multiagent persona simulation for imagination enhancement and business insights.| | 714| rustformers/llm !2025-03-2861010 | Run inference for Large Language Models on CPU, with Rust| | 715|firebase/firebase-ios-sdk !2025-03-2860950|Firebase SDK for Apple App Development| | 716|vespa-engine/vespa !2025-03-2860824|The open big data serving engine. https://vespa.ai| | 717|n4ze3m/page-assist !2025-03-28607610|Use your locally running AI models to assist you in your web browsing| | 718|Dooy/chatgpt-web-midjourney-proxy !2025-03-2860646|chatgpt web, midjourney, gpts,tts, whisper 一套ui全搞定| | 719|ethereum-optimism/optimism !2025-03-2860213|Optimism is Ethereum, scaled.| | 720|sczhou/ProPainter !2025-03-2859971|[ICCV 2023] ProPainter: Improving Propagation and Transformer for Video Inpainting| | 721|MineDojo/Voyager !2025-03-2859951 |An Open-Ended Embodied Agent with Large Language Models| | 722|lavague-ai/LaVague !2025-03-2859800|Automate automation with Large Action Model framework| | 723|SevaSk/ecoute !2025-03-2859770 |Ecoute is a live transcription tool that provides real-time transcripts for both the user's microphone input (You) and the user's speakers output (Speaker) in a textbox. It also generates a suggested response using OpenAI's GPT-3.5 for the user to say based on the live transcription of the conversation.| | 724|google/mesop !2025-03-2859661|| | 725|pengxiao-song/LaWGPT !2025-03-2859542 |Repo for LaWGPT, Chinese-Llama tuned with Chinese Legal knowledge| | 726|fr0gger/Awesome-GPT-Agents !2025-03-2859434|A curated list of GPT agents for cybersecurity| | 727|google-deepmind/graphcast !2025-03-2859412|| | 728|comet-ml/opik !2025-03-28594126|Open-source end-to-end LLM Development Platform| | 729|SciPhi-AI/R2R !2025-03-28594033|A framework for rapid development and deployment of production-ready RAG systems| | 730|SkalskiP/courses !2025-03-2859272 |This repository is a curated collection of links to various courses and resources about Artificial Intelligence (AI)| | 731|QuivrHQ/MegaParse !2025-03-2859122|File Parser optimised for LLM Ingestion with no loss 🧠 Parse PDFs, Docx, PPTx in a format that is ideal for LLMs.| | 732|pytorch-labs/gpt-fast !2025-03-2858971|Simple and efficient pytorch-native transformer text generation in !2025-03-2858886|Curated list of chatgpt prompts from the top-rated GPTs in the GPTs Store. Prompt Engineering, prompt attack & prompt protect. Advanced Prompt Engineering papers.| | 734|nilsherzig/LLocalSearch !2025-03-2858852|LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a chain of LLMs to find the answer. The user can see the progress of the agents and the final answer. No OpenAI or Google API keys are needed.| | 735|kuafuai/DevOpsGPT !2025-03-285874-2|Multi agent system for AI-driven software development. Convert natural language requirements into working software. Supports any development language and extends the existing base code.| | 736|myshell-ai/MeloTTS !2025-03-2858486|High-quality multi-lingual text-to-speech library by MyShell.ai. Support English, Spanish, French, Chinese, Japanese and Korean.| | 737|OpenGVLab/LLaMA-Adapter !2025-03-2858421 |Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters| | 738|volcengine/verl !2025-03-28582563|veRL: Volcano Engine Reinforcement Learning for LLM| | 739|a16z-infra/companion-app !2025-03-2858171|AI companions with memory: a lightweight stack to create and host your own AI companions| | 740|HumanAIGC/OutfitAnyone !2025-03-285816-1|Outfit Anyone: Ultra-high quality virtual try-on for Any Clothing and Any Person| | 741|josStorer/RWKV-Runner !2025-03-2857472|A RWKV management and startup tool, full automation, only 8MB. And provides an interface compatible with the OpenAI API. RWKV is a large language model that is fully open source and available for commercial use.| | 742|648540858/wvp-GB28181-pro !2025-03-2857414|WEB VIDEO PLATFORM是一个基于GB28181-2016标准实现的网络视频平台,支持NAT穿透,支持海康、大华、宇视等品牌的IPC、NVR、DVR接入。支持国标级联,支持rtsp/rtmp等视频流转发到国标平台,支持rtsp/rtmp等推流转发到国标平台。| | 743|ToonCrafter/ToonCrafter !2025-03-2857345|a research paper for generative cartoon interpolation| | 744|PawanOsman/ChatGPT !2025-03-2857191|OpenAI API Free Reverse Proxy| | 745|apache/hudi !2025-03-2857091|Upserts, Deletes And Incremental Processing on Big Data.| | 746| nsarrazin/serge !2025-03-2857081 | A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API| | 747|homanp/superagent !2025-03-2857021|🥷 Superagent - Build, deploy, and manage LLM-powered agents| | 748|ramonvc/freegpt-webui !2025-03-2856910|GPT 3.5/4 with a Chat Web UI. No API key is required.| | 749|baichuan-inc/baichuan-7B !2025-03-2856901|A large-scale 7B pretraining language model developed by BaiChuan-Inc.| | 750|Azure/azure-sdk-for-net !2025-03-2856792|This repository is for active development of the Azure SDK for .NET. For consumers of the SDK we recommend visiting our public developer docs at https://learn.microsoft.com/dotnet/azure/ or our versioned developer docs at https://azure.github.io/azure-sdk-for-net.| | 751|mnotgod96/AppAgent !2025-03-2856643|AppAgent: Multimodal Agents as Smartphone Users, an LLM-based multimodal agent framework designed to operate smartphone apps.| | 752|microsoft/TaskWeaver !2025-03-2856243|A code-first agent framework for seamlessly planning and executing data analytics tasks.| | 753| yetone/bob-plugin-openai-translator !2025-03-285600-1 | A Bob Plugin base ChatGPT API | | 754|PrefectHQ/marvin !2025-03-2855840 |A batteries-included library for building AI-powered software| | 755|microsoft/promptbase !2025-03-2855832|All things prompt engineering| | 756|fullstackhero/dotnet-starter-kit !2025-03-2855560|Production Grade Cloud-Ready .NET 8 Starter Kit (Web API + Blazor Client) with Multitenancy Support, and Clean/Modular Architecture that saves roughly 200+ Development Hours! All Batteries Included.| | 757|deepseek-ai/DeepSeek-Coder-V2 !2025-03-2855435|DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence| | 758|aiwaves-cn/agents !2025-03-2855391|An Open-source Framework for Autonomous Language Agents| | 759|microsoft/Mastering-GitHub-Copilot-for-Paired-Programming !2025-03-2855158|A 6 Lesson course teaching everything you need to know about harnessing GitHub Copilot and an AI Paired Programing resource.| | 760|allenai/OLMo !2025-03-2854506|Modeling, training, eval, and inference code for OLMo| | 761|apify/crawlee-python !2025-03-2854493|Crawlee—A web scraping and browser automation library for Python to build reliable crawlers. Extract data for AI, LLMs, RAG, or GPTs. Download HTML, PDF, JPG, PNG, and other files from websites. Works with BeautifulSoup, Playwright, and raw HTTP. Both headful and headless mode. With proxy rotation.| | 762|k2-fsa/sherpa-onnx !2025-03-28541520|Speech-to-text, text-to-speech, and speaker recongition using next-gen Kaldi with onnxruntime without Internet connection. Support embedded systems, Android, iOS, Raspberry Pi, RISC-V, x86_64 servers, websocket server/client, C/C++, Python, Kotlin, C#, Go, NodeJS, Java, Swift| | 763|TEN-framework/TEN-Agent !2025-03-28541411|TEN Agent is a realtime conversational AI agent powered by TEN. It seamlessly integrates the OpenAI Realtime API, RTC capabilities, and advanced features like weather updates, web search, computer vision, and Retrieval-Augmented Generation (RAG).| | 764|google/gemmapytorch !2025-03-2854010|The official PyTorch implementation of Google's Gemma models| | 765|snakers4/silero-vad !2025-03-2853858|Silero VAD: pre-trained enterprise-grade Voice Activity Detector| | 766|livekit/agents !2025-03-2853836|Build real-time multimodal AI applications 🤖🎙️📹| | 767|pipecat-ai/pipecat !2025-03-28537811|Open Source framework for voice and multimodal conversational AI| | 768|EricLBuehler/mistral.rs !2025-03-28536324|Blazingly fast LLM inference.| | 769|asg017/sqlite-vec !2025-03-28535810|Work-in-progress vector search SQLite extension that runs anywhere.| | 770|albertan017/LLM4Decompile !2025-03-2853563|Reverse Engineering: Decompiling Binary Code with Large Language Models| | 771|Permify/permify !2025-03-2853235|An open-source authorization as a service inspired by Google Zanzibar, designed to build and manage fine-grained and scalable authorization systems for any application.| | 772|imoneoi/openchat !2025-03-2853171|OpenChat: Advancing Open-source Language Models with Imperfect Data| | 773|mosaicml/composer !2025-03-2853140|Train neural networks up to 7x faster| | 774|dsdanielpark/Bard-API !2025-03-285277-1 |The python package that returns a response of Google Bard through API.| | 775|lxfater/inpaint-web !2025-03-2852552|A free and open-source inpainting & image-upscaling tool powered by webgpu and wasm on the browser。| | 776|leanprover/lean4 !2025-03-2852441|Lean 4 programming language and theorem prover| | 777|AILab-CVC/YOLO-World !2025-03-2852415|Real-Time Open-Vocabulary Object Detection| | 778|openchatai/OpenChat !2025-03-2852260 |Run and create custom ChatGPT-like bots with OpenChat, embed and share these bots anywhere, the open-source chatbot console.| | 779|mufeedvh/code2prompt !2025-03-28519414|A CLI tool to convert your codebase into a single LLM prompt with source tree, prompt templating, and token counting.| | 780|biobootloader/wolverine !2025-03-2851700 |Automatically repair python scripts through GPT-4 to give them regenerative abilities.| | 781|huggingface/parler-tts !2025-03-2851671|Inference and training library for high-quality TTS models.| | 782|Akegarasu/lora-scripts !2025-03-2851308 |LoRA training scripts use kohya-ss's trainer, for diffusion model.| | 783|openchatai/OpenCopilot !2025-03-285128-3|🤖 🔥 Let your users chat with your product features and execute things by text - open source Shopify sidekick| | 784|e2b-dev/fragments !2025-03-2851228|Open-source Next.js template for building apps that are fully generated by AI. By E2B.| | 785|microsoft/SynapseML !2025-03-2851132|Simple and Distributed Machine Learning| | 786|aigc-apps/sd-webui-EasyPhoto !2025-03-285108-1|📷 EasyPhoto | | 787|ChaoningZhang/MobileSAM !2025-03-2850944|This is the official code for Faster Segment Anything (MobileSAM) project that makes SAM lightweight| | 788|huggingface/alignment-handbook !2025-03-2850932|Robust recipes for to align language models with human and AI preferences| | 789|alpkeskin/mosint !2025-03-2850920|An automated e-mail OSINT tool| | 790|TaskingAI/TaskingAI !2025-03-2850891|The open source platform for AI-native application development.| | 791|lipku/metahuman-stream !2025-03-28507615|Real time interactive streaming digital human| | 792|OpenInterpreter/01 !2025-03-2850530|The open-source language model computer| | 793|open-compass/opencompass !2025-03-28505111|OpenCompass is an LLM evaluation platform, supporting a wide range of models (InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.| | 794|xxlong0/Wonder3D !2025-03-2850491|A cross-domain diffusion model for 3D reconstruction from a single image| | 795|pytorch/torchtune !2025-03-2850342|A Native-PyTorch Library for LLM Fine-tuning| | 796|SuperDuperDB/superduperdb !2025-03-2850192|🔮 SuperDuperDB: Bring AI to your database: Integrate, train and manage any AI models and APIs directly with your database and your data.| | 797|WhiskeySockets/Baileys !2025-03-2850057|Lightweight full-featured typescript/javascript WhatsApp Web API| | 798| mpociot/chatgpt-vscode !2025-03-2849890 | A VSCode extension that allows you to use ChatGPT | | 799|OpenGVLab/DragGAN !2025-03-2849880|Unofficial Implementation of DragGAN - "Drag Your GAN: Interactive Point-based Manipulation on the Generative Image Manifold" (DragGAN 全功能实现,在线Demo,本地部署试用,代码、模型已全部开源,支持Windows, macOS, Linux)| | 800|microsoft/LLMLingua !2025-03-2849824|To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.| | 801|Zipstack/unstract !2025-03-2849745|No-code LLM Platform to launch APIs and ETL Pipelines to structure unstructured documents| | 802|OpenBMB/ToolBench !2025-03-2849621|An open platform for training, serving, and evaluating large language model for tool learning.| | 803|Fanghua-Yu/SUPIR !2025-03-2849593|SUPIR aims at developing Practical Algorithms for Photo-Realistic Image Restoration In the Wild| | 804|GaiaNet-AI/gaianet-node !2025-03-2849360|Install and run your own AI agent service| | 805|qodo-ai/qodo-cover !2025-03-284922-1|Qodo-Cover: An AI-Powered Tool for Automated Test Generation and Code Coverage Enhancement! 💻🤖🧪🐞| | 806|Zejun-Yang/AniPortrait !2025-03-2849042|AniPortrait: Audio-Driven Synthesis of Photorealistic Portrait Animation| | 807|lvwzhen/law-cn-ai !2025-03-2848901 |⚖️ AI Legal Assistant| | 808|developersdigest/llm-answer-engine !2025-03-2848740|Build a Perplexity-Inspired Answer Engine Using Next.js, Groq, Mixtral, Langchain, OpenAI, Brave & Serper| | 809|Plachtaa/VITS-fast-fine-tuning !2025-03-2848640|This repo is a pipeline of VITS finetuning for fast speaker adaptation TTS, and many-to-many voice conversion| | 810|espeak-ng/espeak-ng !2025-03-2848601|eSpeak NG is an open source speech synthesizer that supports more than hundred languages and accents.| | 811|ant-research/CoDeF !2025-03-2848581|[CVPR'24 Highlight] Official PyTorch implementation of CoDeF: Content Deformation Fields for Temporally Consistent Video Processing| | 812|deepseek-ai/DeepSeek-V2 !2025-03-2848512|| | 813|XRPLF/rippled !2025-03-2848210|Decentralized cryptocurrency blockchain daemon implementing the XRP Ledger protocol in C++| | 814|AutoMQ/automq !2025-03-28478721|AutoMQ is a cloud-first alternative to Kafka by decoupling durability to S3 and EBS. 10x cost-effective. Autoscale in seconds. Single-digit ms latency.| | 815|AILab-CVC/VideoCrafter !2025-03-2847800|VideoCrafter1: Open Diffusion Models for High-Quality Video Generation| | 816|nautechsystems/nautilustrader !2025-03-2847702|A high-performance algorithmic trading platform and event-driven backtester| | 817|kyegomez/swarms !2025-03-2847563|The Enterprise-Grade Production-Ready Multi-Agent Orchestration Framework Join our Community: https://discord.com/servers/agora-999382051935506503| | 818|Deci-AI/super-gradients !2025-03-2847310 |Easily train or fine-tune SOTA computer vision models with one open source training library. The home of Yolo-NAS.| | 819|QwenLM/Qwen2.5-Coder !2025-03-2847236|Qwen2.5-Coder is the code version of Qwen2.5, the large language model series developed by Qwen team, Alibaba Cloud.| | 820|SCIR-HI/Huatuo-Llama-Med-Chinese !2025-03-2847191 |Repo for HuaTuo (华驼), Llama-7B tuned with Chinese medical knowledge| | 821|togethercomputer/RedPajama-Data !2025-03-2846841 |code for preparing large datasets for training large language models| | 822|mishushakov/llm-scraper !2025-03-2846704|Turn any webpage into structured data using LLMs| | 823|1rgs/jsonformer !2025-03-2846663 |A Bulletproof Way to Generate Structured JSON from Language Models| | 824|anti-work/shortest !2025-03-2846565|QA via natural language AI tests| | 825|dnhkng/GlaDOS !2025-03-2846510|This is the Personality Core for GLaDOS, the first steps towards a real-life implementation of the AI from the Portal series by Valve.| | 826|Nukem9/dlssg-to-fsr3 !2025-03-2846380|Adds AMD FSR3 Frame Generation to games by replacing Nvidia DLSS-G Frame Generation (nvngx_dlssg).| | 827|BuilderIO/ai-shell !2025-03-2846373 |A CLI that converts natural language to shell commands.| | 828|facebookincubator/AITemplate !2025-03-2846220 |AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.| | 829|terraform-aws-modules/terraform-aws-eks !2025-03-2846030|Terraform module to create AWS Elastic Kubernetes (EKS) resources 🇺🇦| | 830|timescale/pgai !2025-03-2845915|A suite of tools to develop RAG, semantic search, and other AI applications more easily with PostgreSQL| | 831|awslabs/multi-agent-orchestrator !2025-03-2845788|Flexible and powerful framework for managing multiple AI agents and handling complex conversations| | 832|sanchit-gandhi/whisper-jax !2025-03-2845771 |Optimised JAX code for OpenAI's Whisper Model, largely built on the Hugging Face Transformers Whisper implementation| | 833|NVIDIA/NeMo-Guardrails !2025-03-2845755|NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.| | 834|PathOfBuildingCommunity/PathOfBuilding !2025-03-2845480|Offline build planner for Path of Exile.| | 835|UX-Decoder/Segment-Everything-Everywhere-All-At-Once !2025-03-2845412 |Official implementation of the paper "Segment Everything Everywhere All at Once"| | 836|build-trust/ockam !2025-03-2845171|Orchestrate end-to-end encryption, cryptographic identities, mutual authentication, and authorization policies between distributed applications – at massive scale.| | 837|google-research/timesfm !2025-03-2845135|TimesFM (Time Series Foundation Model) is a pretrained time-series foundation model developed by Google Research for time-series forecasting.| | 838|luosiallen/latent-consistency-model !2025-03-2844842|Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference| | 839|NVlabs/neuralangelo !2025-03-2844740|Official implementation of "Neuralangelo: High-Fidelity Neural Surface Reconstruction" (CVPR 2023)| | 840|kyegomez/tree-of-thoughts !2025-03-2844720 |Plug in and Play Implementation of Tree of Thoughts: Deliberate Problem Solving with Large Language Models that Elevates Model Reasoning by atleast 70%| | 841|sjvasquez/handwriting-synthesis !2025-03-2844720 |Handwriting Synthesis with RNNs ✏️| | 842| madawei2699/myGPTReader !2025-03-2844420 | A slack bot that can read any webpage, ebook or document and summarize it with chatGPT | | 843|OpenBMB/AgentVerse !2025-03-2844413|🤖 AgentVerse 🪐 provides a flexible framework that simplifies the process of building custom multi-agent environments for large language models (LLMs).| | 844|argmaxinc/WhisperKit !2025-03-2844395|Swift native speech recognition on-device for iOS and macOS applications.| | 845|landing-ai/vision-agent !2025-03-2844346|Vision agent| | 846|InternLM/xtuner !2025-03-2844273|An efficient, flexible and full-featured toolkit for fine-tuning large models (InternLM, Llama, Baichuan, Qwen, ChatGLM)| | 847|google-deepmind/alphageometry !2025-03-284421-1|Solving Olympiad Geometry without Human Demonstrations| | 848|ostris/ai-toolkit !2025-03-2844093|Various AI scripts. Mostly Stable Diffusion stuff.| | 849|LLM-Red-Team/kimi-free-api !2025-03-2844004|🚀 KIMI AI 长文本大模型白嫖服务,支持高速流式输出、联网搜索、长文档解读、图像解析、多轮对话,零配置部署,多路token支持,自动清理会话痕迹。| | 850|argilla-io/argilla !2025-03-2843991|Argilla is a collaboration platform for AI engineers and domain experts that require high-quality outputs, full data ownership, and overall efficiency.| | 851|spring-projects/spring-ai !2025-03-28438419|An Application Framework for AI Engineering| | 852|alibaba-damo-academy/FunClip !2025-03-2843555|Open-source, accurate and easy-to-use video clipping tool, LLM based AI clipping intergrated | | 853|yisol/IDM-VTON !2025-03-2843541|IDM-VTON : Improving Diffusion Models for Authentic Virtual Try-on in the Wild| | 854|fchollet/ARC-AGI !2025-03-2843368|The Abstraction and Reasoning Corpus| | 855|MahmoudAshraf97/whisper-diarization !2025-03-2843064|Automatic Speech Recognition with Speaker Diarization based on OpenAI Whisper| | 856|Speykious/cve-rs !2025-03-2843047|Blazingly 🔥 fast 🚀 memory vulnerabilities, written in 100% safe Rust. 🦀| | 857|Blealtan/efficient-kan !2025-03-2842770|An efficient pure-PyTorch implementation of Kolmogorov-Arnold Network (KAN).| | 858|smol-ai/GodMode !2025-03-284249-1|AI Chat Browser: Fast, Full webapp access to ChatGPT / Claude / Bard / Bing / Llama2! I use this 20 times a day.| | 859|openai/plugins-quickstart !2025-03-284235-4 |Get a ChatGPT plugin up and running in under 5 minutes!| | 860|Doriandarko/maestro !2025-03-2842260|A framework for Claude Opus to intelligently orchestrate subagents.| | 861|philz1337x/clarity-upscaler !2025-03-2842204|Clarity-Upscaler: Reimagined image upscaling for everyone| | 862|facebookresearch/co-tracker !2025-03-2842142|CoTracker is a model for tracking any point (pixel) on a video.| | 863|xlang-ai/OpenAgents !2025-03-2842031|OpenAgents: An Open Platform for Language Agents in the Wild| | 864|alibaba/higress !2025-03-28419514|🤖 AI Gateway | | 865|ray-project/llm-numbers !2025-03-2841920 |Numbers every LLM developer should know| | 866|fudan-generative-vision/champ !2025-03-2841820|Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance| | 867|NVIDIA/garak !2025-03-2841795|the LLM vulnerability scanner| | 868|leetcode-mafia/cheetah !2025-03-2841740 |Whisper & GPT-based app for passing remote SWE interviews| | 869|ragapp/ragapp !2025-03-2841710|The easiest way to use Agentic RAG in any enterprise| | 870|collabora/WhisperSpeech !2025-03-2841692|An Open Source text-to-speech system built by inverting Whisper.| | 871|Facico/Chinese-Vicuna !2025-03-2841520 |Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model| | 872|openai/grok !2025-03-2841381|| | 873|CrazyBoyM/llama3-Chinese-chat !2025-03-2841361|Llama3 Chinese Repository with modified versions, and training and deployment resources| | 874|luban-agi/Awesome-AIGC-Tutorials !2025-03-2841301|Curated tutorials and resources for Large Language Models, AI Painting, and more.| | 875|damo-vilab/AnyDoor !2025-03-2841192|Official implementations for paper: Anydoor: zero-shot object-level image customization| | 876|raspberrypi/pico-sdk !2025-03-2841072|| | 877|mshumer/gpt-llm-trainer !2025-03-284097-1|| | 878|metavoiceio/metavoice-src !2025-03-284076-1|AI for human-level speech intelligence| | 879|intelowlproject/IntelOwl !2025-03-2840763|IntelOwl: manage your Threat Intelligence at scale| | 880|a16z-infra/ai-getting-started !2025-03-2840682|A Javascript AI getting started stack for weekend projects, including image/text models, vector stores, auth, and deployment configs| | 881|MarkFzp/mobile-aloha !2025-03-2840641|Mobile ALOHA: Learning Bimanual Mobile Manipulation with Low-Cost Whole-Body Teleoperation| | 882| keijiro/AICommand !2025-03-2840380 | ChatGPT integration with Unity Editor | | 883|Tencent/HunyuanDiT !2025-03-2840214|Hunyuan-DiT : A Powerful Multi-Resolution Diffusion Transformer with Fine-Grained Chinese Understanding| | 884|hengyoush/kyanos !2025-03-2840061|Visualize the time packets spend in the kernel, watch & analyze in command line.| | 885|agiresearch/AIOS !2025-03-2840045|AIOS: LLM Agent Operating System| | 886|truefoundry/cognita !2025-03-2839773|RAG (Retrieval Augmented Generation) Framework for building modular, open source applications for production by TrueFoundry| | 887|X-PLUG/MobileAgent !2025-03-2839557|Mobile-Agent: Autonomous Multi-Modal Mobile Device Agent with Visual Perception| | 888|jackMort/ChatGPT.nvim !2025-03-2839231|ChatGPT Neovim Plugin: Effortless Natural Language Generation with OpenAI's ChatGPT API| | 889|microsoft/RD-Agent !2025-03-28388422|Research and development (R&D) is crucial for the enhancement of industrial productivity, especially in the AI era, where the core aspects of R&D are mainly focused on data and models. We are committed to automate these high-value generic R&D processes through our open source R&D automation tool RD-Agent, which let AI drive data-driven AI.| | 890|Significant-Gravitas/Auto-GPT-Plugins !2025-03-283882-1 |Plugins for Auto-GPT| | 891|apple/ml-mgie !2025-03-2838770|| | 892|OpenDriveLab/UniAD !2025-03-2838727|[CVPR 2023 Best Paper] Planning-oriented Autonomous Driving| | 893|llSourcell/DoctorGPT !2025-03-2838640|DoctorGPT is an LLM that can pass the US Medical Licensing Exam. It works offline, it's cross-platform, & your health data stays private.| | 894|FlagAI-Open/FlagAI !2025-03-2838601|FlagAI (Fast LArge-scale General AI models) is a fast, easy-to-use and extensible toolkit for large-scale model.| | 895|krishnaik06/Roadmap-To-Learn-Generative-AI-In-2024 !2025-03-2838513|Roadmap To Learn Generative AI In 2024| | 896|SysCV/sam-hq !2025-03-2838491|Segment Anything in High Quality| | 897|google/security-research !2025-03-2838420|This project hosts security advisories and their accompanying proof-of-concepts related to research conducted at Google which impact non-Google owned code.| | 898|shroominic/codeinterpreter-api !2025-03-2838330|Open source implementation of the ChatGPT Code Interpreter 👾| | 899|Yonom/assistant-ui !2025-03-2838308|React Components for AI Chat 💬 🚀| | 900|nucleuscloud/neosync !2025-03-2838262|Open source data anonymization and synthetic data orchestration for developers. Create high fidelity synthetic data and sync it across your environments.| | 901|ravenscroftj/turbopilot !2025-03-2838230 |Turbopilot is an open source large-language-model based code completion engine that runs locally on CPU| | 902|NVlabs/Sana !2025-03-28380810|SANA: Efficient High-Resolution Image Synthesis with Linear Diffusion Transformer| | 903|huggingface/distil-whisper !2025-03-2838061|Distilled variant of Whisper for speech recognition. 6x faster, 50% smaller, within 1% word error rate.| | 904|Codium-ai/AlphaCodium !2025-03-2837971|code generation tool that surpasses most human competitors in CodeContests| | 905|fixie-ai/ultravox !2025-03-2837710|A fast multimodal LLM for real-time voice| | 906|unit-mesh/auto-dev !2025-03-28375715|🧙‍AutoDev: The AI-powered coding wizard with multilingual support 🌐, auto code generation 🏗️, and a helpful bug-slaying assistant 🐞! Customizable prompts 🎨 and a magic Auto Dev/Testing/Document/Agent feature 🧪 included! 🚀| | 907|Marker-Inc-Korea/AutoRAG !2025-03-2837432|AutoML tool for RAG| | 908|deepseek-ai/DeepSeek-VL !2025-03-283734-1|DeepSeek-VL: Towards Real-World Vision-Language Understanding| | 909|hiyouga/ChatGLM-Efficient-Tuning !2025-03-283692-1|Fine-tuning ChatGLM-6B with PEFT | | 910| Yue-Yang/ChatGPT-Siri !2025-03-2836921 | Shortcuts for Siri using ChatGPT API gpt-3.5-turbo model | | 911|0hq/WebGPT !2025-03-2836901 |Run GPT model on the browser with WebGPU. An implementation of GPT inference in less than ~2000 lines of vanilla Javascript.| | 912|cvg/LightGlue !2025-03-2836903|LightGlue: Local Feature Matching at Light Speed (ICCV 2023)| | 913|deanxv/coze-discord-proxy !2025-03-2836791|代理Discord-Bot对话Coze-Bot,实现API形式请求GPT4对话模型/微调模型| | 914|MervinPraison/PraisonAI !2025-03-2836764|PraisonAI application combines AutoGen and CrewAI or similar frameworks into a low-code solution for building and managing multi-agent LLM systems, focusing on simplicity, customisation, and efficient human-agent collaboration.| | 915|Ironclad/rivet !2025-03-2836345 |The open-source visual AI programming environment and TypeScript library| | 916|BasedHardware/OpenGlass !2025-03-2835851|Turn any glasses into AI-powered smart glasses| | 917|ricklamers/gpt-code-ui !2025-03-2835840 |An open source implementation of OpenAI's ChatGPT Code interpreter| | 918|whoiskatrin/chart-gpt !2025-03-2835830 |AI tool to build charts based on text input| | 919|github/CopilotForXcode !2025-03-2835788|Xcode extension for GitHub Copilot| | 920|hemansnation/God-Level-Data-Science-ML-Full-Stack !2025-03-2835570 |A collection of scientific methods, processes, algorithms, and systems to build stories & models. This roadmap contains 16 Chapters, whether you are a fresher in the field or an experienced professional who wants to transition into Data Science & AI| | 921|pytorch/torchchat !2025-03-2835461|Run PyTorch LLMs locally on servers, desktop and mobile| | 922| Kent0n-Li/ChatDoctor !2025-03-2835451 | A Medical Chat Model Fine-tuned on LLaMA Model using Medical Domain Knowledge | | 923|xtekky/chatgpt-clone !2025-03-283519-1 |ChatGPT interface with better UI| | 924|jupyterlab/jupyter-ai !2025-03-2835120|A generative AI extension for JupyterLab| | 925|pytorch/torchtitan !2025-03-2835064|A native PyTorch Library for large model training| | 926|minimaxir/simpleaichat !2025-03-2835031|Python package for easily interfacing with chat apps, with robust features and minimal code complexity.| | 927|srush/Tensor-Puzzles !2025-03-2834930|Solve puzzles. Improve your pytorch.| | 928|Helicone/helicone !2025-03-2834918|🧊 Open source LLM-Observability Platform for Developers. One-line integration for monitoring, metrics, evals, agent tracing, prompt management, playground, etc. Supports OpenAI SDK, Vercel AI SDK, Anthropic SDK, LiteLLM, LLamaIndex, LangChain, and more. 🍓 YC W23| | 929|run-llama/llama-hub !2025-03-2834740|A library of data loaders for LLMs made by the community -- to be used with LlamaIndex and/or LangChain| | 930|NExT-GPT/NExT-GPT !2025-03-2834700|Code and models for NExT-GPT: Any-to-Any Multimodal Large Language Model| | 931|souzatharsis/podcastfy !2025-03-2834661|An Open Source Python alternative to NotebookLM's podcast feature: Transforming Multimodal Content into Captivating Multilingual Audio Conversations with GenAI| | 932|Dataherald/dataherald !2025-03-2834450|Interact with your SQL database, Natural Language to SQL using LLMs| | 933|iryna-kondr/scikit-llm !2025-03-2834350 |Seamlessly integrate powerful language models like ChatGPT into scikit-learn for enhanced text analysis tasks.| | 934|Netflix/maestro !2025-03-2834230|Maestro: Netflix’s Workflow Orchestrator| | 935|CanadaHonk/porffor !2025-03-2833560|A from-scratch experimental AOT JS engine, written in JS| | 936|hustvl/Vim !2025-03-2833323|Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model| | 937|pashpashpash/vault-ai !2025-03-2833250 |OP Vault ChatGPT: Give ChatGPT long-term memory using the OP Stack (OpenAI + Pinecone Vector Database). Upload your own custom knowledge base files (PDF, txt, etc) using a simple React frontend.| | 938|tencentmusic/supersonic !2025-03-28330611|SuperSonic is the next-generation BI platform that integrates Chat BI (powered by LLM) and Headless BI (powered by semantic layer) paradigms.| | 939|billmei/every-chatgpt-gui !2025-03-2832981|Every front-end GUI client for ChatGPT| | 940|microsoft/torchgeo !2025-03-2832772|TorchGeo: datasets, samplers, transforms, and pre-trained models for geospatial data| | 941|LLMBook-zh/LLMBook-zh.github.io !2025-03-28326110|《大语言模型》作者:赵鑫,李军毅,周昆,唐天一,文继荣| | 942|dvlab-research/MiniGemini !2025-03-2832601|Official implementation for Mini-Gemini| | 943|rashadphz/farfalle !2025-03-2832460|🔍 AI search engine - self-host with local or cloud LLMs| | 944|Luodian/Otter !2025-03-2832450|🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.| | 945|AprilNEA/ChatGPT-Admin-Web !2025-03-2832370 | ChatGPT WebUI with user management and admin dashboard system| | 946|MarkFzp/act-plus-plus !2025-03-2832365|Imitation Learning algorithms with Co-traing for Mobile ALOHA: ACT, Diffusion Policy, VINN| | 947|ethen8181/machine-learning !2025-03-2832310|🌎 machine learning tutorials (mainly in Python3)| | 948|opengeos/segment-geospatial !2025-03-2832312 |A Python package for segmenting geospatial data with the Segment Anything Model (SAM)| | 949|iusztinpaul/hands-on-llms !2025-03-283225-2|🦖 𝗟𝗲𝗮𝗿𝗻 about 𝗟𝗟𝗠𝘀, 𝗟𝗟𝗠𝗢𝗽𝘀, and 𝘃𝗲𝗰𝘁𝗼𝗿 𝗗𝗕𝘀 for free by designing, training, and deploying a real-time financial advisor LLM system ~ 𝘴𝘰𝘶𝘳𝘤𝘦 𝘤𝘰𝘥𝘦 + 𝘷𝘪𝘥𝘦𝘰 & 𝘳𝘦𝘢𝘥𝘪𝘯𝘨 𝘮𝘢𝘵𝘦𝘳𝘪𝘢𝘭𝘴| | 950|ToTheBeginning/PuLID !2025-03-2832221|Official code for PuLID: Pure and Lightning ID Customization via Contrastive Alignment| | 951|neo4j-labs/llm-graph-builder !2025-03-2832164|Neo4j graph construction from unstructured data using LLMs| | 952|OpenGVLab/InternGPT !2025-03-2832150 |InternGPT (iGPT) is an open source demo platform where you can easily showcase your AI models. Now it supports DragGAN, ChatGPT, ImageBind, multimodal chat like GPT-4, SAM, interactive image editing, etc. Try it at igpt.opengvlab.com (支持DragGAN、ChatGPT、ImageBind、SAM的在线Demo系统)| | 953|PKU-YuanGroup/Video-LLaVA !2025-03-2832060 |Video-LLaVA: Learning United Visual Representation by Alignment Before Projection| | 954|DataTalksClub/llm-zoomcamp !2025-03-2832030|LLM Zoomcamp - a free online course about building an AI bot that can answer questions about your knowledge base| | 955|gptscript-ai/gptscript !2025-03-2832010|Natural Language Programming| |!green-up-arrow.svg 956|isaac-sim/IsaacLab !2025-03-28320113|Unified framework for robot learning built on NVIDIA Isaac Sim| |!red-down-arrow 957|ai-boost/Awesome-GPTs !2025-03-2832003|Curated list of awesome GPTs 👍.| | 958|huggingface/safetensors !2025-03-2831901|Simple, safe way to store and distribute tensors| | 959|linyiLYi/bilibot !2025-03-2831771|A local chatbot fine-tuned by bilibili user comments.| | 960| project-baize/baize-chatbot !2025-03-283168-1 | Let ChatGPT teach your own chatbot in hours with a single GPU! | | 961|Azure-Samples/cognitive-services-speech-sdk !2025-03-2831280|Sample code for the Microsoft Cognitive Services Speech SDK| | 962|microsoft/Phi-3CookBook !2025-03-2831231|This is a Phi-3 book for getting started with Phi-3. Phi-3, a family of open AI models developed by Microsoft. Phi-3 models are the most capable and cost-effective small language models (SLMs) available, outperforming models of the same size and next size up across a variety of language, reasoning, coding, and math benchmarks.| | 963|neuralmagic/deepsparse !2025-03-2831180|Sparsity-aware deep learning inference runtime for CPUs| | 964|sugarforever/chat-ollama !2025-03-2831000|ChatOllama is an open source chatbot based on LLMs. It supports a wide range of language models, and knowledge base management.| | 965|amazon-science/chronos-forecasting !2025-03-2830974|Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting| | 966|damo-vilab/i2vgen-xl !2025-03-2830902|Official repo for VGen: a holistic video generation ecosystem for video generation building on diffusion models| | 967|google-deepmind/gemma !2025-03-2830733|Open weights LLM from Google DeepMind.| | 968|iree-org/iree !2025-03-2830733|A retargetable MLIR-based machine learning compiler and runtime toolkit.| | 969|NVlabs/VILA !2025-03-2830724|VILA - a multi-image visual language model with training, inference and evaluation recipe, deployable from cloud to edge (Jetson Orin and laptops)| | 970|microsoft/torchscale !2025-03-2830661|Foundation Architecture for (M)LLMs| | 971|openai/openai-realtime-console !2025-03-2830656|React app for inspecting, building and debugging with the Realtime API| | 972|daveshap/OpenAIAgentSwarm !2025-03-2830610|HAAS = Hierarchical Autonomous Agent Swarm - "Resistance is futile!"| | 973|microsoft/PromptWizard !2025-03-2830555|Task-Aware Agent-driven Prompt Optimization Framework| | 974|CVI-SZU/Linly !2025-03-2830490 |Chinese-LLaMA basic model; ChatFlow Chinese conversation model; NLP pre-training/command fine-tuning dataset| | 975|cohere-ai/cohere-toolkit !2025-03-2830130|Toolkit is a collection of prebuilt components enabling users to quickly build and deploy RAG applications.| | 976|adamcohenhillel/ADeus !2025-03-2830131|An open source AI wearable device that captures what you say and hear in the real world and then transcribes and stores it on your own server. You can then chat with Adeus using the app, and it will have all the right context about what you want to talk about - a truly personalized, personal AI.| | 977|Lightning-AI/LitServe !2025-03-2830132|Lightning-fast serving engine for AI models. Flexible. Easy. Enterprise-scale.| | 978|potpie-ai/potpie !2025-03-2829973|Prompt-To-Agent : Create custom engineering agents for your codebase| | 979|ant-design/x !2025-03-28299529|Craft AI-driven interfaces effortlessly 🤖| | 980|meta-llama/PurpleLlama !2025-03-2829832|Set of tools to assess and improve LLM security.| | 981|williamyang1991/RerenderAVideo !2025-03-2829800|[SIGGRAPH Asia 2023] Rerender A Video: Zero-Shot Text-Guided Video-to-Video Translation| | 982|baichuan-inc/Baichuan-13B !2025-03-2829790|A 13B large language model developed by Baichuan Intelligent Technology| | 983|Stability-AI/stable-audio-tools !2025-03-2829761|Generative models for conditional audio generation| | 984|li-plus/chatglm.cpp !2025-03-2829720|C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & more LLMs| | 985|NVIDIA/GenerativeAIExamples !2025-03-2829546|Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.| | 986|Josh-XT/AGiXT !2025-03-2829521 |AGiXT is a dynamic AI Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.| | 987|MrForExample/ComfyUI-3D-Pack !2025-03-2829515|An extensive node suite that enables ComfyUI to process 3D inputs (Mesh & UV Texture, etc) using cutting edge algorithms (3DGS, NeRF, etc.)| | 988|olimorris/codecompanion.nvim !2025-03-28295111|✨ AI-powered coding, seamlessly in Neovim. Supports Anthropic, Copilot, Gemini, Ollama, OpenAI and xAI LLMs| | 989|salesforce/CodeT5 !2025-03-282940-1 |Home of CodeT5: Open Code LLMs for Code Understanding and Generation| | 990|facebookresearch/ijepa !2025-03-2829391|Official codebase for I-JEPA, the Image-based Joint-Embedding Predictive Architecture. First outlined in the CVPR paper, "Self-supervised learning from images with a joint-embedding predictive architecture."| | 991|eureka-research/Eureka !2025-03-2829351|Official Repository for "Eureka: Human-Level Reward Design via Coding Large Language Models"| | 992|NVIDIA/trt-llm-rag-windows !2025-03-282934-1|A developer reference project for creating Retrieval Augmented Generation (RAG) chatbots on Windows using TensorRT-LLM| | 993|gmpetrov/databerry !2025-03-282930-1|The no-code platform for building custom LLM Agents| | 994|AI4Finance-Foundation/FinRobot !2025-03-28291946|FinRobot: An Open-Source AI Agent Platform for Financial Applications using LLMs 🚀 🚀 🚀| | 995|nus-apr/auto-code-rover !2025-03-2829013|A project structure aware autonomous software engineer aiming for autonomous program improvement| | 996|deepseek-ai/DreamCraft3D !2025-03-2828921|[ICLR 2024] Official implementation of DreamCraft3D: Hierarchical 3D Generation with Bootstrapped Diffusion Prior| | 997|mlabonne/llm-datasets !2025-03-2828848|High-quality datasets, tools, and concepts for LLM fine-tuning.| | 998|facebookresearch/jepa !2025-03-2828712|PyTorch code and models for V-JEPA self-supervised learning from video.| | 999|facebookresearch/habitat-sim !2025-03-2828604|A flexible, high-performance 3D simulator for Embodied AI research.| | 1000|xenova/whisper-web !2025-03-2828581|ML-powered speech recognition directly in your browser| | 1001|cvlab-columbia/zero123 !2025-03-2828530|Zero-1-to-3: Zero-shot One Image to 3D Object: https://zero123.cs.columbia.edu/| | 1002|yuruotong1/autoMate !2025-03-28285121|Like Manus, Computer Use Agent(CUA) and Omniparser, we are computer-using agents.AI-driven local automation assistant that uses natural language to make computers work by themselves| | 1003|muellerberndt/mini-agi !2025-03-282845-1 |A minimal generic autonomous agent based on GPT3.5/4. Can analyze stock prices, perform network security tests, create art, and order pizza.| | 1004|allenai/open-instruct !2025-03-2828432|| | 1005|CodingChallengesFYI/SharedSolutions !2025-03-2828360|Publicly shared solutions to Coding Challenges| | 1006|hegelai/prompttools !2025-03-2828220|Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate).| | 1007|mazzzystar/Queryable !2025-03-2828222|Run CLIP on iPhone to Search Photos.| | 1008|Doubiiu/DynamiCrafter !2025-03-2828173|DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors| | 1009|SamurAIGPT/privateGPT !2025-03-282805-1 |An app to interact privately with your documents using the power of GPT, 100% privately, no data leaks| | 1010|facebookresearch/Pearl !2025-03-2827951|A Production-ready Reinforcement Learning AI Agent Library brought by the Applied Reinforcement Learning team at Meta.| | 1011|intuitem/ciso-assistant-community !2025-03-2827954|CISO Assistant is a one-stop-shop for GRC, covering Risk, AppSec and Audit Management and supporting +70 frameworks worldwide with auto-mapping: NIST CSF, ISO 27001, SOC2, CIS, PCI DSS, NIS2, CMMC, PSPF, GDPR, HIPAA, Essential Eight, NYDFS-500, DORA, NIST AI RMF, 800-53, 800-171, CyFun, CJIS, AirCyber, NCSC, ECC, SCF and so much more| | 1012|facebookresearch/audio2photoreal !2025-03-2827840|Code and dataset for photorealistic Codec Avatars driven from audio| | 1013|Azure/azure-rest-api-specs !2025-03-2827770|The source for REST API specifications for Microsoft Azure.| | 1014|SCUTlihaoyu/open-chat-video-editor !2025-03-2827690 |Open source short video automatic generation tool| | 1015|Alpha-VLLM/LLaMA2-Accessory !2025-03-2827642|An Open-source Toolkit for LLM Development| | 1016|johnma2006/mamba-minimal !2025-03-2827601|Simple, minimal implementation of the Mamba SSM in one file of PyTorch.| | 1017|nerfstudio-project/gsplat !2025-03-2827576|CUDA accelerated rasterization of gaussian splatting| | 1018|Physical-Intelligence/openpi !2025-03-28274617|| | 1019|leptonai/leptonai !2025-03-2827246|A Pythonic framework to simplify AI service building| |!green-up-arrow.svg 1020|joanrod/star-vector !2025-03-28271149|StarVector is a foundation model for SVG generation that transforms vectorization into a code generation task. Using a vision-language modeling architecture, StarVector processes both visual and textual inputs to produce high-quality SVG code with remarkable precision.| |!red-down-arrow 1021|jqnatividad/qsv !2025-03-2827092|CSVs sliced, diced & analyzed.| | 1022|FranxYao/chain-of-thought-hub !2025-03-2826991|Benchmarking large language models' complex reasoning ability with chain-of-thought prompting| | 1023|princeton-nlp/SWE-bench !2025-03-2826965|[ICLR 2024] SWE-Bench: Can Language Models Resolve Real-world Github Issues?| | 1024|elastic/otel-profiling-agent !2025-03-2826930|The production-scale datacenter profiler| | 1025|src-d/hercules !2025-03-2826900|Gaining advanced insights from Git repository history.| | 1026|lanqian528/chat2api !2025-03-2826695|A service that can convert ChatGPT on the web to OpenAI API format.| | 1027|ishan0102/vimGPT !2025-03-2826681|Browse the web with GPT-4V and Vimium| | 1028|TMElyralab/MuseV !2025-03-2826650|MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising| | 1029|georgia-tech-db/eva !2025-03-2826600 |AI-Relational Database System | | 1030|kubernetes-sigs/controller-runtime !2025-03-2826590|Repo for the controller-runtime subproject of kubebuilder (sig-apimachinery)| | 1031|gptlink/gptlink !2025-03-2826550 |Build your own free commercial ChatGPT environment in 10 minutes. The setup is simple and includes features such as user management, orders, tasks, and payments| | 1032|pytorch/executorch !2025-03-2826534|On-device AI across mobile, embedded and edge for PyTorch| | 1033|NVIDIA/nv-ingest !2025-03-2826290|NVIDIA Ingest is an early access set of microservices for parsing hundreds of thousands of complex, messy unstructured PDFs and other enterprise documents into metadata and text to embed into retrieval systems.| | 1034|SuperTux/supertux !2025-03-2826081|SuperTux source code| | 1035|abi/secret-llama !2025-03-2826050|Fully private LLM chatbot that runs entirely with a browser with no server needed. Supports Mistral and LLama 3.| | 1036|liou666/polyglot !2025-03-2825841 |Desktop AI Language Practice Application| | 1037|janhq/nitro !2025-03-2825821|A fast, lightweight, embeddable inference engine to supercharge your apps with local AI. OpenAI-compatible API| | 1038|deepseek-ai/DeepSeek-Math !2025-03-2825825|DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models| | 1039|anthropics/prompt-eng-interactive-tutorial !2025-03-2825781|Anthropic's Interactive Prompt Engineering Tutorial| | 1040|microsoft/promptbench !2025-03-2825741|A unified evaluation framework for large language models| | 1041|baaivision/Painter !2025-03-2825580 |Painter & SegGPT Series: Vision Foundation Models from BAAI| | 1042|OpenPipe/OpenPipe !2025-03-2825581|Turn expensive prompts into cheap fine-tuned models| | 1043|TracecatHQ/tracecat !2025-03-2825531|😼 The AI-native, open source alternative to Tines / Splunk SOAR.| | 1044|JoshuaC215/agent-service-toolkit !2025-03-2825528|Full toolkit for running an AI agent service built with LangGraph, FastAPI and Streamlit| | 1045|databricks/dbrx !2025-03-2825460|Code examples and resources for DBRX, a large language model developed by Databricks| | 1046|lamini-ai/lamini !2025-03-2825271 |Official repo for Lamini's data generator for generating instructions to train instruction-following LLMs| | 1047|mshumer/gpt-author !2025-03-282510-1|| | 1048|TMElyralab/MusePose !2025-03-2824971|MusePose: a Pose-Driven Image-to-Video Framework for Virtual Human Generation| | 1049|Kludex/fastapi-tips !2025-03-2824974|FastAPI Tips by The FastAPI Expert!| | 1050|openai/simple-evals !2025-03-2824813|| | 1051|iterative/datachain !2025-03-2824732|AI-data warehouse to enrich, transform and analyze data from cloud storages| | 1052|girafe-ai/ml-course !2025-03-2824703|Open Machine Learning course| | 1053|kevmo314/magic-copy !2025-03-2824620 |Magic Copy is a Chrome extension that uses Meta's Segment Anything Model to extract a foreground object from an image and copy it to the clipboard.| | 1054|Eladlev/AutoPrompt !2025-03-2824432|A framework for prompt tuning using Intent-based Prompt Calibration| | 1055|OpenBMB/CPM-Bee !2025-03-282434-1 |A bilingual large-scale model with trillions of parameters| | 1056|IDEA-Research/T-Rex !2025-03-2824310|T-Rex2: Towards Generic Object Detection via Text-Visual Prompt Synergy| | 1057|microsoft/genaiscript !2025-03-2824202|Automatable GenAI Scripting| | 1058|paulpierre/RasaGPT !2025-03-2824090 |💬 RasaGPT is the first headless LLM chatbot platform built on top of Rasa and Langchain. Built w/ Rasa, FastAPI, Langchain, LlamaIndex, SQLModel, pgvector, ngrok, telegram| | 1059|ashishpatel26/LLM-Finetuning !2025-03-2823911|LLM Finetuning with peft| | 1060|SoraWebui/SoraWebui !2025-03-2823570|SoraWebui is an open-source Sora web client, enabling users to easily create videos from text with OpenAI's Sora model.| | 1061|6drf21e/ChatTTScolab !2025-03-2823491|🚀 一键部署(含离线整合包)!基于 ChatTTS ,支持音色抽卡、长音频生成和分角色朗读。简单易用,无需复杂安装。| | 1062|Azure/PyRIT !2025-03-2823343|The Python Risk Identification Tool for generative AI (PyRIT) is an open access automation framework to empower security professionals and machine learning engineers to proactively find risks in their generative AI systems.| | 1063|tencent-ailab/V-Express !2025-03-2823201|V-Express aims to generate a talking head video under the control of a reference image, an audio, and a sequence of V-Kps images.| | 1064|THUDM/CogVLM2 !2025-03-2823170|GPT4V-level open-source multi-modal model based on Llama3-8B| | 1065|dvmazur/mixtral-offloading !2025-03-2823001|Run Mixtral-8x7B models in Colab or consumer desktops| | 1066|semanser/codel !2025-03-2822950|✨ Fully autonomous AI Agent that can perform complicated tasks and projects using terminal, browser, and editor.| | 1067|mshumer/gpt-investor !2025-03-2822590|| | 1068|aixcoder-plugin/aiXcoder-7B !2025-03-2822550|official repository of aiXcoder-7B Code Large Language Model| | 1069|Azure-Samples/graphrag-accelerator !2025-03-2822503|One-click deploy of a Knowledge Graph powered RAG (GraphRAG) in Azure| | 1070|emcf/engshell !2025-03-2821830 |An English-language shell for any OS, powered by LLMs| | 1071|hncboy/chatgpt-web-java !2025-03-2821771|ChatGPT project developed in Java, based on Spring Boot 3 and JDK 17, supports both AccessToken and ApiKey modes| | 1072|openai/consistencydecoder !2025-03-2821692|Consistency Distilled Diff VAE| | 1073|Alpha-VLLM/Lumina-T2X !2025-03-2821681|Lumina-T2X is a unified framework for Text to Any Modality Generation| | 1074|bghira/SimpleTuner !2025-03-2821612|A general fine-tuning kit geared toward Stable Diffusion 2.1, Stable Diffusion 3, DeepFloyd, and SDXL.| | 1075|JiauZhang/DragGAN !2025-03-2821530 |Implementation of DragGAN: Interactive Point-based Manipulation on the Generative Image Manifold| | 1076|cgpotts/cs224u !2025-03-2821390|Code for Stanford CS224u| | 1077|PKU-YuanGroup/MoE-LLaVA !2025-03-2821300|Mixture-of-Experts for Large Vision-Language Models| | 1078|darrenburns/elia !2025-03-2820831|A snappy, keyboard-centric terminal user interface for interacting with large language models. Chat with ChatGPT, Claude, Llama 3, Phi 3, Mistral, Gemma and more.| | 1079|ageerle/ruoyi-ai !2025-03-28207898|RuoYi AI 是一个全栈式 AI 开发平台,旨在帮助开发者快速构建和部署个性化的 AI 应用。| | 1080|NVIDIA/gpu-operator !2025-03-2820510|NVIDIA GPU Operator creates/configures/manages GPUs atop Kubernetes| | 1081|BAAI-Agents/Cradle !2025-03-2820481|The Cradle framework is a first attempt at General Computer Control (GCC). Cradle supports agents to ace any computer task by enabling strong reasoning abilities, self-improvment, and skill curation, in a standardized general environment with minimal requirements.| | 1082|microsoft/aici !2025-03-2820080|AICI: Prompts as (Wasm) Programs| | 1083|PRIS-CV/DemoFusion !2025-03-2820040|Let us democratise high-resolution generation! (arXiv 2023)| | 1084|apple/axlearn !2025-03-2820012|An Extensible Deep Learning Library| | 1085|naver/mast3r !2025-03-2819685|Grounding Image Matching in 3D with MASt3R| | 1086|liltom-eth/llama2-webui !2025-03-281958-1|Run Llama 2 locally with gradio UI on GPU or CPU from anywhere (Linux/Windows/Mac). Supporting Llama-2-7B/13B/70B with 8-bit, 4-bit. Supporting GPU inference (6 GB VRAM) and CPU inference.| | 1087|GaParmar/img2img-turbo !2025-03-2819582|One-step image-to-image with Stable Diffusion turbo: sketch2image, day2night, and more| | 1088|Niek/chatgpt-web !2025-03-2819560|ChatGPT web interface using the OpenAI API| | 1089|huggingface/cookbook !2025-03-2819421|Open-source AI cookbook| | 1090|pytorch/ao !2025-03-2819241|PyTorch native quantization and sparsity for training and inference| | 1091|emcie-co/parlant !2025-03-2819053|The behavior guidance framework for customer-facing LLM agents| | 1092|ymcui/Chinese-LLaMA-Alpaca-3 !2025-03-2818980|中文羊驼大模型三期项目 (Chinese Llama-3 LLMs) developed from Meta Llama 3| | 1093|Nutlope/notesGPT !2025-03-2818811|Record voice notes & transcribe, summarize, and get tasks| | 1094|InstantStyle/InstantStyle !2025-03-2818791|InstantStyle: Free Lunch towards Style-Preserving in Text-to-Image Generation 🔥| | 1095|idaholab/moose !2025-03-2818771|Multiphysics Object Oriented Simulation Environment| | 1096|The-OpenROAD-Project/OpenROAD !2025-03-2818351|OpenROAD's unified application implementing an RTL-to-GDS Flow. Documentation at https://openroad.readthedocs.io/en/latest/| | 1097|alibaba/spring-ai-alibaba !2025-03-281831121|Agentic AI Framework for Java Developers| | 1098|ytongbai/LVM !2025-03-2817990|Sequential Modeling Enables Scalable Learning for Large Vision Models| | 1099|microsoft/sample-app-aoai-chatGPT !2025-03-2817981|[PREVIEW] Sample code for a simple web chat experience targeting chatGPT through AOAI.| | 1100|AI-Citizen/SolidGPT !2025-03-2817830|Chat everything with your code repository, ask repository level code questions, and discuss your requirements. AI Scan and learning your code repository, provide you code repository level answer🧱 🧱| | 1101|YangLing0818/RPG-DiffusionMaster !2025-03-2817784|Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs (PRG)| | 1102|kyegomez/BitNet !2025-03-2817710|Implementation of "BitNet: Scaling 1-bit Transformers for Large Language Models" in pytorch| | 1103|eloialonso/diamond !2025-03-2817671|DIAMOND (DIffusion As a Model Of eNvironment Dreams) is a reinforcement learning agent trained in a diffusion world model.| | 1104|flowdriveai/flowpilot !2025-03-2817250|flow-pilot is an openpilot based driver assistance system that runs on linux, windows and android powered machines.| | 1105|xlang-ai/OSWorld !2025-03-2817200|OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments| | 1106|linyiLYi/snake-ai !2025-03-2817031|An AI agent that beats the classic game "Snake".| | 1107|baaivision/Emu !2025-03-2816991|Emu Series: Generative Multimodal Models from BAAI| | 1108|kevmo314/scuda !2025-03-2816870|SCUDA is a GPU over IP bridge allowing GPUs on remote machines to be attached to CPU-only machines.| | 1109|SharifiZarchi/IntroductiontoMachineLearning !2025-03-2816701|دوره‌ی مقدمه‌ای بر یادگیری ماشین، برای دانشجویان| | 1110|google/maxtext !2025-03-2816670|A simple, performant and scalable Jax LLM!| | 1111|ml-explore/mlx-swift-examples !2025-03-2816471|Examples using MLX Swift| | 1112|unitreerobotics/unitreerlgym !2025-03-2816256|| | 1113|collabora/WhisperFusion !2025-03-2815901|WhisperFusion builds upon the capabilities of WhisperLive and WhisperSpeech to provide a seamless conversations with an AI.| | 1114|lichao-sun/Mora !2025-03-2815520|Mora: More like Sora for Generalist Video Generation| | 1115|GoogleCloudPlatform/localllm !2025-03-2815370|Run LLMs locally on Cloud Workstations| | 1116|TencentARC/BrushNet !2025-03-2815330|The official implementation of paper "BrushNet: A Plug-and-Play Image Inpainting Model with Decomposed Dual-Branch Diffusion"| | 1117|ai-christianson/RA.Aid !2025-03-2815288|Develop software autonomously.| | 1118|stephansturges/WALDO !2025-03-2815170|Whereabouts Ascertainment for Low-lying Detectable Objects. The SOTA in FOSS AI for drones!| | 1119|skills/copilot-codespaces-vscode !2025-03-2815112|Develop with AI-powered code suggestions using GitHub Copilot and VS Code| | 1120|andrewnguonly/Lumos !2025-03-2814920|A RAG LLM co-pilot for browsing the web, powered by local LLMs| | 1121|TeamNewPipe/NewPipeExtractor !2025-03-2814811|NewPipe's core library for extracting data from streaming sites| | 1122|mhamilton723/FeatUp !2025-03-2814770|Official code for "FeatUp: A Model-Agnostic Frameworkfor Features at Any Resolution" ICLR 2024| | 1123|AnswerDotAI/fsdpqlora !2025-03-2814671|Training LLMs with QLoRA + FSDP| | 1124|jgravelle/AutoGroq !2025-03-2814330|| | 1125|OpenGenerativeAI/llm-colosseum !2025-03-2814130|Benchmark LLMs by fighting in Street Fighter 3! The new way to evaluate the quality of an LLM| | 1126|microsoft/vscode-ai-toolkit !2025-03-2814000|| | 1127|McGill-NLP/webllama !2025-03-2813930|Llama-3 agents that can browse the web by following instructions and talking to you| | 1128|lucidrains/self-rewarding-lm-pytorch !2025-03-2813760|Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI| | 1129|ishaan1013/sandbox !2025-03-2813650|A cloud-based code editing environment with an AI copilot and real-time collaboration.| | 1130|goatcorp/Dalamud !2025-03-2813275|FFXIV plugin framework and API| | 1131|Lightning-AI/lightning-thunder !2025-03-2813151|Make PyTorch models Lightning fast! Thunder is a source to source compiler for PyTorch. It enables using different hardware executors at once.| | 1132|PKU-YuanGroup/MagicTime !2025-03-2813052|MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators| | 1133|SakanaAI/evolutionary-model-merge !2025-03-2813000|Official repository of Evolutionary Optimization of Model Merging Recipes| | 1134|a-real-ai/pywinassistant !2025-03-2812950|The first open source Large Action Model generalist Artificial Narrow Intelligence that controls completely human user interfaces by only using natural language. PyWinAssistant utilizes Visualization-of-Thought Elicits Spatial Reasoning in Large Language Models.| | 1135|TraceMachina/nativelink !2025-03-2812630|NativeLink is an open source high-performance build cache and remote execution server, compatible with Bazel, Buck2, Reclient, and other RBE-compatible build systems. It offers drastically faster builds, reduced test flakiness, and significant infrastructure cost savings.| | 1136|MLSysOps/MLE-agent !2025-03-2812500|🤖 MLE-Agent: Your intelligent companion for seamless AI engineering and research. 🔍 Integrate with arxiv and paper with code to provide better code/research plans 🧰 OpenAI, Ollama, etc supported. 🎆 Code RAG| | 1137|wpilibsuite/allwpilib !2025-03-2811610|Official Repository of WPILibJ and WPILibC| | 1138|elfvingralf/macOSpilot-ai-assistant !2025-03-2811470|Voice + Vision powered AI assistant that answers questions about any application, in context and in audio.| | 1139|langchain-ai/langchain-extract !2025-03-2811210|🦜⛏️ Did you say you like data?| | 1140|FoundationVision/GLEE !2025-03-2811120|【CVPR2024】GLEE: General Object Foundation Model for Images and Videos at Scale| | 1141|Profluent-AI/OpenCRISPR !2025-03-2810990|AI-generated gene editing systems| | 1142|zju3dv/EasyVolcap !2025-03-2810821|[SIGGRAPH Asia 2023 (Technical Communications)] EasyVolcap: Accelerating Neural Volumetric Video Research| | 1143|PaddlePaddle/PaddleHelix !2025-03-2810560|Bio-Computing Platform Featuring Large-Scale Representation Learning and Multi-Task Deep Learning “螺旋桨”生物计算工具集| | 1144|myshell-ai/JetMoE !2025-03-289800|Reaching LLaMA2 Performance with 0.1M Dollars| | 1145|likejazz/llama3.np !2025-03-289770|llama3.np is pure NumPy implementation for Llama 3 model.| | 1146|mustafaaljadery/gemma-2B-10M !2025-03-289500|Gemma 2B with 10M context length using Infini-attention.| | 1147|HITsz-TMG/FilmAgent !2025-03-289382|Resources of our paper "FilmAgent: A Multi-Agent Framework for End-to-End Film Automation in Virtual 3D Spaces". New versions in the making!| | 1148|aws-samples/amazon-bedrock-samples !2025-03-289362|This repository contains examples for customers to get started using the Amazon Bedrock Service. This contains examples for all available foundational models| | 1149|Akkudoktor-EOS/EOS !2025-03-2893154|This repository features an Energy Optimization System (EOS) that optimizes energy distribution, usage for batteries, heat pumps& household devices. It includes predictive models for electricity prices (planned), load forecasting& dynamic optimization to maximize energy efficiency & minimize costs. Founder Dr. Andreas Schmitz (YouTube @akkudoktor)| Tip: | symbol| rule | | :----| :---- | |🔥 | 256 1k| |!green-up-arrow.svg !red-down-arrow | ranking up / down| |⭐ | on trending page today| [Back to Top] Tools | No. | Tool | Description | | ----:|:----------------------------------------------- |:------------------------------------------------------------------------------------------- | | 1 | ChatGPT | A sibling model to InstructGPT, which is trained to follow instructions in a prompt and provide a detailed response | | 2 | DALL·E 2 | Create original, realistic images and art from a text description | | 3 | Murf AI | AI enabled, real people's voices| | 4 | Midjourney | An independent research lab that produces an artificial intelligence program under the same name that creates images from textual descriptions, used in Discord | 5 | Make-A-Video | Make-A-Video is a state-of-the-art AI system that generates videos from text | | 6 | Creative Reality™ Studio by D-ID| Use generative AI to create future-facing videos| | 7 | chat.D-ID| The First App Enabling Face-to-Face Conversations with ChatGPT| | 8 | Notion AI| Access the limitless power of AI, right inside Notion. Work faster. Write better. Think bigger. | | 9 | Runway| Text to Video with Gen-2 | | 10 | Resemble AI| Resemble’s AI voice generator lets you create human–like voice overs in seconds | | 11 | Cursor| Write, edit, and chat about your code with a powerful AI | | 12 | Hugging Face| Build, train and deploy state of the art models powered by the reference open source in machine learning | | 13 | Claude | A next-generation AI assistant for your tasks, no matter the scale | | 14 | Poe| Poe lets you ask questions, get instant answers, and have back-and-forth conversations with AI. Gives access to GPT-4, gpt-3.5-turbo, Claude from Anthropic, and a variety of other bots| [Back to Top] Websites | No. | WebSite |Description | | ----:|:------------------------------------------ |:---------------------------------------------------------------------------------------- | | 1 | OpenAI | An artificial intelligence research lab | | 2 | Bard | Base Google's LaMDA chatbots and pull from internet | | 3 | ERNIE Bot | Baidu’s new generation knowledge-enhanced large language model is a new member of the Wenxin large model family | | 4 | DALL·E 2 | An AI system that can create realistic images and art from a description in natural language | | 5 | Whisper | A general-purpose speech recognition model | | 6| CivitAI| A platform that makes it easy for people to share and discover resources for creating AI art| | 7|D-ID| D-ID’s Generative AI enables users to transform any picture or video into extraordinary experiences| | 8| Nvidia eDiff-I| Text-to-Image Diffusion Models with Ensemble of Expert Denoisers | | 9| Stability AI| The world's leading open source generative AI company which opened source Stable Diffusion | | 10| Meta AI| Whether it be research, product or infrastructure development, we’re driven to innovate responsibly with AI to benefit the world | | 11| ANTHROPIC| AI research and products that put safety at the frontier | [Back to Top] Reports&Papers | No. | Report&Paper | Description | |:---- |:-------------------------------------------------------------------------------------------------------------- |:---------------------------------------------------- | | 1 | GPT-4 Technical Report | GPT-4 Technical Report | | 2 | mli/paper-reading | Deep learning classics and new papers are read carefully paragraph by paragraph. | | 3 | labmlai/annotateddeeplearningpaperimplementations| A collection of simple PyTorch implementations of neural networks and related algorithms, which are documented with explanations | | 4 | Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models | Talking, Drawing and Editing with Visual Foundation Models | | 5 | OpenAI Research | The latest research report and papers from OpenAI | | 6 | Make-A-Video: Text-to-Video Generation without Text-Video Data|Meta's Text-to-Video Generation| | 7 | eDiff-I: Text-to-Image Diffusion Models with Ensemble of Expert Denoisers| Nvidia eDiff-I - New generation of generative AI content creation tool | | 8 | Training an Assistant-style Chatbot with Large Scale Data Distillation from GPT-3.5-Turbo | 2023 GPT4All Technical Report | | 9 | Segment Anything| Meta Segment Anything | | 10 | LLaMA: Open and Efficient Foundation Language Models| LLaMA: a collection of foundation language models ranging from 7B to 65B parameters| | 11 | papers-we-love/papers-we-love |Papers from the computer science community to read and discuss| | 12 | CVPR 2023 papers |The most exciting and influential CVPR 2023 papers| [Back to Top] Tutorials | No. | Tutorial | Description| |:---- |:---------------------------------------------------------------- | --- | | 1 | Coursera - Machine Learning | The Machine Learning Specialization Course taught by Dr. Andrew Ng| | 2 | microsoft/ML-For-Beginners | 12 weeks, 26 lessons, 52 quizzes, classic Machine Learning for all| | 3 | ChatGPT Prompt Engineering for Developers | This short course taught by Isa Fulford (OpenAI) and Andrew Ng (DeepLearning.AI) will teach how to use a large language model (LLM) to quickly build new and powerful applications | | 4 | Dive into Deep Learning |Targeting Chinese readers, functional and open for discussion. The Chinese and English versions are used for teaching in over 400 universities across more than 60 countries | | 5 | AI Expert Roadmap | Roadmap to becoming an Artificial Intelligence Expert in 2022 | | 6 | Computer Science courses |List of Computer Science courses with video lectures| | 7 | Machine Learning with Python | Machine Learning with Python Certification on freeCodeCamp| | 8 | Building Systems with the ChatGPT API | This short course taught by Isa Fulford (OpenAI) and Andrew Ng (DeepLearning.AI), you will learn how to automate complex workflows using chain calls to a large language model| | 9 | LangChain for LLM Application Development | This short course taught by Harrison Chase (Co-Founder and CEO at LangChain) and Andrew Ng. you will gain essential skills in expanding the use cases and capabilities of language models in application development using the LangChain framework| | 10 | How Diffusion Models Work | This short course taught by Sharon Zhou (CEO, Co-founder, Lamini). you will gain a deep familiarity with the diffusion process and the models which carry it out. More than simply pulling in a pre-built model or using an API, this course will teach you to build a diffusion model from scratch| | 11 | Free Programming Books For AI |📚 Freely available programming books for AI | | 12 | microsoft/AI-For-Beginners |12 Weeks, 24 Lessons, AI for All!| | 13 | hemansnation/God-Level-Data-Science-ML-Full-Stack |A collection of scientific methods, processes, algorithms, and systems to build stories & models. This roadmap contains 16 Chapters, whether you are a fresher in the field or an experienced professional who wants to transition into Data Science & AI| | 14 | datawhalechina/prompt-engineering-for-developers |Chinese version of Andrew Ng's Big Model Series Courses, including "Prompt Engineering", "Building System", and "LangChain"| | 15 | ossu/computer-science |🎓 Path to a free self-taught education in Computer Science!| | 16 | microsoft/Data-Science-For-Beginners | 10 Weeks, 20 Lessons, Data Science for All! | |17 |jwasham/coding-interview-university !2023-09-29268215336 |A complete computer science study plan to become a software engineer.| [Back to Top] Thanks If this project has been helpful to you in any way, please give it a ⭐️ by clicking on the star.

obsei
github
LLM Vibe Score0.545
Human Vibe Score0.10175553624190911
obseiMar 27, 2025

obsei

Note: Obsei is still in alpha stage hence carefully use it in Production. Also, as it is constantly undergoing development hence master branch may contain many breaking changes. Please use released version. Obsei (pronounced "Ob see" | /əb-'sē/) is an open-source, low-code, AI powered automation tool. Obsei consists of - Observer: Collect unstructured data from various sources like tweets from Twitter, Subreddit comments on Reddit, page post's comments from Facebook, App Stores reviews, Google reviews, Amazon reviews, News, Website, etc. Analyzer: Analyze unstructured data collected with various AI tasks like classification, sentiment analysis, translation, PII, etc. Informer: Send analyzed data to various destinations like ticketing platforms, data storage, dataframe, etc so that the user can take further actions and perform analysis on the data. All the Observers can store their state in databases (Sqlite, Postgres, MySQL, etc.), making Obsei suitable for scheduled jobs or serverless applications. !Obsei diagram Future direction - Text, Image, Audio, Documents and Video oriented workflows Collect data from every possible private and public channels Add every possible workflow to an AI downstream application to automate manual cognitive workflows Use cases Obsei use cases are following, but not limited to - Social listening: Listening about social media posts, comments, customer feedback, etc. Alerting/Notification: To get auto-alerts for events such as customer complaints, qualified sales leads, etc. Automatic customer issue creation based on customer complaints on Social Media, Email, etc. Automatic assignment of proper tags to tickets based content of customer complaint for example login issue, sign up issue, delivery issue, etc. Extraction of deeper insight from feedbacks on various platforms Market research Creation of dataset for various AI tasks Many more based on creativity 💡 Installation Prerequisite Install the following (if not present already) - Install Python 3.7+ Install PIP Install Obsei You can install Obsei either via PIP or Conda based on your preference. To install latest released version - Install from master branch (if you want to try the latest features) - Note: all option will install all the dependencies which might not be needed for your workflow, alternatively following options are available to install minimal dependencies as per need - pip install obsei[source]: To install dependencies related to all observers pip install obsei[sink]: To install dependencies related to all informers pip install obsei[analyzer]: To install dependencies related to all analyzers, it will install pytorch as well pip install obsei[twitter-api]: To install dependencies related to Twitter observer pip install obsei[google-play-scraper]: To install dependencies related to Play Store review scrapper observer pip install obsei[google-play-api]: To install dependencies related to Google official play store review API based observer pip install obsei[app-store-scraper]: To install dependencies related to Apple App Store review scrapper observer pip install obsei[reddit-scraper]: To install dependencies related to Reddit post and comment scrapper observer pip install obsei[reddit-api]: To install dependencies related to Reddit official api based observer pip install obsei[pandas]: To install dependencies related to TSV/CSV/Pandas based observer and informer pip install obsei[google-news-scraper]: To install dependencies related to Google news scrapper observer pip install obsei[facebook-api]: To install dependencies related to Facebook official page post and comments api based observer pip install obsei[atlassian-api]: To install dependencies related to Jira official api based informer pip install obsei[elasticsearch]: To install dependencies related to elasticsearch informer pip install obsei[slack-api]:To install dependencies related to Slack official api based informer You can also mix multiple dependencies together in single installation command. For example to install dependencies Twitter observer, all analyzer, and Slack informer use following command - How to use Expand the following steps and create a workflow - Step 1: Configure Source/Observer Twitter Youtube Scrapper Facebook Email Google Maps Reviews Scrapper AppStore Reviews Scrapper Play Store Reviews Scrapper Reddit Reddit Scrapper Note: Reddit heavily rate limit scrappers, hence use it to fetch small data during long period Google News Web Crawler Pandas DataFrame Step 2: Configure Analyzer Note: To run transformers in an offline mode, check transformers offline mode. Some analyzer support GPU and to utilize pass device parameter. List of possible values of device parameter (default value auto): auto: GPU (cuda:0) will be used if available otherwise CPU will be used cpu: CPU will be used cuda:{id} - GPU will be used with provided CUDA device id Text Classification Text classification: Classify text into user provided categories. Sentiment Analyzer Sentiment Analyzer: Detect the sentiment of the text. Text classification can also perform sentiment analysis but if you don't want to use heavy-duty NLP model then use less resource hungry dictionary based Vader Sentiment detector. NER Analyzer NER (Named-Entity Recognition) Analyzer: Extract information and classify named entities mentioned in text into pre-defined categories such as person names, organizations, locations, medical codes, time expressions, quantities, monetary values, percentages, etc Translator PII Anonymizer Dummy Analyzer Dummy Analyzer: Does nothing. Its simply used for transforming the input (TextPayload) to output (TextPayload) and adding the user supplied dummy data. Step 3: Configure Sink/Informer Slack Zendesk Jira ElasticSearch Http Pandas DataFrame Logger This is useful for testing and dry running the pipeline. Step 4: Join and create workflow source will fetch data from the selected source, then feed it to the analyzer for processing, whose output we feed into a sink to get notified at that sink. Step 5: Execute workflow Copy the code snippets from Steps 1 to 4 into a python file, for example example.py and execute the following command - Demo We have a minimal streamlit based UI that you can use to test Obsei. !Screenshot Watch UI demo video Check demo at (Note: Sometimes the Streamlit demo might not work due to rate limiting, use the docker image (locally) in such cases.) To test locally, just run To run Obsei workflow easily using GitHub Actions (no sign ups and cloud hosting required), refer to this repo. Companies/Projects using Obsei Here are some companies/projects (alphabetical order) using Obsei. To add your company/project to the list, please raise a PR or contact us via email. Oraika: Contextually understand customer feedback 1Page: Giving a better context in meetings and calls Spacepulse: The operating system for spaces Superblog: A blazing fast alternative to WordPress and Medium Zolve: Creating a financial world beyond borders Utilize: No-code app builder for businesses with a deskless workforce Articles Sr. No. Title Author 1 AI based Comparative Customer Feedback Analysis Using Obsei Reena Bapna 2 LinkedIn App - User Feedback Analysis Himanshu Sharma Tutorials Sr. No. Workflow Colab Binder 1 Observe app reviews from Google play store, Analyze them by performing text classification and then Inform them on console via logger PlayStore Reviews → Classification → Logger 2 Observe app reviews from Google play store, PreProcess text via various text cleaning functions, Analyze them by performing text classification, Inform them to Pandas DataFrame and store resultant CSV to Google Drive PlayStore Reviews → PreProcessing → Classification → Pandas DataFrame → CSV in Google Drive 3 Observe app reviews from Apple app store, PreProcess text via various text cleaning function, Analyze them by performing text classification, Inform them to Pandas DataFrame and store resultant CSV to Google Drive AppStore Reviews → PreProcessing → Classification → Pandas DataFrame → CSV in Google Drive 4 Observe news article from Google news, PreProcess text via various text cleaning function, Analyze them via performing text classification while splitting text in small chunks and later computing final inference using given formula Google News → Text Cleaner → Text Splitter → Classification → Inference Aggregator 💡Tips: Handle large text classification via Obsei Documentation For detailed installation instructions, usages and examples, refer to our documentation. Support and Release Matrix Linux Mac Windows Remark Tests ✅ ✅ ✅ Low Coverage as difficult to test 3rd party libs PIP ✅ ✅ ✅ Fully Supported Conda ❌ ❌ ❌ Not Supported Discussion forum Discussion about Obsei can be done at community forum Changelogs Refer releases for changelogs Security Issue For any security issue please contact us via email Stargazers over time Maintainers This project is being maintained by Oraika Technologies. Lalit Pagaria and Girish Patel are maintainers of this project. License Copyright holder: Oraika Technologies Overall Apache 2.0 and you can read License file. Multiple other secondary permissive or weak copyleft licenses (LGPL, MIT, BSD etc.) for third-party components refer Attribution. To make project more commercial friendly, we void third party components which have strong copyleft licenses (GPL, AGPL etc.) into the project. Attribution This could not have been possible without these open source softwares. Contribution First off, thank you for even considering contributing to this package, every contribution big or small is greatly appreciated. Please refer our Contribution Guideline and Code of Conduct. Thanks so much to all our contributors

bytom
github
LLM Vibe Score0.537
Human Vibe Score0.038940878121795156
BytomDAOMar 14, 2025

bytom

Bytom ====== Official golang implementation of the Bytom protocol. Automated builds are available for stable releases and the unstable master branch. Binary archives are published at https://github.com/Bytom/bytom/releases. What is Bytom? Bytom is software designed to operate and connect to highly scalable blockchain networks confirming to the Bytom Blockchain Protocol, which allows partipicants to define, issue and transfer digitial assets on a multi-asset shared ledger. Please refer to the White Paper for more details. In the current state bytom is able to: Manage key, account as well as asset Send transactions, i.e., issue, spend and retire asset Installing with Homebrew Building from source Requirements Go version 1.8 or higher, with $GOPATH set to your preferred directory Installation Ensure Go with the supported version is installed properly: Get the source code Build source code When successfully building the project, the bytomd and bytomcli binary should be present in cmd/bytomd and cmd/bytomcli directory, respectively. Executables The Bytom project comes with several executables found in the cmd directory. | Command | Description | | ------------ | ------------------------------------------------------------ | | bytomd | bytomd command can help to initialize and launch bytom domain by custom parameters. bytomd --help for command line options. | | bytomcli | Our main Bytom CLI client. It is the entry point into the Bytom network (main-, test- or private net), capable of running as a full node archive node (retaining all historical state). It can be used by other processes as a gateway into the Bytom network via JSON RPC endpoints exposed on top of HTTP, WebSocket and/or IPC transports. bytomcli --help and the bytomcli Wiki page for command line options. | Running bytom Currently, bytom is still in active development and a ton of work needs to be done, but we also provide the following content for these eager to do something with bytom. This section won't cover all the commands of bytomd and bytomcli at length, for more information, please the help of every command, e.g., bytomcli help. Initialize First of all, initialize the node: There are three options for the flag --chain_id: mainnet: connect to the mainnet. testnet: connect to the testnet wisdom. solonet: standalone mode. After that, you'll see config.toml generated, then launch the node. launch available flags for bytomd node: Given the bytomd node is running, the general workflow is as follows: create key, then you can create account and asset. send transaction, i.e., build, sign and submit transaction. query all kinds of information, let's say, avaliable key, account, key, balances, transactions, etc. Dashboard Access the dashboard: In Docker Ensure your Docker version is 17.05 or higher. For the usage please refer to running-in-docker-wiki. Contributing Thank you for considering helping out with the source code! Any contributions are highly appreciated, and we are grateful for even the smallest of fixes! If you run into an issue, feel free to bytom issues in this repository. We are glad to help! License AGPL v3