VibeBuilders.ai Logo
VibeBuilders.ai

Vesting

Explore resources related to vesting to help implement AI solutions for your business.

I single-handedly built the world’s best AI investing platform. Here’s NexusTrade’s 2024 year in review
reddit
LLM Vibe Score0
Human Vibe Score1
No-Definition-2886This week

I single-handedly built the world’s best AI investing platform. Here’s NexusTrade’s 2024 year in review

I copy-pasted the content of this article to save you a click! I’ve been developing an AI investing platform for 4 years, and I’m blown away by all of the new features I’ve gotten done! Here’s my project’s 2024 year in review —- When someone asks me what is the best way to learn how to trade and invest, I have an unbiased answer – NexusTrade.io. I started NexusTrade to empower everybody, including beginners and non-technical investors, to learn how to make smarter investing decisions. NexusTrade is the best way for a new investor to learn algorithmic trading and financial research, and I’m not the only person to think so. Just this year alone, user growth has skyrocketed from 1,703 users to 14,319 users. This is driven by new features, better research tools, and the launch of algorithmic trading. Here’s NexusTrade’s 2024 year in review, a semi-complete list of the features I’ve launched. Summarizing this year in review TL;DR: I implemented a variety of new features to enhance NexusTrade’s algorithmic trading and financial research capabilities. This includes: Cryptocurrency support Enhanced financial research, like the AI-Powered Stock Screener Unique watchlists and daily market summaries Live-trading with Alpaca. Next year, I plan to implement features to make NexusTrade more tailored for each user’s experience, and launch several unique features including copy trading and fully automated algorithmic trading. Feature-by-feature: What have I done so far in 2024? Algorithmic Cryptocurrency Trading Picture: Algorithmic Cryptocurrency Trading I kicked off the year by adding cryptocurrency support to NexusTrade. Users can now research, design, and implement automated strategies for popular cryptocurrencies, such as Bitcoin, Dogecoin, and Ethereum. AI-Powered Stock Screener and research capabilities Picture: AI-Powered Stock Screener In tandem with cryptocurrency support, I made a huge update to Aurora, the AI Assistant in NexusTrade, by implementing a natural language stock screener. This screener makes it easy to find fundamentally strong stocks. Throughout the year, I’ve made several enhancements to it. Over time, I’ve made the screener faster, more accurate, and expanded its capabilities. Using fundamental indicators within trading strategies Picture: Using fundamental indicators Doing financial research for companies isn’t enough; we also need a way to integrate this type of research into trading strategies. Thus, I’ve expanded the NexusTrade indicators, and made it possible to create strategies using metrics like revenue, net income, free cash flow, and P/E ratio. Stock watchlists with tailored, automated daily emails Picture: Stock watchlists In addition, I didn’t want the research you may have done for a stock (or list of stocks) to be forgotten. Thus, I created the most useful watchlist page of any investing platform. This watchlist makes it easy to keep track of your favorite stocks, track them over time, and even receive curated, daily emails about them. Enhanced user profile page, Google sign-ins, and two-factor authentication Picture: Enhanced user profile Keeping in theme with adding new pages to NexusTrade, many pages, such as the profile page, got a huge revamp. The new profile page is cleaner, easier to use, and allows you to secure your account more effectively, for example, by using two-factor authentication. GPT-Reports: an AI-generated analysis of every stock in the market Picture: GPT-Reports I created GPT-Stock Reports, an AI-Generated analysis of every stock in the market. This report was generated by taking each company’s earnings data and asking GPT to analyze the stock and give it a rating. Manual and semi-automated algorithmic trading with Alpaca Picture: Manual and semi-automated trading Finally, I’ve fully launched the Alpaca integration, and enabled users to execute real trades directly in the NexusTrade app! This integration has transformed NexusTrade from a financial research app into a real, algorithmic trading platform for retail investors. Concluding Thoughts When I say that NexusTrade is the best platform for traders and investors to make more money in the stock market, you may naively think that I’m biased. I created the app, and the rose-tinted glasses is bound to make every red flag look like a regular flag, right? Wrong. NexusTrade is objectively a completely new way for investors to approach financial markets. The fact that the app is so expansive is nothing short of miraculous.

160 of Y Combinators 229 Startup Cohort are AI Startups with and 75% of the Cohort has 0 revenue
reddit
LLM Vibe Score0
Human Vibe Score1
DemocratizingfinanceThis week

160 of Y Combinators 229 Startup Cohort are AI Startups with and 75% of the Cohort has 0 revenue

Y Combinator (YC), one of the most prestigious startup accelerators in the world, has just unveiled its latest batch of innovative startups, providing key insights into what the future might hold. Y Combinators Summer 2023 Batch In a recent post by Garry Tan, YC's president, Tan offers a nostalgic look back at his first YC Demo Day in 2008, where he, as a budding entrepreneur, pitched his startup. Now, fifteen years later, he's at the helm, proudly launching the 37th Demo Day, this time for the Summer 2023 batch. Tan proudly declares this batch as one of YC's most impressive yet, emphasizing the deep technical talent of the participants. From a staggering pool of over 24,000 applications, only 229 startups were chosen, making this one of the most competitive batches to date. This batch marks a number of firsts and solidifies several rising trends within the startups landscape. 75% of these companies began their YC journey with zero revenue, and 81% hadn't raised any funding before joining the accelerator. YC's decision to focus on early-stage startups this round signals their commitment to nurturing raw, untapped potential. A Return to Face-to-Face Interaction After three years, YC has brought back the in-person Demo Day format, allowing startups, investors, and mentors to connect directly. While the virtual format has its merits, there's an unmistakable magic in the YC Demo Day room, filled with anticipation, hope, and innovation. AI Takes Center Stage Artificial Intelligence is the standout sector in the Summer 2023 batch. With recent advancements making waves across various industries, there's arguably no better time to launch an AI-focused startup, and no better platform than YC to foster its growth. This signals a clear trend in the startup investing and venture capital space: AI is just getting started. Of the entire Summer 2023 batch, 160 out of the entire 229 Summer 2023 batch that are utilizing or implementing artificial intelligence in some capacity. This means over 2 out of every 3 startups accepted is focused on artificial intelligence in some capacity. Some of the startups include: Quill AI: Automating the job of a financial analyst Fiber AI: Automating prospecting and outbound marketing Reworkd AI: Open Source Zapier of AI Agents Watto AI: AI-powered McKinsey-quality reports in seconds Agentive: AI-powered auditing platform Humanlike: Replace your call center with voice bots that sound human Greenlite: AI compliance team for fintech and banking atla: AI assistants to help in-house lawyers answer legal questions Studdy: An AI Match tutor Glade: League of Legends with AI-generated maps and gameplay and literally over 100 others. As you can see, there's a startup covering nearly every sector of AI in the new batch. YC By The Numbers YC continues to grow as a community. The accelerator now boasts over 10,000 founders spanning more than 4,500 startups. The success stories are impressive: over 350 startups valued at over $150 million and 90 valued at more than $1 billion. The unicorn creation rate of 5% is truly unparalleled in the industry. To cater to the ever-growing community, YC has added more full-time Group Partners than ever. This includes industry veterans such as Tom Blomfield, co-founder of billion-dollar startups GoCardless and Monzo, and YC alumni like Wayne Crosby (Zenter) and Emmett Shear (Twitch). YC Core Values YC's commitment to diversity is evident in the demographics of the S23 batch. They've also spotlighted the industries these startups operate in, with 70% in B2B SaaS/Enterprise, followed by fintech, healthcare, consumer, and proptech/industrials. Garry Tan emphasizes three core tenets for YC investors: to act ethically, to make decisions swiftly, and to commit long-term. He underlines the importance of the YC community, urging investors to provide valuable introductions and guidance to founders. The Road Ahead With YC's track record and the promise shown by the Summer 2023 batch, the future of the startup ecosystem looks promising. As always, YC remains at the forefront, championing innovation and shaping the next generation of global startups. Original Post: https://www.democratizing.finance/post/take-a-peek-into-the-future-with-y-combinators-finalized-summer-2023-batch

I studied how 7 Founders found their first 100 customers for their businesses. Summarizing it here!
reddit
LLM Vibe Score0
Human Vibe Score1
adriannelestrangeThis week

I studied how 7 Founders found their first 100 customers for their businesses. Summarizing it here!

I am learning marketing, and so I combed through the internet to find specific advice that helped founders reach 100 users and not random Google answers. Here’s what I found: Llama Life by Marie Marie founder of Llama Life, a productivity app ($51.4K+ revenue) got her first 100 users using Snowballing effect. She shared great advice that I want to add here verbatim, “Need to think about what you have that you can leverage based on your current situation. eg..When you have no customers, think about where you can post to get the 1st customer eg Product Hunt. If you do well on PH, say you get #3 product of the day, then you post somewhere else saying ‘I got #3 product of the day’.. to get your next few customers. Maybe that post is on reddit with some learnings that you found. If the reddit post does well, then you might post it on Twitter, saying reddit did well and what learnings you got from that etc. or even if it doesn’t do well you can still post about it.” Another tip she shared is to build related products that get more viral than the product itself. These are small stand-alone sites that would appeal to the same target audience, but by nature, are more shareable. On these sites, you can mention your startup like: ‘brought to you by Llama Life’ and then provide a link to the main website if someone is interested. If one of those gets viral or ranks on Google, you’ll have a passive traffic source. Scraping bee by Pierre Pierre, founder of Scraping Bee, a web scraping tool has now reached $1.5M ARR. Pierre and his cofounder Kevin started with 10 Free Beta Users in 2019, and after 6 months asked them to take a paid subscription if they wanted to continue using the product. That’s how they got their first user within 50 minutes of that email. Then they listed it on dozens of startup directories but their core strategy was writing the best possible content for their target audience — Developers. 3 very successful pieces of content that worked were : A small tutorial on how to scrape single-page application An extensive general guide about web scraping without getting blocked A complete introduction to web scraping with Python They didn’t do content marketing for the sake of content marketing but deep-dived into the value they were providing their customer. One of these got 70K visits, and all this together got them to over 100 users. WePay by Bill Clerico Bill Clerico left his cushy corporate job to build WePay which was then acquired for $400M got his first users by using his app. He got his first users by using his app! The app was for group payments. So he hosted a Poker tournament at his house and collected payments only with his app. Then they hosted a barbecue for fraternity treasurers at San Jose State & helped them do their annual dues collection. Good old word-of-mouth marketing, that however, started with an event where they used what they made! RealWorld by Genevieve Genevieve — Founder and CEO of Realworld stands by the old-school advice of value giving. RealWorld is an app that helps GenZ navigate adulthood. So, before launching their direct-to-consumer platform, they had an educational course that they sold to college career centers and students. They already had a pipeline of adults who turned to Realworld for their adulting challenges. From there, she gained her first 100 followers. Saner dot ai by Austin Austin got 100 users from Reddit for his startup Saner.ai. Reddit hates advertising, and so his tips to market your startup on Reddit is to Write value-driven posts on your niche. Instead of writing posts, find posts where people are looking for solutions DM people facing problems that your SaaS solves. But instead of selling, ask about their problem to see if your product is a good fit Heartfelt posts about why you built it, aren’t gonna cut it To find posts and people, search Reddit with relevant keywords and join all the subreddits A Stock Portfolio Newsletter A financial investor got his first 100 paid newsletter subscribers for his stock portfolio newsletter. His tips : Don’t reinvent the wheel. Work what’s already working. He saw a company making $500M+ from stock picking newsletter, so decided to try that. Find the gaps in “already working” and leverage them. That newsletter did not have portfolios of advisors writing them. That was his USP. He added his own portfolio to his newsletter. Now to 100 users, he partnered with a guy running an investing website and getting good traffic. That guy got a cut of his revenue, in exchange. That one simple step got him to 100 users. Hypefury by Yannick and Samy Yannick and Samy from Hypefury, Twitter and Social Media Automation tool got their first beta testers and users from a paid community. They launched Hypefury there and asked if someone wanted to try it. A couple of people tried it and gave feedback. Samy conducted user interviews and product demos for them, And shared the reviews on Twitter. That alone, along with word-of-mouth marketing on Twitter got them their first 100 users. To conclude: Don’t reinvent the wheel, try what’s working. Find the gaps in what’s working, and leverage that. Instead of thinking about millions of customers, think about the first 10. Then first 100. Leverage what you have. Get the first 10 customers, then talk about this to get the next 100. Use your app. Find ways, events, and opportunities to use your app in front of people. And get them to use it. Write content not only for SEO but also to help people. It won’t work tomorrow, but it will work for years after it picks up. Leverage other sources of traffic by partnering up! Do things that don’t scale. I’m also doing SaaS marketing deep dives over 30 pieces of content. I'm posting here for the first time, so I'm not sure if it will stay or not, sorry if it doesn't. I've helped a SaaS grow from $19K to $100K MRR as a marketer in last 2 years, and now I wanna dive deep. Cheers! (1/30)

I studied how 7 Founders found their first 100 customers for their businesses. Summarizing it here!
reddit
LLM Vibe Score0
Human Vibe Score1
adriannelestrangeThis week

I studied how 7 Founders found their first 100 customers for their businesses. Summarizing it here!

I am learning marketing, and so I combed through the internet to find specific advice that helped founders reach 100 users and not random Google answers. Here’s what I found: Llama Life by Marie Marie founder of Llama Life, a productivity app ($51.4K+ revenue) got her first 100 users using Snowballing effect. She shared great advice that I want to add here verbatim, “Need to think about what you have that you can leverage based on your current situation. eg..When you have no customers, think about where you can post to get the 1st customer eg Product Hunt. If you do well on PH, say you get #3 product of the day, then you post somewhere else saying ‘I got #3 product of the day’.. to get your next few customers. Maybe that post is on reddit with some learnings that you found. If the reddit post does well, then you might post it on Twitter, saying reddit did well and what learnings you got from that etc. or even if it doesn’t do well you can still post about it.” Another tip she shared is to build related products that get more viral than the product itself. These are small stand-alone sites that would appeal to the same target audience, but by nature, are more shareable. On these sites, you can mention your startup like: ‘brought to you by Llama Life’ and then provide a link to the main website if someone is interested. If one of those gets viral or ranks on Google, you’ll have a passive traffic source. Scraping bee by Pierre Pierre, founder of Scraping Bee, a web scraping tool has now reached $1.5M ARR. Pierre and his cofounder Kevin started with 10 Free Beta Users in 2019, and after 6 months asked them to take a paid subscription if they wanted to continue using the product. That’s how they got their first user within 50 minutes of that email. Then they listed it on dozens of startup directories but their core strategy was writing the best possible content for their target audience — Developers. 3 very successful pieces of content that worked were : A small tutorial on how to scrape single-page application An extensive general guide about web scraping without getting blocked A complete introduction to web scraping with Python They didn’t do content marketing for the sake of content marketing but deep-dived into the value they were providing their customer. One of these got 70K visits, and all this together got them to over 100 users. WePay by Bill Clerico Bill Clerico left his cushy corporate job to build WePay which was then acquired for $400M got his first users by using his app. He got his first users by using his app! The app was for group payments. So he hosted a Poker tournament at his house and collected payments only with his app. Then they hosted a barbecue for fraternity treasurers at San Jose State & helped them do their annual dues collection. Good old word-of-mouth marketing, that however, started with an event where they used what they made! RealWorld by Genevieve Genevieve — Founder and CEO of Realworld stands by the old-school advice of value giving. RealWorld is an app that helps GenZ navigate adulthood. So, before launching their direct-to-consumer platform, they had an educational course that they sold to college career centers and students. They already had a pipeline of adults who turned to Realworld for their adulting challenges. From there, she gained her first 100 followers. Saner dot ai by Austin Austin got 100 users from Reddit for his startup Saner.ai. Reddit hates advertising, and so his tips to market your startup on Reddit is to Write value-driven posts on your niche. Instead of writing posts, find posts where people are looking for solutions DM people facing problems that your SaaS solves. But instead of selling, ask about their problem to see if your product is a good fit Heartfelt posts about why you built it, aren’t gonna cut it To find posts and people, search Reddit with relevant keywords and join all the subreddits A Stock Portfolio Newsletter A financial investor got his first 100 paid newsletter subscribers for his stock portfolio newsletter. His tips : Don’t reinvent the wheel. Work what’s already working. He saw a company making $500M+ from stock picking newsletter, so decided to try that. Find the gaps in “already working” and leverage them. That newsletter did not have portfolios of advisors writing them. That was his USP. He added his own portfolio to his newsletter. Now to 100 users, he partnered with a guy running an investing website and getting good traffic. That guy got a cut of his revenue, in exchange. That one simple step got him to 100 users. Hypefury by Yannick and Samy Yannick and Samy from Hypefury, Twitter and Social Media Automation tool got their first beta testers and users from a paid community. They launched Hypefury there and asked if someone wanted to try it. A couple of people tried it and gave feedback. Samy conducted user interviews and product demos for them, And shared the reviews on Twitter. That alone, along with word-of-mouth marketing on Twitter got them their first 100 users. To conclude: Don’t reinvent the wheel, try what’s working. Find the gaps in what’s working, and leverage that. Instead of thinking about millions of customers, think about the first 10. Then first 100. Leverage what you have. Get the first 10 customers, then talk about this to get the next 100. Use your app. Find ways, events, and opportunities to use your app in front of people. And get them to use it. Write content not only for SEO but also to help people. It won’t work tomorrow, but it will work for years after it picks up. Leverage other sources of traffic by partnering up! Do things that don’t scale. I’m also doing SaaS marketing deep dives over 30 pieces of content. I'm posting here for the first time, so I'm not sure if it will stay or not, sorry if it doesn't. I've helped a SaaS grow from $19K to $100K MRR as a marketer in last 2 years, and now I wanna dive deep. Cheers! (1/30)

Building in the open with Founder University - I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
Tim-SylvesterThis week

Building in the open with Founder University - I will not promote

Published Oct 30, 2024 I am on my fifth startup. I ran the last one for a decade, that’s a whole story. A hell of a story. But a different story. I’ll tell it to you when I can, but not right now. The one before that was an e-commerce site that did pretty well but I didn’t love it. Before that were two service businesses. The first one I did for the love of the game, the second one was an attempt to make people stop asking me to fix their computer by charging them outrageous prices, which backfired horribly when they were eager to pay. None are relevant except to say I’ve been around the block and have the scars to prove it. When it was time to get back out there, I wanted to use all I’ve learned to do better. Before I talk about what those lessons produced, I’m going to talk about what those lessons were. Cause before effect, after all. One thing I wanted to do better this time was pattern matching - making the startup look the way that the industry and investors “expect” a startup to look. My last startup was an awesome idea with awesome tech (still is, but like I said, another story), but that one didn’t match patterns. It didn’t match investor patterns, industry buying patterns, patterns of existing, immediate, recognized and admitted needs. Because it didn’t “look” right to anyone, everything about it was way harder than necessary. The “make it look right” approach runs the risk of building a cargo cult, imitating the trappings of something but without understanding the essence of that something, but then again, a thing that looks like a knife is going to make a better knife that a thing that looks like a bowling ball, so sometimes just sharing apparent similarities can get you pretty far, even if it doesn’t get you all the way there. Like how mimicking someone’s accent makes it easier for them to understand you. For this one, I wanted to adopt every tool, method, and pattern that I knew “the industry” wanted to see to minimize the friction from development, go-to-market, scaling, adoption, and that would make investment optional (and, therefore, available if desired) instead of necessary (and, therefore, largely unavailable). That required establishing some expectations for successful patterns I could match against. What patterns am I matching to? Here’s a general sketch of my pattern matching thought process: Software first and software only. It’s the easiest industry to start a business in, lowest startup costs, and easiest customer acquisition. I wanted to build software for an element of the industry that’s actively emerging (and therefore has room to grow) and part of an optimistic investor thesis (and therefore has a cohort of people who are intent on injecting capital into the market to help it grow). It needs to fills a niche that is underexplored (low competition) and highly potent (lots of opportunity), while being aligned to recognized and emerging needs within the industry (readily adopted). I wanted it to have evidence supporting the business thesis that proves the demand exists, but demonstrates that the demand is unanswered (as of yet) by sufficient or adequate supply.* I wanted the lowest number of dominoes to line up and tip for everything to work correctly - the more dominoes in the line, the less likely the last one will fall. I wanted to implement modern toolsets for everything, wherever possible. I wanted to obey the maxim, “When there’s a gold rush, don’t mine the gold, sell the picks and shovels.” Whatever I chose would need to produce cash flow almost immediately with minimal development time or go-to-market delays, because the end of ZIRP killed the “trust me bro” investment thesis predominant over the last 15 years. I wanted to match to YC best practices, not because YC can predict what will definitely work, but because they’ve churned through so many startups in the last 15 years that they have a good sense of what will definitely not work. And I wanted to build client-centric, because if my intent is to to produce cash flow immediately, we need to get clients immediately, and if we need to get clients immediately, we need to focus on what clients need right now. Extra credit: What’s the difference between a customer and a client? Note: Competition is awesome! Competition is validating and not scary, because competition proves a market exists. But competition, especially mature competition against an immature startup, makes it harder to break into a space. A first mover advantage isn’t everything, but seeing demand before it’s sufficiently supplied is a great advantage if you’re capital constrained or otherwise unproven. Think about how much money the first guy to sell fidget spinners or Silly Bandz made versus how much money the last guy to order a pallet of each made. Finding demand that exists already but is as of yet insufficiently satisfied is a great place to start. What opportunity spaces are most relevant? The industries and markets I chose to observe were: AI, because if I’m following a theme & pattern for today, it’s AI. Fintech, because cash is king, and fintech puts your hands on cash flow. Crypto/blockchain, because that’s the “new” fintech (or maybe the “old-new” fintech?), and crypto creates powerful incentives and capital formation strategies, along with a lot of flexibility for transaction systems. Tools, particularly unmet demand in tools, that enable these industries. If you wanted to do some brief and simple homework, you could map each of those bullets to several of the numbered list items preceding them. The reasoning was pretty simplistic - AI is what people want to build and invest in now, while fintech and crypto/blockchain are what people were building and investing in for the last major investment thesis. That means that there’s demand in the market for AI and AI-adjacent startups, while there’s a glut of underutilized and highly developed tools within fintech and crypto/blockchain, with a lot of motivated capital behind the adoption. When someone is thinking “I built this thing and not enough people are using it”, and you then build something that uses it creates a great way to find allies. This rationale harnesses technology that is being built and financed now (which means it needs tools and support methods, and a lot of other “picks and shovels”), while leveraging technology that was recently built and financed and is eager for more widespread adoption of the existing toolkits, which makes it suitable for using to build the AI-adjacent tools that are in demand now. It’s like two harmonics producing constructive interference - it makes two waves into one larger wave, which gives me more momentum to surf against. This was a learning process, and I iterated against my general paradigm repeatedly as I learned more. Neither of us have the patience to go through that in excruciating detail, so I’ll cover the highlights in my next post. Extra credit answer: A customer gets a product, a client gets a service. Challenge: Is software a product or a service?

36 startup ideas found by analyzing podcasts (problem, solution & source episode)
reddit
LLM Vibe Score0
Human Vibe Score1
joepigeonThis week

36 startup ideas found by analyzing podcasts (problem, solution & source episode)

Hey, I've been a bit of a podcast nerd for a long time. Around a year ago I began experimenting with transcription of podcasts for a SaaS I was running. I realized pretty quickly that there's a lot of knowledge and value in podcast discussions that is for all intents and purposes entirely unsearchable or discoverable to most people. I ended up stopping work on that SaaS product (party for lack of product/market fit, and partly because podcasting was far more interesting), and focusing on the podcast technology full-time instead. I'm a long-time lurker and poster of r/startups and thought this would make for some interesting content and inspiration for folks. Given I'm in this space, have millions of transcripts, and transcribe thousands daily... I've been exploring fun ways to expose some of the interesting knowledge and conversations taking place that utilize our own data/API. I'm a big fan of the usual startup podcasts (My First Million, Greg Isenberg, etc. etc.) and so I built an automation that turns all of the startup ideas discussed into a weekly email digest. I always struggle to listen to as many episodes as I'd actually like to, so I thought I'd summarise the stuff I care about instead (startup opportunities being discussed). I thought it would be interesting to post some of the ideas extracted so far. They range from being completely whacky and blue sky, to pretty boring but realistic. A word of warning before anyone complains – this is a big mixture of tech, ai, non-tech, local services, etc. ideas: Some of the ideas are completely mundane, but realistic (e.g. local window cleaning service) Some of the ideas are completely insane, blue sky, but sound super interesting Here's the latest 36 ideas: |Idea Name|Problem|Solution|Source| |:-|:-|:-|:-| |SalesForce-as-a-Service - White Label Enterprise Sales Teams|White-label enterprise sales teams for B2B SaaS. Companies need sales but can't hire/train. Recruit retail sellers, train for tech, charge 30% of deals closed.|Create a white-label enterprise sales team by recruiting natural salespeople from retail and direct sales backgrounds (e.g. mall kiosks, cutco knives). Train them specifically in B2B SaaS sales techniques and processes. Offer this trained sales force to tech companies on a contract basis.|My First Million - "Life Hacks From The King of Introverts + 7 Business Ideas| |TechButler - Mobile Device Maintenance Service|Mobile tech maintenance service. Clean/optimize devices, improve WiFi, basic support. $100/visit to homes. Target affluent neighborhoods.|Mobile tech support service providing in-home device cleaning, optimization, and setup. Focus on common issues like WiFi improvement, device maintenance, and basic tech support.|My First Million - "Life Hacks From The King of Introverts + 7 Business Ideas| |MemoryBox - At-Home Video Digitization Service|Door-to-door VHS conversion service. Parents have boxes of old tapes. Pick up, digitize, deliver. $30/tape with minimum order. Going extinct.|Door-to-door VHS to digital conversion service that handles everything from pickup to digital delivery. Make it extremely convenient for customers to preserve their memories.|My First Million - "Life Hacks From The King of Introverts + 7 Business Ideas| |Elite Match Ventures - Success-Based Luxury Matchmaking|High-end matchmaking for 50M+ net worth individuals. Only charge $1M+ when they get married. No upfront fees. Extensive vetting process.|Premium matchmaking service exclusively for ultra-high net worth individuals with a pure contingency fee model - only get paid ($1M+) upon successful marriage. Focus on quality over quantity with extensive vetting and personalized matching.|My First Million - "Life Hacks From The King of Introverts + 7 Business Ideas| |LocalHost - Simple Small Business Websites|Simple WordPress sites for local businesses. $50/month includes hosting, updates, security. Target restaurants and shops. Recurring revenue play.|Simplified web hosting and WordPress management service targeting local small businesses. Focus on basic sites with standard templates, ongoing maintenance, and reliable support for a fixed monthly fee.|My First Million - "Life Hacks From The King of Introverts + 7 Business Ideas| |VoiceJournal AI - Voice-First Smart Journaling|Voice-to-text journaling app with AI insights. 8,100 monthly searches. $15/month subscription. Partners with journaling YouTubers.|AI-powered journaling app that combines voice recording, transcription, and intelligent insights. Users can speak their thoughts, which are automatically transcribed and analyzed for patterns, emotions, and actionable insights.|Where It Happens - "7 $1M+ AI startup ideas you can launch tomorrow with $0"| |AIGenAds - AI-Generated UGC Content Platform|AI platform turning product briefs into UGC-style video ads. Brands spending $500/video for human creators. Generate 100 variations for $99/month.|AI platform that generates UGC-style video ads using AI avatars and scripting. System would allow rapid generation of multiple ad variations at a fraction of the cost. Platform would use existing AI avatar technology combined with script generation to create authentic-looking testimonial-style content.|Where It Happens - "7 $1M+ AI startup ideas you can launch tomorrow with $0"| |InfographAI - Automated Infographic Generation Platform|AI turning blog posts into branded infographics. Marketers spending hours on design. $99/month unlimited generation.|AI-powered platform that automatically converts blog posts and articles into visually appealing infographics. System would analyze content, extract key points, and generate professional designs using predefined templates and brand colors.|Where It Happens - "7 $1M+ AI startup ideas you can launch tomorrow with $0"| |KidFinance - Children's Financial Education Entertainment|Children's media franchise teaching financial literacy. Former preschool teacher creating 'Dora for money'. Books, videos, merchandise potential.|Character-driven financial education content for kids, including books, videos, and potentially TV show. Focus on making money concepts fun and memorable.|The Side Hustle Show - "How a Free Challenge Turned Into a $500,000 a Year Business (Greatest Hits)"| |FinanceTasker - Daily Financial Task Challenge|Free 30-day financial challenge with daily action items. People overwhelmed by money management. Makes $500k/year through books, speaking, and premium membership.|A free 30-day financial challenge delivering one simple, actionable task per day via email. Each task includes detailed scripts and instructions. Participants join a Facebook community for support and accountability. The program focuses on quick wins to build momentum. Automated delivery allows scaling.|The Side Hustle Show - "How a Free Challenge Turned Into a $500,000 a Year Business (Greatest Hits)"| |FinanceAcademy - Expert Financial Training Platform|Premium financial education platform. $13/month for expert-led courses and live Q&As. 4000+ members generating $40k+/month.|Premium membership site with expert-led courses, live Q&As, and community support. Focus on specific topics like real estate investing, business creation, and advanced money management.|The Side Hustle Show - "How a Free Challenge Turned Into a $500,000 a Year Business (Greatest Hits)"| |SecurityFirst Compliance - Real Security + Compliance Platform|Security-first compliance platform built by hackers. Companies spending $50k+ on fake security. Making $7M/year showing why current solutions don't work.|A compliance platform built by security experts that combines mandatory compliance requirements with real security measures. The solution includes hands-on security testing, expert guidance, and a focus on actual threat prevention rather than just documentation. It merges traditional compliance workflows with practical security implementations.|In the Pit with Cody Schneider| |LinkedInbound - Automated Professional Visibility Engine|LinkedIn automation for inbound job offers. Professionals spending hours on manual outreach. $99/month per job seeker.|Automated system for creating visibility and generating inbound interest on LinkedIn through coordinated profile viewing and engagement. Uses multiple accounts to create visibility patterns that trigger curiosity and inbound messages.|In the Pit with Cody Schneider| |ConvoTracker - Community Discussion Monitoring Platform|Community discussion monitoring across Reddit, Twitter, HN. Companies missing sales opportunities. $499/month per brand tracked.|Comprehensive monitoring system that tracks competitor mentions and industry discussions across multiple platforms (Reddit, Twitter, Hacker News, etc.) with automated alerts and engagement suggestions.|In the Pit with Cody Schneider| |ContentAds Pro - Smart Display Ad Implementation|Display ad implementation service for content creators. Bloggers losing thousands in ad revenue monthly. Makes $3-5k per site setup plus ongoing optimization fees.|Implementation of professional display advertising through networks like Mediavine that specialize in optimizing ad placement and revenue while maintaining user experience. Include features like turning off ads for email subscribers and careful placement to minimize impact on core metrics.|The Side Hustle Show - "636: Is Business Coaching Worth It? A Look Inside the last 12 months of Side Hustle Nation"| |MoneyAppReviews - Professional Side Hustle App Testing|Professional testing service for money-making apps. People wasting time on low-paying apps. Makes $20k/month from affiliate commissions and ads.|Professional app testing service that systematically reviews money-making apps and creates detailed, honest reviews including actual earnings data, time investment, and practical tips.|The Side Hustle Show - "636: Is Business Coaching Worth It? A Look Inside the last 12 months of Side Hustle Nation"| |LightPro - Holiday Light Installation Service|Professional Christmas light installation service. Homeowners afraid of ladders. $500-2000 per house plus storage.|Professional Christmas light installation service targeting residential and commercial properties. Full-service offering including design, installation, maintenance, removal and storage. Focus on safety and premium aesthetic results.|The Side Hustle Show - "639: 30 Ways to Make Extra Money for the Holidays"| |FocusMatch - Research Participant Marketplace|Marketplace connecting companies to paid research participants. Companies spending weeks finding people. $50-150/hour per study.|Online platform connecting companies directly with paid research participants. Participants create detailed profiles and get matched to relevant studies. Companies get faster access to their target demographic while participants earn money sharing opinions.|The Side Hustle Show - "639: 30 Ways to Make Extra Money for the Holidays"| |SolarShine Pro - Specialized Solar Panel Cleaning Service|Solar panel cleaning service using specialized equipment. Panels lose 50% efficiency when dirty. $650 per job, automated scheduling generates $18k/month from repeat customers.|Professional solar panel cleaning service using specialized deionized water system and European cleaning equipment. Includes automated 6-month scheduling, professional liability coverage, and warranty-safe cleaning processes. Service is bundled with inspection and performance monitoring.|The UpFlip Podcast - "156. $18K/Month with This ONE Service — Niche Business Idea"| |ExteriorCare Complete - One-Stop Exterior Maintenance Service|One-stop exterior home cleaning service (solar, windows, gutters, bird proofing). Automated scheduling. $650 average ticket. 60% repeat customers on 6-month contracts.|All-in-one exterior cleaning service offering comprehensive maintenance packages including solar, windows, gutters, roof cleaning and bird proofing. Single point of contact, consistent quality, and automated scheduling for all services.|The UpFlip Podcast - "156. $18K/Month with This ONE Service — Niche Business Idea"| |ContentMorph - Automated Cross-Platform Content Adaptation|AI platform converting blog posts into platform-optimized social content. Marketing teams spending 5hrs/post on manual adaptation. $199/mo per brand with 50% margins.|An AI-powered platform that automatically transforms long-form content (blog posts, podcasts, videos) into platform-specific formats (Instagram reels, TikToks, tweets). The system would preserve brand voice while optimizing for each platform's unique requirements and best practices.|Entrepreneurs on Fire - "Digital Threads: The Entrepreneur Playbook for Digital-First Marketing with Neal Schaffer"| |MarketerMatch - Verified Digital Marketing Talent Marketplace|Marketplace for pre-vetted digital marketing specialists. Entrepreneurs spending 15hrs/week on marketing tasks. Platform takes 15% commission averaging $900/month per active client.|A specialized marketplace exclusively for digital marketing professionals, pre-vetted for specific skills (video editing, social media, SEO, etc.). Platform includes skill verification, portfolio review, and specialization matching.|Entrepreneurs on Fire - "Digital Threads: The Entrepreneur Playbook for Digital-First Marketing with Neal Schaffer"| |Tiger Window Cleaning - Premium Local Window Service|Local window cleaning service targeting homeowners. Traditional companies charging 2x market rate. Making $10k/month from $200 initial investment.|Local window cleaning service combining competitive pricing ($5/pane), excellent customer service, and quality guarantees. Uses modern tools like water-fed poles for efficiency. Implements systematic approach to customer communication and follow-up.|The Side Hustle Show - "630: How this College Student’s Side Hustle Brings in $10k a Month"| |RealViz3D - Real Estate Visualization Platform|3D visualization service turning architectural plans into photorealistic renderings for real estate agents. Agents struggling with unbuilt property sales. Making $30-40k/year per operator.|Professional 3D modeling and rendering service that creates photorealistic visualizations of properties before they're built or renovated. The service transforms architectural plans into immersive 3D representations that show lighting, textures, and realistic details. This helps potential buyers fully understand and connect with the space before it physically exists.|Side Hustle School - "#2861 - TBT: An Architect’s Side Hustle in 3D Real Estate Modeling"| |Somewhere - Global Talent Marketplace|Platform connecting US companies with vetted overseas talent. Tech roles costing $150k locally filled for 50% less. Grew from $15M to $52M valuation in 9 months.|Platform connecting US companies with pre-vetted overseas talent at significantly lower rates while maintaining high quality. Handles payments, contracts, and quality assurance to remove friction from global hiring.|My First Million - "I Lost Everything Twice… Then Made $26M In 18 Months| |GymLaunch - Rapid Gym Turnaround Service|Consultants flying to struggling gyms to implement proven member acquisition systems. Gym owners lacking sales expertise. Made $100k in first 21 days.|Expert consultants fly in to implement proven member acquisition systems, train staff, and rapidly fill gyms with new members. The service combines sales training, marketing automation, and proven conversion tactics to transform struggling gyms into profitable businesses within weeks.|My First Million - "I Lost Everything Twice… Then Made $26M In 18 Months| |PublishPlus - Publishing Backend Monetization|Backend monetization system for publishing companies. One-time customers becoming recurring revenue. Grew business from $2M to $110M revenue.|Add complementary backend products and services to increase customer lifetime value. Develop software tools and additional services that natural extend from initial publishing product. Focus on high-margin recurring revenue streams.|My First Million - "I Lost Everything Twice… Then Made $26M In 18 Months| |WelcomeBot - Automated Employee Onboarding Platform|Automated employee welcome platform. HR teams struggling with consistent onboarding. $99/month per 100 employees.|An automated onboarding platform that creates personalized welcome experiences through pre-recorded video messages, scheduled check-ins, and automated swag delivery. The platform would ensure consistent high-quality onboarding regardless of timing or location.|Entrepreneurs on Fire - "Free Training on Building Systems and Processes to Scale Your Business with Chris Ronzio: An EOFire Classic from 2021"| |ProcessBrain - Business Knowledge Documentation Platform|SaaS platform turning tribal knowledge into documented processes. Business owners spending hours training new hires. $199/month per company.|A software platform that makes it easy to document and delegate business processes and procedures. The platform would include templates, guided documentation flows, and tools to easily share and update procedures. It would help businesses create a comprehensive playbook of their operations.|Entrepreneurs on Fire - "Free Training on Building Systems and Processes to Scale Your Business with Chris Ronzio: An EOFire Classic from 2021"| |TradeMatch - Modern Manufacturing Job Marketplace|Modern job board making manufacturing sexy again. Factory jobs paying $40/hr but can't recruit. $500 per successful referral.|A specialized job marketplace and recruitment platform focused exclusively on modern manufacturing and trade jobs. The platform would combine TikTok-style content marketing, referral programs, and modern UX to make manufacturing jobs appealing to Gen Z and young workers. Would leverage existing $500 referral fees and industry demand.|My First Million - "He Sold His Company For $15M, Then Got A Job At McDonald’s"| |GroundLevel - Executive Immersion Program|Structured program putting CEOs in front-line jobs. Executives disconnected from workers. $25k per placement.|A structured program that places executives and founders in front-line jobs (retail, warehouse, service) for 2-4 weeks with documentation and learning framework. Similar to Scott Heiferman's McDonald's experience but productized.|My First Million - "He Sold His Company For $15M, Then Got A Job At McDonald’s"| |OneStepAhead - Micro-Mentorship Marketplace|Marketplace for 30-min mentorship calls with people one step ahead. Professionals seeking specific guidance. Takes 15% of session fees.|MicroMentor Marketplace - Platform connecting people with mentors who are just one step ahead in their journey for focused, affordable micro-mentorship sessions.|Entrepreneurs on Fire - "How to Create an Unbroken Business with Michael Unbroken: An EOFire Classic from 2021"| |VulnerableLeader - Leadership Authenticity Training Platform|Leadership vulnerability training platform. Leaders struggling with authentic communication. $2k/month per company subscription.|Leadership Vulnerability Platform - A digital training platform combining assessment tools, guided exercises, and peer support to help leaders develop authentic communication skills. The platform would include real-world scenarios, video coaching, and measurable metrics for tracking leadership growth through vulnerability.|Entrepreneurs on Fire - "How to Create an Unbroken Business with Michael Unbroken: An EOFire Classic from 2021"| |NetworkAI - Smart Network Intelligence Platform|AI analyzing your network to find hidden valuable connections. Professionals missing opportunities in existing contacts. $49/month per user.|AI Network Navigator - Smart tool that analyzes your professional network across platforms, identifies valuable hidden connections, and suggests specific actionable ways to leverage relationships for mutual benefit.|Entrepreneurs on Fire - "How to Create an Unbroken Business with Michael Unbroken: An EOFire Classic from 2021"| |Porch Pumpkins - Seasonal Decoration Service|Full-service porch pumpkin decoration. Homeowners spend $300-1350 per season. One operator making $1M in 8 weeks seasonal revenue.|Full-service seasonal porch decoration service focused on autumn/Halloween, including design, installation, maintenance, and removal. Offering premium curated pumpkin arrangements with various package tiers.|My First Million - "The guy who gets paid $80K/yr to do nothing"| |Silent Companion - Professional Presence Service|Professional silent companions for lonely people. Huge problem in Japan/globally. $68/session, $80k/year per companion. Non-sexual, just presence.|A professional companion service where individuals can rent a non-judgmental, quiet presence for various activities. The companion provides silent company without the pressure of conversation or social performance. They accompany clients to events, meals, or just sit quietly together.|My First Million - "The guy who gets paid $80K/yr to do nothing"| Hope this is useful. If anyone would like to ensure I include any particular podcasts or episodes etc. in future posts, very happy to do so. I'll generally send \~5 ideas per week in a short weekly digest format (you can see the format I'd usually use in here: podcastmarketwatch.beehiiv.com). I find it mindblowing that the latest models with large context windows make it even possible to analyze full transcripts at such scale. It's a very exciting time we're living through! Would love some feedback on this stuff, happy to iterate and improve the analysis/ideas... or create a new newsletter on a different topic if anyone would like. Cheers!

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
GummyBear8659This week

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)

Hey everyone! Long-time lurker, first-time founder here. I’m reaching out to get feedback on a recent startup experience—what went wrong, what I could have done better, and how I should approach future opportunities. The Background There were three founders in this venture: • Founder A (CEO, 50%) – The product/growth guy who identified the problem space. • Founder B (Me, CTO, 37.5%) – A software engineer with a software dev shop and multiple clients. I wanted to diversify into building my own products but am not inherently a “product person.” • Founder C (COO, 12.5%) – Brought into the mix by Founder A, with the goal of leveraging his network for traction once the product was built. The idea was to create Product X, a solution targeting the SMB space while competitors were moving upmarket. It wasn’t revolutionary—more of a strategic market play. The Initial Plan & My Role • Founder A would define and prioritize product specs, guiding what needed to be built. • I (Founder B) didn’t have time to code myself, so I allocated engineers from my dev shop (which I personally paid for). My stake was adjusted from 32.5% to 37.5% to reflect this contribution. • Founder C was more of an observer early on, planning to help with traction once we had a product ready. We agreed on a 1-year cliff and a 4-year vesting schedule for equity. Where Things Started to Go Wrong • Lack of a Clear Product Roadmap – Founder A was very focused on getting something built fast, but we never signed off on a structured roadmap or milestones. I underestimated the complexity of what was actually needed for customer conversations. • Engineering Expectations vs. Reality – The team (one part-time lead + two full-time juniors from my dev shop) faced early feedback that development was too slow. In response, I ramped up the lead to full-time and added a part-time PM. But Founder A continued pushing for speed, despite real hurdles (OAuth integrations, etc.). • Shifting MVP Goalposts – Midway, Founder A concluded that an MVP wouldn’t cut it—we needed a more complete product to be competitive. This meant more engineering, more delays, and more of my own money spent on development. The Breaking Point Near the 1-year vesting mark, we had an opportunity: a paying client willing to fund an app. I didn’t have devs on the bench, so I asked Founder A to hold off our project briefly while I hired more engineers to avoid stalling either effort. This was the final straw. Founder A (with Founder C somewhat aligned) decided the arrangement wasn’t working—citing past disagreements and the “slowness” issue. The decision was made to end the partnership. Now, Founder A, as majority holder, is requesting a full handover of the code, Founder C is indifferent, and all engineering costs I covered are essentially lost. Key Takeaways (So Far) Crystal-Clear Agreements Upfront – A formalized product roadmap and timeline should’ve been locked in from day one. Business Needs > Engineering Standards – I wanted to build something solid and scalable, but in an early-stage startup, speed to market is king. This was before AI tools became mainstream, so our approach wasn’t as optimized. Don’t Overextend Without Protection – I personally financed all engineering, but without clear safeguards, that investment became a sunk cost. Expenses Must Be Distributed – I was solely covering engineering salaries, which created an imbalance in financial risk. Future partnerships should ensure costs are shared proportionally, rather than one person shouldering the burden. Where I Need Advice Looking back, I want to improve as an engineer, CEO, and co-founder. • What should I have done differently in structuring this partnership? • How do you balance engineering quality with the startup need for speed? • As a dev shop owner, how can I better navigate equity deals where I’m also bringing in engineering resources? I really appreciate everyone who went through this long post and provide any insights from founders, engineers, or anyone who has been in a similar situation. Thanks for reading! ===================================================================== For readers who might be thinking what set this type of expectation? Because I had a dev shop and I thought my co-founders will be understanding of my business circumstance and I was a bit trigger to build a product with a C-exec team, I gave the impression of "unlimited" engineering which I later realized down the line that it was not feasible for me. Something I learned that I have to be more careful with and set expectations accordingly from the very beginning. And from the feedback of the commenters here, I am much more aware what I should offer and how to set expectations, esp. in the early stages of execution. So thank you all! 🙏🏾 EDIT: I would like to thank everyone who contributed to this thread. You not only helped me but future founders who are considering to get into the startup scene!

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
GummyBear8659This week

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)

Hey everyone! Long-time lurker, first-time founder here. I’m reaching out to get feedback on a recent startup experience—what went wrong, what I could have done better, and how I should approach future opportunities. The Background There were three founders in this venture: • Founder A (CEO, 50%) – The product/growth guy who identified the problem space. • Founder B (Me, CTO, 37.5%) – A software engineer with a software dev shop and multiple clients. I wanted to diversify into building my own products but am not inherently a “product person.” • Founder C (COO, 12.5%) – Brought into the mix by Founder A, with the goal of leveraging his network for traction once the product was built. The idea was to create Product X, a solution targeting the SMB space while competitors were moving upmarket. It wasn’t revolutionary—more of a strategic market play. The Initial Plan & My Role • Founder A would define and prioritize product specs, guiding what needed to be built. • I (Founder B) didn’t have time to code myself, so I allocated engineers from my dev shop (which I personally paid for). My stake was adjusted from 32.5% to 37.5% to reflect this contribution. • Founder C was more of an observer early on, planning to help with traction once we had a product ready. We agreed on a 1-year cliff and a 4-year vesting schedule for equity. Where Things Started to Go Wrong • Lack of a Clear Product Roadmap – Founder A was very focused on getting something built fast, but we never signed off on a structured roadmap or milestones. I underestimated the complexity of what was actually needed for customer conversations. • Engineering Expectations vs. Reality – The team (one part-time lead + two full-time juniors from my dev shop) faced early feedback that development was too slow. In response, I ramped up the lead to full-time and added a part-time PM. But Founder A continued pushing for speed, despite real hurdles (OAuth integrations, etc.). • Shifting MVP Goalposts – Midway, Founder A concluded that an MVP wouldn’t cut it—we needed a more complete product to be competitive. This meant more engineering, more delays, and more of my own money spent on development. The Breaking Point Near the 1-year vesting mark, we had an opportunity: a paying client willing to fund an app. I didn’t have devs on the bench, so I asked Founder A to hold off our project briefly while I hired more engineers to avoid stalling either effort. This was the final straw. Founder A (with Founder C somewhat aligned) decided the arrangement wasn’t working—citing past disagreements and the “slowness” issue. The decision was made to end the partnership. Now, Founder A, as majority holder, is requesting a full handover of the code, Founder C is indifferent, and all engineering costs I covered are essentially lost. Key Takeaways (So Far) Crystal-Clear Agreements Upfront – A formalized product roadmap and timeline should’ve been locked in from day one. Business Needs > Engineering Standards – I wanted to build something solid and scalable, but in an early-stage startup, speed to market is king. This was before AI tools became mainstream, so our approach wasn’t as optimized. Don’t Overextend Without Protection – I personally financed all engineering, but without clear safeguards, that investment became a sunk cost. Expenses Must Be Distributed – I was solely covering engineering salaries, which created an imbalance in financial risk. Future partnerships should ensure costs are shared proportionally, rather than one person shouldering the burden. Where I Need Advice Looking back, I want to improve as an engineer, CEO, and co-founder. • What should I have done differently in structuring this partnership? • How do you balance engineering quality with the startup need for speed? • As a dev shop owner, how can I better navigate equity deals where I’m also bringing in engineering resources? I really appreciate everyone who went through this long post and provide any insights from founders, engineers, or anyone who has been in a similar situation. Thanks for reading! ===================================================================== For readers who might be thinking what set this type of expectation? Because I had a dev shop and I thought my co-founders will be understanding of my business circumstance and I was a bit trigger to build a product with a C-exec team, I gave the impression of "unlimited" engineering which I later realized down the line that it was not feasible for me. Something I learned that I have to be more careful with and set expectations accordingly from the very beginning. And from the feedback of the commenters here, I am much more aware what I should offer and how to set expectations, esp. in the early stages of execution. So thank you all! 🙏🏾 EDIT: I would like to thank everyone who contributed to this thread. You not only helped me but future founders who are considering to get into the startup scene!

I studied how 7 Founders found their first 100 customers for their businesses. Summarizing it here!
reddit
LLM Vibe Score0
Human Vibe Score1
adriannelestrangeThis week

I studied how 7 Founders found their first 100 customers for their businesses. Summarizing it here!

I am learning marketing, and so I combed through the internet to find specific advice that helped founders reach 100 users and not random Google answers. Here’s what I found: Llama Life by Marie Marie founder of Llama Life, a productivity app ($51.4K+ revenue) got her first 100 users using Snowballing effect. She shared great advice that I want to add here verbatim, “Need to think about what you have that you can leverage based on your current situation. eg..When you have no customers, think about where you can post to get the 1st customer eg Product Hunt. If you do well on PH, say you get #3 product of the day, then you post somewhere else saying ‘I got #3 product of the day’.. to get your next few customers. Maybe that post is on reddit with some learnings that you found. If the reddit post does well, then you might post it on Twitter, saying reddit did well and what learnings you got from that etc. or even if it doesn’t do well you can still post about it.” Another tip she shared is to build related products that get more viral than the product itself. These are small stand-alone sites that would appeal to the same target audience, but by nature, are more shareable. On these sites, you can mention your startup like: ‘brought to you by Llama Life’ and then provide a link to the main website if someone is interested. If one of those gets viral or ranks on Google, you’ll have a passive traffic source. Scraping bee by Pierre Pierre, founder of Scraping Bee, a web scraping tool has now reached $1.5M ARR. Pierre and his cofounder Kevin started with 10 Free Beta Users in 2019, and after 6 months asked them to take a paid subscription if they wanted to continue using the product. That’s how they got their first user within 50 minutes of that email. Then they listed it on dozens of startup directories but their core strategy was writing the best possible content for their target audience — Developers. 3 very successful pieces of content that worked were : A small tutorial on how to scrape single-page application An extensive general guide about web scraping without getting blocked A complete introduction to web scraping with Python They didn’t do content marketing for the sake of content marketing but deep-dived into the value they were providing their customer. One of these got 70K visits, and all this together got them to over 100 users. WePay by Bill Clerico Bill Clerico left his cushy corporate job to build WePay which was then acquired for $400M got his first users by using his app. He got his first users by using his app! The app was for group payments. So he hosted a Poker tournament at his house and collected payments only with his app. Then they hosted a barbecue for fraternity treasurers at San Jose State & helped them do their annual dues collection. Good old word-of-mouth marketing, that however, started with an event where they used what they made! RealWorld by Genevieve Genevieve — Founder and CEO of Realworld stands by the old-school advice of value giving. RealWorld is an app that helps GenZ navigate adulthood. So, before launching their direct-to-consumer platform, they had an educational course that they sold to college career centers and students. They already had a pipeline of adults who turned to Realworld for their adulting challenges. From there, she gained her first 100 followers. Saner dot ai by Austin Austin got 100 users from Reddit for his startup Saner.ai. Reddit hates advertising, and so his tips to market your startup on Reddit is to Write value-driven posts on your niche. Instead of writing posts, find posts where people are looking for solutions DM people facing problems that your SaaS solves. But instead of selling, ask about their problem to see if your product is a good fit Heartfelt posts about why you built it, aren’t gonna cut it To find posts and people, search Reddit with relevant keywords and join all the subreddits A Stock Portfolio Newsletter A financial investor got his first 100 paid newsletter subscribers for his stock portfolio newsletter. His tips : Don’t reinvent the wheel. Work what’s already working. He saw a company making $500M+ from stock picking newsletter, so decided to try that. Find the gaps in “already working” and leverage them. That newsletter did not have portfolios of advisors writing them. That was his USP. He added his own portfolio to his newsletter. Now to 100 users, he partnered with a guy running an investing website and getting good traffic. That guy got a cut of his revenue, in exchange. That one simple step got him to 100 users. Hypefury by Yannick and Samy Yannick and Samy from Hypefury, Twitter and Social Media Automation tool got their first beta testers and users from a paid community. They launched Hypefury there and asked if someone wanted to try it. A couple of people tried it and gave feedback. Samy conducted user interviews and product demos for them, And shared the reviews on Twitter. That alone, along with word-of-mouth marketing on Twitter got them their first 100 users. To conclude: Don’t reinvent the wheel, try what’s working. Find the gaps in what’s working, and leverage that. Instead of thinking about millions of customers, think about the first 10. Then first 100. Leverage what you have. Get the first 10 customers, then talk about this to get the next 100. Use your app. Find ways, events, and opportunities to use your app in front of people. And get them to use it. Write content not only for SEO but also to help people. It won’t work tomorrow, but it will work for years after it picks up. Leverage other sources of traffic by partnering up! Do things that don’t scale. I’m also doing SaaS marketing deep dives over 30 pieces of content. I'm posting here for the first time, so I'm not sure if it will stay or not, sorry if it doesn't. I've helped a SaaS grow from $19K to $100K MRR as a marketer in last 2 years, and now I wanna dive deep. Cheers! (1/30)

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age
reddit
LLM Vibe Score0
Human Vibe Score1
ExpenectThis week

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age Introduction In an era characterized by rapid technological advancements, the field of finance is undergoing a transformative journey. The emergence of financial technology, or fintech, is reshaping the way businesses manage their finances, and Chartered Accountants (CAs) are at the forefront of this evolution. In this blog post, we'll explore how CAs are embracing fintech and leveraging its potential to enhance financial management, analysis, and advisory services. Fintech's Impact on Financial Services Fintech encompasses a wide range of technologies that leverage data analytics, artificial intelligence, blockchain, and automation to improve financial services. For CAs, this means new tools to streamline processes, enhance decision-making, and offer innovative solutions to clients. Automation of Routine Tasks CAs are increasingly using automation tools to handle repetitive tasks such as data entry, reconciliations, and transaction processing. This not only reduces the risk of human error but also frees up CAs to focus on higher-value tasks like strategic planning and analysis. Advanced Data Analytics Data analytics tools enable CAs to extract meaningful insights from large volumes of financial data. These insights can help businesses identify trends, anticipate risks, and make informed decisions to drive growth. Real-Time Financial Reporting Fintech enables CAs to provide clients with real-time financial reporting, giving businesses immediate access to critical information. This enhances transparency and empowers business owners to respond quickly to changing market conditions. Enhancing Audit Efficiency Fintech tools are revolutionizing the audit process. CAs can use AI-powered algorithms to analyze vast amounts of data, detect anomalies, and identify potential instances of fraud more efficiently. Personalized Financial Planning CAs can leverage fintech to offer personalized financial planning services. With access to detailed financial data, CAs can create tailored strategies that align with a client's unique goals and circumstances. Strengthening Cybersecurity As businesses become more reliant on digital tools, cybersecurity becomes paramount. CAs are playing a critical role in advising clients on cybersecurity measures to protect sensitive financial information. Virtual CFO Services Fintech enables CAs to offer virtual CFO services to startups and small businesses. Through digital platforms, CAs can provide expert financial advice and guidance remotely, making their expertise accessible to a wider range of clients. Embracing Blockchain Technology Blockchain's potential for secure and transparent record-keeping is of interest to CAs. They can explore applications in supply chain finance, smart contracts, and even audit trail verification. Continuous Learning in Fintech CAs recognize the importance of staying updated with fintech trends. Many are investing in continuous learning to master the use of new tools and technologies that can optimize their services. Conclusion The integration of fintech into the realm of finance is reshaping the landscape in profound ways. CAs are embracing these technologies to elevate their roles from traditional number-crunchers to strategic advisors, equipped with tools that enhance efficiency, accuracy, and insight. As fintech continues to evolve, CAs will remain pivotal in guiding businesses through the ever-changing financial landscape, leveraging technology to drive growth, innovation, and success. Find the top verified CA in your City Feel free to let me know if you'd like more blogs on different topics or if you have specific requirements for the content.

Here’s How Chatbots Can Boost Your Small Business
reddit
LLM Vibe Score0
Human Vibe Score1
smanwerThis week

Here’s How Chatbots Can Boost Your Small Business

Chatbots are the next big thing in the tech world that are meant for business use. Almost every business can benefit from chatbots in one way or the other. They are now everywhere – the fastest rising star are basically computer-operated machines that can play a variety of roles such as customer service representative, social media manager, personal assistant and much more. Virtually every industry is seemingly investing in it. Chatbots became the flavor of the season because of their task management and problem solving skills. This is why companies are aggressively deploying chatbots to their business strategy to make it work right. What are Chatbots – How They Can Benefit Your Small Business? In essence, chatbots are simply a computer program tailor-made to mimic conversations with the help of artificial intelligence (AI). These computer-based programs are capable enough to respond to natural language text and voice inputs in a human way. Chatbots can take over a lot of time consuming tasks, allowing project managers to focus on other important matters and take high level decisions. Chatbots are not just the next big thing for digital and tech brands, small businesses can also get the most out from them. Small businesses should get into chatbots to streamline their routine project management practices and support other business operations – thereby saving budget, time, energy, while improving ROI. If you are not completely getting into it, here are some ways that help you deploy this rising technology in order to boost your small business strategy. Instant Customer Support One of the effective ways small businesses can implement a chatbot is an immediate customer support. If you belong to an industry that offers products and services, chances are you get so many phone calls and emails to educate people. Prior to allowing customers to clog up your inbox with unlimited queries, try using a chatbot that will save your valuable time. You can simply create an immediate customer support presence for customers who engage with your chatbot. Craft answers for all the popular queries so that your project management team can focus on other complex and important issues while chatbots addressing the most commonly asked questions. Moreover, it will add a consistency to your brand voice. You can control the tone and ensure that the chatbot will deliver your crafted messages. Boost Sales Leads Generation Chatbots are not just about sharing or collecting information. They can actually boost sales. But, how? Though they can’t replace your sales and marketing team, they can smartly assist them by being an immediate point of contact. Create an automated conversation for a new visitor and it can directly influence sales. Though chatbots are rising, they will ultimately carry on artificial intelligence that is capable for gathering the data required to curate a specific set of products for customers. For instance, if a user asks the chatbot for blue shirt in cotton, the chatbot can pull items with the particular details for the user. This process is cumulative and when next time the user communicates with the chatbot, it will consider their preferences. Increase Your Business Efficiency Though chatbots can’t perform every business operation, what they can do is eliminate few of the menial but important operations. Consider all the important tasks that your employees need to perform, such as answering customer queries, compiling data for a user, filling out form etc. Most of these tasks are monotonous in nature that allows you to train your chatbot to manage all these repetitive tasks with a low risk and high return of your valuable time. Reducing Cost and Resource Consumption Like any online task management system , chatbots are great to reduce manpower. From performing as a personal assistant to a customer sales representative, you can easily cut down the total number of resources that deal with customer complaints and feedback. You can utilize a chatbot, as it can do this work easily a human would usually do. Read Full article here

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age
reddit
LLM Vibe Score0
Human Vibe Score1
ExpenectThis week

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age Introduction In an era characterized by rapid technological advancements, the field of finance is undergoing a transformative journey. The emergence of financial technology, or fintech, is reshaping the way businesses manage their finances, and Chartered Accountants (CAs) are at the forefront of this evolution. In this blog post, we'll explore how CAs are embracing fintech and leveraging its potential to enhance financial management, analysis, and advisory services. Fintech's Impact on Financial Services Fintech encompasses a wide range of technologies that leverage data analytics, artificial intelligence, blockchain, and automation to improve financial services. For CAs, this means new tools to streamline processes, enhance decision-making, and offer innovative solutions to clients. Automation of Routine Tasks CAs are increasingly using automation tools to handle repetitive tasks such as data entry, reconciliations, and transaction processing. This not only reduces the risk of human error but also frees up CAs to focus on higher-value tasks like strategic planning and analysis. Advanced Data Analytics Data analytics tools enable CAs to extract meaningful insights from large volumes of financial data. These insights can help businesses identify trends, anticipate risks, and make informed decisions to drive growth. Real-Time Financial Reporting Fintech enables CAs to provide clients with real-time financial reporting, giving businesses immediate access to critical information. This enhances transparency and empowers business owners to respond quickly to changing market conditions. Enhancing Audit Efficiency Fintech tools are revolutionizing the audit process. CAs can use AI-powered algorithms to analyze vast amounts of data, detect anomalies, and identify potential instances of fraud more efficiently. Personalized Financial Planning CAs can leverage fintech to offer personalized financial planning services. With access to detailed financial data, CAs can create tailored strategies that align with a client's unique goals and circumstances. Strengthening Cybersecurity As businesses become more reliant on digital tools, cybersecurity becomes paramount. CAs are playing a critical role in advising clients on cybersecurity measures to protect sensitive financial information. Virtual CFO Services Fintech enables CAs to offer virtual CFO services to startups and small businesses. Through digital platforms, CAs can provide expert financial advice and guidance remotely, making their expertise accessible to a wider range of clients. Embracing Blockchain Technology Blockchain's potential for secure and transparent record-keeping is of interest to CAs. They can explore applications in supply chain finance, smart contracts, and even audit trail verification. Continuous Learning in Fintech CAs recognize the importance of staying updated with fintech trends. Many are investing in continuous learning to master the use of new tools and technologies that can optimize their services. Conclusion The integration of fintech into the realm of finance is reshaping the landscape in profound ways. CAs are embracing these technologies to elevate their roles from traditional number-crunchers to strategic advisors, equipped with tools that enhance efficiency, accuracy, and insight. As fintech continues to evolve, CAs will remain pivotal in guiding businesses through the ever-changing financial landscape, leveraging technology to drive growth, innovation, and success. Find the top verified CA in your City Feel free to let me know if you'd like more blogs on different topics or if you have specific requirements for the content.

How I Started Learning Machine Learning
reddit
LLM Vibe Score0
Human Vibe Score1
TechPrimoThis week

How I Started Learning Machine Learning

Hello, everyone. As promised, I'll write a longer post about how I entered the world of ML, hoping it will help someone shape their path. I'll include links to all the useful materials I used alongside the story, which you can use for learning. I like to call myself an AI Research Scientist who enjoys exploring new AI trends, delving deeper into understanding their background, and applying them to real products. This way, I try to connect science and entrepreneurship because I believe everything that starts as scientific research ends up "on the shelves" as a product that solves a specific user problem. I began my journey in ML in 2016 when it wasn't such a popular field. Everyone had heard of it, but few were applying it. I have several years of development experience and want to try my hand at ML. The first problem I encountered was where to start - whether to learn mathematics, statistics, or something else. That's when I came across a name and a course that completely changed my career. Let's start You guessed it. It was Professor Andrew Ng and his globally popular Machine Learning course available on Coursera (I still have the certificate, hehe). This was also my first official online course ever. Since that course no longer exists as it's been replaced by a new one, I recommend you check out: Machine Learning (Stanford CS229) Machine Learning Specialization These two courses start from the basics of ML and all the necessary calculus you need to know. Many always ask questions like whether to learn linear algebra, statistics, or probability, but you don't need to know everything in depth. This knowledge helps if you're a scientist developing a new architecture, but as an engineer, not really. You need to know some basics to understand, such as how the backpropagation algorithm works. I know that Machine Learning (Stanford CS229) is a very long and arduous course, but it's the right start if you want to be really good at ML. In my time, I filled two thick notebooks by hand while taking the course mentioned above. TensorFlow and Keras After the course, I didn't know how to apply my knowledge because I hadn't learned specifically how to code things. Then, I was looking for ways to learn how to code it. That's when I came across a popular framework called Keras, now part of TensorFlow. I started with a new course and acquiring practical knowledge: Deep Learning Specialization Deep Learning by Ian Goodfellow Machine Learning Yearning by Andrew Ng These resources above were my next step. I must admit that I learned the most from that course and from the book Deep Learning by Ian Goodfellow because I like reading books (although this one is quite difficult to read). Learn by coding To avoid just learning, I went through various GitHub repositories that I manually retyped and learned that way. It may be an old-fashioned technique, but it helped me a lot. Now, most of those repositories don't exist, so I'll share some that I found to be good: Really good Jupyter notebooks that can teach you the basics of TensorFlow Another good repo for learning TF and Keras Master the challenge After mastering the basics in terms of programming in TF/Keras, I wanted to try solving some real problems. There's no better place for that challenge than Kaggle and the popular Titanic dataset. Here, you can really find a bunch of materials and simple examples of ML applications. Here are some of my favorites: Titanic - Machine Learning from Disaster Home Credit Default Risk House Prices - Advanced Regression Techniques Two Sigma: Using News to Predict Stock Movements I then decided to further develop my career in the direction of applying ML to the stock market, first using predictions on time series and then using natural language processing. I've remained in this field until today and will defend my doctoral dissertation soon. How to deploy models To continue, before I move on to the topic of specialization, we need to address the topic of deployment. Now that we've learned how to make some basic models in Keras and how to use them, there are many ways and services, but I'll only mention what I use today. For all my ML models, whether simple regression models or complex GPT models, I use FastAPI. It's a straightforward framework, and you can quickly create API endpoints. I'll share a few older and useful tutorials for beginners: AI as an API tutorial series A step-by-step guide Productizing an ML Model with FastAPI and Cloud Run Personally, I've deployed on various cloud providers, of which I would highlight GCP and AWS because they have everything needed for model deployment, and if you know how to use them, they can be quite cheap. Chose your specialization The next step in developing my career, besides choosing finance as the primary area, was my specialization in the field of NLP. This happened in early 2020 when I started working with models based on the Transformer architecture. The first model I worked with was BERT, and the first tasks were related to classifications. My recommendations are to master the Transformer architecture well because 99% of today's LLM models are based on it. Here are some resources: The legendary paper "Attention Is All You Need" Hugging Face Course on Transformers Illustrated Guide to Transformers - Step by Step Explanation Good repository How large language models work, a visual intro to transformers After spending years using encoder-based Transformer models, I started learning GPT models. Good open-source models like Llama 2 then appear. Then, I started fine-tuning these models using the excellent Unsloth library: How to Finetune Llama-3 and Export to Ollama Fine-tune Llama 3.1 Ultra-Efficiently with Unsloth After that, I focused on studying various RAG techniques and developing Agent AI systems. This is now called AI engineering, and, as far as I can see, it has become quite popular. So I'll write more about that in another post, but here I'll leave what I consider to be the three most famous representatives, i.e., their tutorials: LangChain tutorial LangGraph tutorial CrewAI examples Here I am today Thanks to the knowledge I've generated over all these years in the field of ML, I've developed and worked on numerous projects. The most significant publicly available project is developing an agent AI system for well-being support, which I turned into a mobile application. Also, my entire doctoral dissertation is related to applying ML to the stock market in combination with the development of GPT models and reinforcement learning (more on that in a separate post). After long 6 years, I've completed my dissertation, and now I'm just waiting for its defense. I'll share everything I'm working on for the dissertation publicly on the project, and in tutorials I'm preparing to write. If you're interested in these topics, I announce that I'll soon start with activities of publishing content on Medium and a blog, but I'll share all of that here on Reddit as well. Now that I've gathered years of experience and knowledge in this field, I'd like to share it with others and help as much as possible. If you have any questions, feel free to ask them, and I'll try to answer all of them. Thank you for reading.

How I Started Learning Machine Learning
reddit
LLM Vibe Score0
Human Vibe Score1
TechPrimoThis week

How I Started Learning Machine Learning

Hello, everyone. As promised, I'll write a longer post about how I entered the world of ML, hoping it will help someone shape their path. I'll include links to all the useful materials I used alongside the story, which you can use for learning. I like to call myself an AI Research Scientist who enjoys exploring new AI trends, delving deeper into understanding their background, and applying them to real products. This way, I try to connect science and entrepreneurship because I believe everything that starts as scientific research ends up "on the shelves" as a product that solves a specific user problem. I began my journey in ML in 2016 when it wasn't such a popular field. Everyone had heard of it, but few were applying it. I have several years of development experience and want to try my hand at ML. The first problem I encountered was where to start - whether to learn mathematics, statistics, or something else. That's when I came across a name and a course that completely changed my career. Let's start You guessed it. It was Professor Andrew Ng and his globally popular Machine Learning course available on Coursera (I still have the certificate, hehe). This was also my first official online course ever. Since that course no longer exists as it's been replaced by a new one, I recommend you check out: Machine Learning (Stanford CS229) Machine Learning Specialization These two courses start from the basics of ML and all the necessary calculus you need to know. Many always ask questions like whether to learn linear algebra, statistics, or probability, but you don't need to know everything in depth. This knowledge helps if you're a scientist developing a new architecture, but as an engineer, not really. You need to know some basics to understand, such as how the backpropagation algorithm works. I know that Machine Learning (Stanford CS229) is a very long and arduous course, but it's the right start if you want to be really good at ML. In my time, I filled two thick notebooks by hand while taking the course mentioned above. TensorFlow and Keras After the course, I didn't know how to apply my knowledge because I hadn't learned specifically how to code things. Then, I was looking for ways to learn how to code it. That's when I came across a popular framework called Keras, now part of TensorFlow. I started with a new course and acquiring practical knowledge: Deep Learning Specialization Deep Learning by Ian Goodfellow Machine Learning Yearning by Andrew Ng These resources above were my next step. I must admit that I learned the most from that course and from the book Deep Learning by Ian Goodfellow because I like reading books (although this one is quite difficult to read). Learn by coding To avoid just learning, I went through various GitHub repositories that I manually retyped and learned that way. It may be an old-fashioned technique, but it helped me a lot. Now, most of those repositories don't exist, so I'll share some that I found to be good: Really good Jupyter notebooks that can teach you the basics of TensorFlow Another good repo for learning TF and Keras Master the challenge After mastering the basics in terms of programming in TF/Keras, I wanted to try solving some real problems. There's no better place for that challenge than Kaggle and the popular Titanic dataset. Here, you can really find a bunch of materials and simple examples of ML applications. Here are some of my favorites: Titanic - Machine Learning from Disaster Home Credit Default Risk House Prices - Advanced Regression Techniques Two Sigma: Using News to Predict Stock Movements I then decided to further develop my career in the direction of applying ML to the stock market, first using predictions on time series and then using natural language processing. I've remained in this field until today and will defend my doctoral dissertation soon. How to deploy models To continue, before I move on to the topic of specialization, we need to address the topic of deployment. Now that we've learned how to make some basic models in Keras and how to use them, there are many ways and services, but I'll only mention what I use today. For all my ML models, whether simple regression models or complex GPT models, I use FastAPI. It's a straightforward framework, and you can quickly create API endpoints. I'll share a few older and useful tutorials for beginners: AI as an API tutorial series A step-by-step guide Productizing an ML Model with FastAPI and Cloud Run Personally, I've deployed on various cloud providers, of which I would highlight GCP and AWS because they have everything needed for model deployment, and if you know how to use them, they can be quite cheap. Chose your specialization The next step in developing my career, besides choosing finance as the primary area, was my specialization in the field of NLP. This happened in early 2020 when I started working with models based on the Transformer architecture. The first model I worked with was BERT, and the first tasks were related to classifications. My recommendations are to master the Transformer architecture well because 99% of today's LLM models are based on it. Here are some resources: The legendary paper "Attention Is All You Need" Hugging Face Course on Transformers Illustrated Guide to Transformers - Step by Step Explanation Good repository How large language models work, a visual intro to transformers After spending years using encoder-based Transformer models, I started learning GPT models. Good open-source models like Llama 2 then appear. Then, I started fine-tuning these models using the excellent Unsloth library: How to Finetune Llama-3 and Export to Ollama Fine-tune Llama 3.1 Ultra-Efficiently with Unsloth After that, I focused on studying various RAG techniques and developing Agent AI systems. This is now called AI engineering, and, as far as I can see, it has become quite popular. So I'll write more about that in another post, but here I'll leave what I consider to be the three most famous representatives, i.e., their tutorials: LangChain tutorial LangGraph tutorial CrewAI examples Here I am today Thanks to the knowledge I've generated over all these years in the field of ML, I've developed and worked on numerous projects. The most significant publicly available project is developing an agent AI system for well-being support, which I turned into a mobile application. Also, my entire doctoral dissertation is related to applying ML to the stock market in combination with the development of GPT models and reinforcement learning (more on that in a separate post). After long 6 years, I've completed my dissertation, and now I'm just waiting for its defense. I'll share everything I'm working on for the dissertation publicly on the project, and in tutorials I'm preparing to write. If you're interested in these topics, I announce that I'll soon start with activities of publishing content on Medium and a blog, but I'll share all of that here on Reddit as well. Now that I've gathered years of experience and knowledge in this field, I'd like to share it with others and help as much as possible. If you have any questions, feel free to ask them, and I'll try to answer all of them. Thank you for reading.

MMML | Deploy HuggingFace training model rapidly based on MetaSpore
reddit
LLM Vibe Score0
Human Vibe Score1
qazmkoppThis week

MMML | Deploy HuggingFace training model rapidly based on MetaSpore

A few days ago, HuggingFace announced a $100 million Series C funding round, which was big news in open source machine learning and could be a sign of where the industry is headed. Two days before the HuggingFace funding announcement, open-source machine learning platform MetaSpore released a demo based on the HuggingFace Rapid deployment pre-training model. As deep learning technology makes innovative breakthroughs in computer vision, natural language processing, speech understanding, and other fields, more and more unstructured data are perceived, understood, and processed by machines. These advances are mainly due to the powerful learning ability of deep learning. Through pre-training of deep models on massive data, the models can capture the internal data patterns, thus helping many downstream tasks. With the industry and academia investing more and more energy in the research of pre-training technology, the distribution warehouses of pre-training models such as HuggingFace and Timm have emerged one after another. The open-source community release pre-training significant model dividends at an unprecedented speed. In recent years, the data form of machine modeling and understanding has gradually evolved from single-mode to multi-mode, and the semantic gap between different modes is being eliminated, making it possible to retrieve data across modes. Take CLIP, OpenAI’s open-source work, as an example, to pre-train the twin towers of images and texts on a dataset of 400 million pictures and texts and connect the semantics between pictures and texts. Many researchers in the academic world have been solving multimodal problems such as image generation and retrieval based on this technology. Although the frontier technology through the semantic gap between modal data, there is still a heavy and complicated model tuning, offline data processing, high performance online reasoning architecture design, heterogeneous computing, and online algorithm be born multiple processes and challenges, hindering the frontier multimodal retrieval technologies fall to the ground and pratt &whitney. DMetaSoul aims at the above technical pain points, abstracting and uniting many links such as model training optimization, online reasoning, and algorithm experiment, forming a set of solutions that can quickly apply offline pre-training model to online. This paper will introduce how to use the HuggingFace community pre-training model to conduct online reasoning and algorithm experiments based on MetaSpore technology ecology so that the benefits of the pre-training model can be fully released to the specific business or industry and small and medium-sized enterprises. And we will give the text search text and text search graph two multimodal retrieval demonstration examples for your reference. Multimodal semantic retrieval The sample architecture of multimodal retrieval is as follows: Our multimodal retrieval system supports both text search and text search application scenarios, including offline processing, model reasoning, online services, and other core modules: https://preview.redd.it/mdyyv1qmdz291.png?width=1834&format=png&auto=webp&s=e9e10710794c78c64cc05adb75db385aa53aba40 Offline processing, including offline data processing processes for different application scenarios of text search and text search, including model tuning, model export, data index database construction, data push, etc. Model inference. After the offline model training, we deployed our NLP and CV large models based on the MetaSpore Serving framework. MetaSpore Serving helps us conveniently perform online inference, elastic scheduling, load balancing, and resource scheduling in heterogeneous environments. Online services. Based on MetaSpore’s online algorithm application framework, MetaSpore has a complete set of reusable online search services, including Front-end retrieval UI, multimodal data preprocessing, vector recall and sorting algorithm, AB experimental framework, etc. MetaSpore also supports text search by text and image scene search by text and can be migrated to other application scenarios at a low cost. The HuggingFace open source community has provided several excellent baseline models for similar multimodal retrieval problems, which are often the starting point for actual optimization in the industry. MetaSpore also uses the pre-training model of the HuggingFace community in its online services of searching words by words and images by words. Searching words by words is based on the semantic similarity model of the question and answer field optimized by MetaSpore, and searching images by words is based on the community pre-training model. These community open source pre-training models are exported to the general ONNX format and loaded into MetaSpore Serving for online reasoning. The following sections will provide a detailed description of the model export and online retrieval algorithm services. The reasoning part of the model is standardized SAAS services with low coupling with the business. Interested readers can refer to my previous post: The design concept of MetaSpore, a new generation of the one-stop machine learning platform. 1.1 Offline Processing Offline processing mainly involves the export and loading of online models and index building and pushing of the document library. You can follow the step-by-step instructions below to complete the offline processing of text search and image search and see how the offline pre-training model achieves reasoning at MetaSpore. 1.1.1 Search text by text Traditional text retrieval systems are based on literal matching algorithms such as BM25. Due to users’ diverse query words, a semantic gap between query words and documents is often encountered. For example, users misspell “iPhone” as “Phone,” and search terms are incredibly long, such as “1 \~ 3 months old baby autumn small size bag pants”. Traditional text retrieval systems will use spelling correction, synonym expansion, search terms rewriting, and other means to alleviate the semantic gap but fundamentally fail to solve this problem. Only when the retrieval system fully understands users’ query terms and documents can it meet users’ retrieval demands at the semantic level. With the continuous progress of pre-training and representational learning technology, some commercial search engines continue to integrate semantic vector retrieval methods based on symbolic learning into the retrieval ecology. Semantic retrieval model This paper introduces a set of semantic vector retrieval applications. MetaSpore built a set of semantic retrieval systems based on encyclopedia question and answer data. MetaSpore adopted the Sentence-Bert model as the semantic vector representation model, which fine-tunes the twin tower BERT in supervised or unsupervised ways to make the model more suitable for retrieval tasks. The model structure is as follows: The query-Doc symmetric two-tower model is used in text search and question and answer retrieval. The vector representation of online Query and offline DOC share the same vector representation model, so it is necessary to ensure the consistency of the offline DOC library building model and online Query inference model. The case uses MetaSpore’s text representation model Sbert-Chinese-QMC-domain-V1, optimized in the open-source semantically similar data set. This model will express the question and answer data as a vector in offline database construction. The user query will be expressed as a vector by this model in online retrieval, ensuring that query-doc in the same semantic space, users’ semantic retrieval demands can be guaranteed by vector similarity metric calculation. Since the text presentation model does vector encoding for Query online, we need to export the model for use by the online service. Go to the q&A data library code directory and export the model concerning the documentation. In the script, Pytorch Tracing is used to export the model. The models are exported to the “./export “directory. The exported models are mainly ONNX models used for wired reasoning, Tokenizer, and related configuration files. The exported models are loaded into MetaSpore Serving by the online Serving system described below for model reasoning. Since the exported model will be copied to the cloud storage, you need to configure related variables in env.sh. \Build library based on text search \ The retrieval database is built on the million-level encyclopedia question and answer data set. According to the description document, you need to download the data and complete the database construction. The question and answer data will be coded as a vector by the offline model, and then the database construction data will be pushed to the service component. The whole process of database construction is described as follows: Preprocessing, converting the original data into a more general JSonline format for database construction; Build index, use the same model as online “sbert-Chinese-qmc-domain-v1” to index documents (one document object per line); Push inverted (vector) and forward (document field) data to each component server. The following is an example of the database data format. After offline database construction is completed, various data are pushed to corresponding service components, such as Milvus storing vector representation of documents and MongoDB storing summary information of documents. Online retrieval algorithm services will use these service components to obtain relevant data. 1.1.2 Search by text Text and images are easy for humans to relate semantically but difficult for machines. First of all, from the perspective of data form, the text is the discrete ID type of one-dimensional data based on words and words. At the same time, images are continuous two-dimensional or three-dimensional data. Secondly, the text is a subjective creation of human beings, and its expressive ability is vibrant, including various turning points, metaphors, and other expressions, while images are machine representations of the objective world. In short, bridging the semantic gap between text and image data is much more complex than searching text by text. The traditional text search image retrieval technology generally relies on the external text description data of the image or the nearest neighbor retrieval technology and carries out the retrieval through the image associated text, which in essence degrades the problem to text search. However, it will also face many issues, such as obtaining the associated text of pictures and whether the accuracy of text search by text is high enough. The depth model has gradually evolved from single-mode to multi-mode in recent years. Taking the open-source project of OpenAI, CLIP, as an example, train the model through the massive image and text data of the Internet and map the text and image data into the same semantic space, making it possible to implement the text and image search technology based on semantic vector. CLIP graphic model The text search pictures introduced in this paper are implemented based on semantic vector retrieval, and the CLIP pre-training model is used as the two-tower retrieval architecture. Because the CLIP model has trained the semantic alignment of the twin towers’ text and image side models on the massive graphic and text data, it is particularly suitable for the text search graph scene. Due to the different image and text data forms, the Query-Doc asymmetric twin towers model is used for text search image retrieval. The image-side model of the twin towers is used for offline database construction, and the text-side model is used for the online return. In the final online retrieval, the database data of the image side model will be searched after the text side model encodes Query, and the CLIP pre-training model guarantees the semantic correlation between images and texts. The model can draw the graphic pairs closer in vector space by pre-training on a large amount of visual data. Here we need to export the text-side model for online MetaSpore Serving inference. Since the retrieval scene is based on Chinese, the CLIP model supporting Chinese understanding is selected. The exported content includes the ONNX model used for online reasoning and Tokenizer, similar to the text search. MetaSpore Serving can load model reasoning through the exported content. Build library on Image search You need to download the Unsplash Lite library data and complete the construction according to the instructions. The whole process of database construction is described as follows: Preprocessing, specify the image directory, and then generate a more general JSOnline file for library construction; Build index, use OpenAI/Clip-Vit-BASE-Patch32 pre-training model to index the gallery, and output one document object for each line of index data; Push inverted (vector) and forward (document field) data to each component server. Like text search, after offline database construction, relevant data will be pushed to service components, called by online retrieval algorithm services to obtain relevant data. 1.2 Online Services The overall online service architecture diagram is as follows: ​ https://preview.redd.it/nz8zrbbpdz291.png?width=1280&format=png&auto=webp&s=28dae7e031621bc8819519667ed03d8d085d8ace Multi-mode search online service system supports application scenarios such as text search and text search. The whole online service consists of the following parts: Query preprocessing service: encapsulate preprocessing logic (including text/image, etc.) of pre-training model, and provide services through gRPC interface; Retrieval algorithm service: the whole algorithm processing link includes AB experiment tangent flow configuration, MetaSpore Serving call, vector recall, sorting, document summary, etc.; User entry service: provides a Web UI interface for users to debug and track down problems in the retrieval service. From a user request perspective, these services form invocation dependencies from back to front, so to build up a multimodal sample, you need to run each service from front to back first. Before doing this, remember to export the offline model, put it online and build the library first. This article will introduce the various parts of the online service system and make the whole service system step by step according to the following guidance. See the ReadME at the end of this article for more details. 1.2.1 Query preprocessing service Deep learning models tend to be based on tensors, but NLP/CV models often have a preprocessing part that translates raw text and images into tensors that deep learning models can accept. For example, NLP class models often have a pre-tokenizer to transform text data of string type into discrete tensor data. CV class models also have similar processing logic to complete the cropping, scaling, transformation, and other processing of input images through preprocessing. On the one hand, considering that this part of preprocessing logic is decoupled from tensor reasoning of the depth model, on the other hand, the reason of the depth model has an independent technical system based on ONNX, so MetaSpore disassembled this part of preprocessing logic. NLP pretreatment Tokenizer has been integrated into the Query pretreatment service. MetaSpore dismantlement with a relatively general convention. Users only need to provide preprocessing logic files to realize the loading and prediction interface and export the necessary data and configuration files loaded into the preprocessing service. Subsequent CV preprocessing logic will also be integrated in this manner. The preprocessing service currently provides the gRPC interface invocation externally and is dependent on the Query preprocessing (QP) module in the retrieval algorithm service. After the user request reaches the retrieval algorithm service, it will be forwarded to the service to complete the data preprocessing and continue the subsequent processing. The ReadMe provides details on how the preprocessing service is started, how the preprocessing model exported offline to cloud storage enters the service, and how to debug the service. To further improve the efficiency and stability of model reasoning, MetaSpore Serving implements a Python preprocessing submodule. So MetaSpore can provide gRPC services through user-specified preprocessor.py, complete Tokenizer or CV-related preprocessing in NLP, and translate requests into a Tensor that deep models can handle. Finally, the model inference is carried out by MetaSpore, Serving subsequent sub-modules. Presented here on the lot code: https://github.com/meta-soul/MetaSpore/compare/add\python\preprocessor 1.2.2 Retrieval algorithm services Retrieval algorithm service is the core of the whole online service system, which is responsible for the triage of experiments, the assembly of algorithm chains such as preprocessing, recall, sorting, and the invocation of dependent component services. The whole retrieval algorithm service is developed based on the Java Spring framework and supports multi-mode retrieval scenarios of text search and text search graph. Due to good internal abstraction and modular design, it has high flexibility and can be migrated to similar application scenarios at a low cost. Here’s a quick guide to configuring the environment to set up the retrieval algorithm service. See ReadME for more details: Install dependent components. Use Maven to install the online-Serving component Search for service configurations. Copy the template configuration file and replace the MongoDB, Milvus, and other configurations based on the development/production environment. Install and configure Consul. Consul allows you to synchronize the search service configuration in real-time, including cutting the flow of experiments, recall parameters, and sorting parameters. The project’s configuration file shows the current configuration parameters of text search and text search. The parameter modelName in the stage of pretreatment and recall is the corresponding model exported in offline processing. Start the service. Once the above configuration is complete, the retrieval service can be started from the entry script. Once the service is started, you can test it! For example, for a user with userId=10 who wants to query “How to renew ID card,” access the text search service. 1.2.3 User Entry Service Considering that the retrieval algorithm service is in the form of the API interface, it is difficult to locate and trace the problem, especially for the text search image scene can intuitively display the retrieval results to facilitate the iterative optimization of the retrieval algorithm. This paper provides a lightweight Web UI interface for text search and image search, a search input box, and results in a display page for users. Developed by Flask, the service can be easily integrated with other retrieval applications. The service calls the retrieval algorithm service and displays the returned results on the page. It’s also easy to install and start the service. Once you’re done, go to http://127.0.0.1:8090 to see if the search UI service is working correctly. See the ReadME at the end of this article for details. Multimodal system demonstration The multimodal retrieval service can be started when offline processing and online service environment configuration have been completed following the above instructions. Examples of textual searches are shown below. Enter the entry of the text search map application, enter “cat” first, and you can see that the first three digits of the returned result are cats: https://preview.redd.it/d7syq47rdz291.png?width=1280&format=png&auto=webp&s=b43df9abd380b7d9a52e3045dd787f4feeb69635 If you add a color constraint to “cat” to retrieve “black cat,” you can see that it does return a black cat: ​ https://preview.redd.it/aa7pxx8tdz291.png?width=1280&format=png&auto=webp&s=e3727c29d1bde6eea2e1cccf6c46d3cae3f4750e Further, strengthen the constraint on the search term, change it to “black cat on the bed,” and return results containing pictures of a black cat climbing on the bed: ​ https://preview.redd.it/2mw4qpjudz291.png?width=1280&format=png&auto=webp&s=1cf1db667892b9b3a40451993680fbd6980b5520 The cat can still be found through the text search system after the color and scene modification in the above example. Conclusion The cutting-edge pre-training technology can bridge the semantic gap between different modes, and the HuggingFace community can greatly reduce the cost for developers to use the pre-training model. Combined with the technological ecology of MetaSpore online reasoning and online microservices provided by DMetaSpore, the pre-training model is no longer mere offline dabbling. Instead, it can truly achieve end-to-end implementation from cutting-edge technology to industrial scenarios, fully releasing the dividends of the pre-training large model. In the future, DMetaSoul will continue to improve and optimize the MetaSpore technology ecosystem: More automated and wider access to HuggingFace community ecology. MetaSpore will soon release a common model rollout mechanism to make HuggingFace ecologically accessible and will later integrate preprocessing services into online services. Multi-mode retrieval offline algorithm optimization. For multimodal retrieval scenarios, MetaSpore will continuously iteratively optimize offline algorithm components, including text recall/sort model, graphic recall/sort model, etc., to improve the accuracy and efficiency of the retrieval algorithm. For related code and reference documentation in this article, please visit: https://github.com/meta-soul/MetaSpore/tree/main/demo/multimodal/online Some images source: https://github.com/openai/CLIP/raw/main/CLIP.png https://www.sbert.net/examples/training/sts/README.html

My Marketing App made $10,000 in 2024. Here is how I target to make $100,000 in 2025:
reddit
LLM Vibe Score0
Human Vibe Score1
MonkDiThis week

My Marketing App made $10,000 in 2024. Here is how I target to make $100,000 in 2025:

You totally get me, I think. It’s a bizarre feeling when you build something, and people appreciate it and are even ready to pay! Pleasant though) In early 2024 my mate and I created a marketing tool that generates ads, content and strategy blocks with a click – Aiter.io. Users can just insert a URL, hit the button and everything is ready. TBH, I built this tool because I’m too lazy to chat with ChatGPT) https://preview.redd.it/ew2kud7ceyde1.png?width=1140&format=png&auto=webp&s=f3fe5b67075858cea3d52278e8063113efa3b97e In 2024 we made $10,000, here is what worked for us: AI directories. Still is the best channel of traffic and clients for us. We listed on TAAFT and other directories scrape TAAFT, so, eventually, we became listed on all major ones. I wrote a Reddit post earlier that explained this process in detail. Email marketing. Gosh, I thought it was dead – I have never been so wrong! We set up automatic emails that share marketing insights and they have a \~25% open rate + consistently convert people. It works great. Product marketing. Having a free version really helps with word-of-mouth and leads, which can be converted via email. Also, we consistently worked on product improvement. I’d say, that our free updates give people a feeling that the devs care about their stuff that’s why they are more confident investing in it. Google Ads. TBH, we had a shitty landing page all the time because were busy with the product. So, Google Ads didn’t work well for us. But we’ve launched the 2.0 version which has a better landing page, and will try it again. Influencers. Worked well for us, but we didn’t pay a dime for this. They just found our tool on directories and created videos about Aiter, so it was a sporadic marketing channel for us. We hope to change it in 2025. We see that our product works and attracts the audience, so we want to deliver and get more in 2025. Here is the plan: Product: add ad banners and video generation. So far, we generate only text data and it’s not so valuable in the time of ChatGPT and Claude. But to generate a high-quality ad banner is still challenging, so we put this on our roadmap. Another feature – one-click market analysis to get marketing insights. Become a TOP50 tool on TAAFT. We’ve become a top tool in our category (content generation) but will need to promote our profile on the profile far more aggressively to get into TOP50 Email marketing. We are fools because we almost didn’t have product emails that explain how it works. Will fix it. Also, we are considering participating more in paid newsletters, like collaborating with Substack influencers. Youtube marketing. Search for low-tail marketing keywords on YouTube and create videos on them, placing my product in them. Blog. Our new platform is Webflow which gives a lot of flexibility in terms of blogging. So, we will repeat the YouTube strategy with blogging. Paid marketing. With an updated landing page, we hope that paid campaigns will work better. We plan to launch campaigns that target different jobs to be done and customer objections to find the right message. Product Management. For 2025, our two key product metrics are retention and product activation rate. For this, we plan to simplify onboarding and make it simpler as well as conduct a lot of in-depth interviews to understand how we can retain users better. Funding. All of this exciting stuff requires money, so we are in the process of securing funding (fingers crossed). Having an indie project is exciting and invigorating. With all these activities, I hope we will achieve the goal of $100,000 in 2025. And what are your goals and marketing steps for 2025? Or maybe you could share some exciting marketing ideas I overlooked?

I grew my mobile app to 1.4 million downloads
reddit
LLM Vibe Score0
Human Vibe Score1
TechPrimoThis week

I grew my mobile app to 1.4 million downloads

I started developing the app in early 2017, well before the AI era, when mobile apps were at their peak popularity. My idea was to create an app for emotional and psychological support in the form of helpful articles and various quizzes, such as personality assessments and life satisfaction tests. I named the app "Emotional Intelligence" because this keyword showed good ASO potential for positioning at the top of mobile stores. This proved to be accurate, and the app quickly gained traction in terms of downloads. A major problem I faced then was monetization. Unfortunately, in my country, it wasn't possible to sell through Google Play then, so I could only display ads. I started with Google AdMob, earning $2000 monthly after just a few months. The app then got about 1500 organic downloads daily and quickly surpassed 500,000. Three years after launching the app, I decided it was time for branding to build recognition. By combining the words "sentiment" and "intelligence," I came up with "Sintelly." I then pushed the app toward a social network, which differed from the right move. Adding features like discussion forums for problems, likes, and comments would result in even more growth, but the opposite happened. The app started declining, and I began investing in advertising campaigns. I managed to maintain a balance between income and expenses but without any profit. Then COVID-19 hit, and everything went downhill. I had to give up development and find a job as a developer to ensure my livelihood. Two years passed since I gave up, and that's when ChatGPT started gaining popularity. This immediately showed me how to steer the app towards active support for well-being questions. As I'm not an expert in psychology, I found several external psychotherapists who helped me put together CBT therapy, which I then implemented through a chatbot. This is how the new Sintelly app was born, with its main feature being a chatbot system composed of 17 AI agents that adapt to the user and guide them through a five-phase CBT therapy (I'll write a post about the technology). In addition to the agents, I added various exercises and tests to provide better personalization for the user. Initially, I made all of this free, which was also a mistake. I followed the principle of first showing what the app can do and gathering enough new users before starting to charge. I started selling subscriptions at the beginning of July, and since then, the app has had stable growth. If you want a check app, here is the link. Lessons learned: If things are working, don't touch them Start selling immediately upon app release; there's no need to wait Regularly test prices and types of subscriptions Onboarding is the most essential part of the app because most users buy subscriptions during onboarding It's essential to listen to user feedback. From day one, have a website and work on content to generate organic visits and redirect users from the web to the mobile app Stats: Over 1.4 million downloads 4.4 rating Only 40,000 active users (I had a massive loss during the period when I gave up) 280 active subscribers $3000 monthly revenue Next steps: Work on improving the Agent AI approach Setting up email campaigns and transactional emails Introducing in-app and push notifications Introducing gamification Potential for B2B I hope you can extract useful information from my example and avoid repeating my mistakes. I'm interested in your thoughts and if you have any recommendations for the next steps. I'm always looking to learn and improve.

I am building my agency to help founders build AI startups after 2 successful AI SaaS exits and 4 failures
reddit
LLM Vibe Score0
Human Vibe Score1
_Gautam19This week

I am building my agency to help founders build AI startups after 2 successful AI SaaS exits and 4 failures

Hey everyone, I have been building AI products before ChatGPT was launched. In these years, I have managed to launch, scale and exit 2 SaaS products successfully. Today I am launching a new service offering - Query Labs - Helping you build AI agents for your startups. Like all my previous products, I will be building this in public and share my learning along the way. Here's what I have built so far : Microsponsors ( Fail ) My first product ever. I tried to create a marketplace for newsletter writers to find sponsorship opportunity. Got a few very big newsletter listed on the marketplace as well. However, building marketplace is tough. I found it very difficult to bring in sponsors. Ended up shutting it down, AI Query (Exit - Pre revenue ) It was the second half of 2022 and GPT-3 was the most advance AI on the market. I decided to build a tool that can help developers and non-technical folks write SQL queries by just asking in plain english. I got my first taste of success with this. Had a decent offer even before I figured out monetisation. Accepted the offer to focus on my next product which had already started gaining traction AI Excel Bot ( Exit - Revenue Generating ) AI Excel Bot was my wild success. I had worked hard on the SEO for the site, along with the UI / UX to make it the best AI to write excel formulas and general excel task. There was already a large competitor in the market. However, the reality is that you don't need to be the top player. There is always room for multiple players to survive in a large market. You just need to find the good differentiating factor For AI Excel Bot, the differentiator was the chrome extension, that helped users access it anywhere on the internet. Scaled the product to more than 40k users at the time of exit. However, in the end I decided to exit and focus on my software service business that needed more time. Tutore AI ( Fail ) I wanted to build something useful for students to help them learn better. Tutore was my idea to build AI tools for students. I did launch quickly with multiple tools. However, wasn't motivated enough to continue with the grind. I have decided to sell the product. Have had some meetings with potential buyers but didn't agree on price. Prompt Hackers ( 1k users but no revenue ) Prompt Hackers is a directory of AI prompts for all the use cases you can image. I focused a lot on bringing traffic and newsletter subscription from the day 1. I have never had a problem bringing initial set of users to my products. Prompt Hackers was getting close to 20k page views a month. At the same time we had close to 1k newsletter subscribers. Since our target customers were people choosing to use ChatGPT / Bard instead of some specific software for their task, I built a Prompt Generation and Prompt Optimisation AI. Along with this I also created features to build private prompt library. To make the experience even better, I launched a Chrome Extension that helps users access the prompt generation AI and their prompt library while using ChatGPT. However, I couldn't figure out monetisation. I still get close to 4k page views per month with no marketing at all. There are users who use the AI tools and the prompt library feature daily. But, since I couldn't figure out monetisation, I decided to not put time into the project. There you go. These are all the products I have built in the last 3 years. I have been heavy investing myself in the latest tech in LLMs and AI agents. I know the biggest challenge for AI founders is the AI agents and backend pipelines. That's why I am launching Query Labs. To help you build the best AI implementation for your innovative AI startup. I would love to hear feedback from the community. I will be sharing my learning with my new service along the way. Thanks!

Introducing Vest: Your AI-Powered Due Diligence Partner - Looking for feedback!
reddit
LLM Vibe Score0
Human Vibe Score1
nervousslinkyThis week

Introducing Vest: Your AI-Powered Due Diligence Partner - Looking for feedback!

TLDR; We are introducing Vest, an AI powered due-diligence and stock recommendation platform. We have bootstrapped ourselves so far and are wanting to get as much feedback from Reddit as we can to see where we can improve, but also what we are doing right. So please have a look around, give us feedback and if you like it, feel free to use it. Hi Reddit, My name is Drian and I'm one of the founders of Vest. We believe we are crafting something special at Vest and we want to get the word out and gather as much feedback as possible! Our major goal at Vest is to help new retail investors make sense of the investment landscape and get AI powered assistance, or even help experienced investors get confirmation of their potential moves. Overall, we want people to start their journey to financial freedom and not be daunted by the complexity of it. So how do we do this? Vest is a user-friendly service that harnesses fundamental metrics, social and news sentiment, and technical analysis, that we feed into some advanced AI models to generate clear buy, sell, or hold signals for US-based (for now!) stocks, offering our users transparent due-diligence for confident investing. The service is currently free with no ads - however, at some point we do plan on adding a paid tier. What's included: ​ Financial Metrics. Our financial metrics take all the potentially complex mathematical equations and present the fundamentals of a company to users in a simple 1 pager, with a score displaying if the metric is positive for a stock. We also provide publicly available analyst ratings from investment banks as well as price targets they have set. News Sentiment. We take publications about a specific stock from new articles, journals and socials and give these all a rating to determine if social sentiment is positive around a stock or not. Each article and its rating is visible to our users through through our dashboard. AI assisted Stock Signals. We have developed an algorithm to take all the metrics, sentiment and technical analysis we collate and analyze this with historic performance data for every stock to attempt to figure out if a stock is undervalued (great time to buy) or overvalued (great time to sell). 155 US stock tickers and counting. We currently have trained our models for around 155 US based stocks on the NASDAQ and NYSE exchanges. As we get more funding/runway we do plan on adding more, with the eventual goal to expand to more exchanges, countries and securities. Knowledge base and community. Our knowledge base & community contains explanations and articles for all metrics and the other good stuff behind Vest. We don’t want to just tell users what to do, but to also assist in their financial education. We hope our knowledge base can also become a thriving community where users can interact with us and each, ask questions around investing and keep gaining knowledge. Is it 100% accurate? Absolutely not. While we do a pretty great job at tracking and surfacing signals, we are not presenting a fool-proof, silver bullet with a guarantee here - rather a starting point for users to make more informed decisions, find potential new investment opportunities and hopefully learn about investing as they do so. We encourage our users to do their own research and due-diligence and not just take our signals as gospel - we know each and every person has a different risk appetite and goals, and we encourage you to use Vest in a way that fits with your own financial goals and risk appetite. We also display our win rates, average returns, and comparisons with buy and hold for each stock - and we are transparent about it when we’ve fallen short. Next steps: ​ Hope over to vestapp.ai and sign-up From the dashboard, play around, inspect our stock information and add some stocks to your watchlist. If you like what you see, and you’ve done your homework - use your favourite brokerage account to make an investment and watch Vest for changes in a stocks signals. If you don’t have one, we have a pop-up when you click buy/sell on any given stock with some non-affiliated brokerage options for the US, Australia and New Zealand - we don’t get a kickback from these brokerages, they are just what we’ve personally been using. FEEDBACK - We’re just getting started and we know the value of a fresh pair of eyes - our current mission is to get as much feedback as possible - anything you think of please send it through here or on the dedicated feedback form on our website in the sidebar on the left. Features we’re working on We're quietly thrilled about the direction Vest is headed, and we want to give you a sneak peek of what's in store for the next couple of quarters. Some of these may roll out as premium features, but we're diligently fine-tuning the details. Here's what you can expect: ​ Insider Trading Insights: Get daily reports on major stock moves by whales and company insiders. Institutional Holders: We're adding daily reports on institutional holders, keeping you informed about their moves. Lobbying Activity: We're actively working on daily updates about lobbying activities, so you can stay informed. Government Contracts Data: We'll provide a quarterly snapshot of government contract values for the companies you're tracking. US Congress Stock Activity: Keep an eye on daily trading actions of House and Senate members. Daily Summaries & Signal Alerts: We're currently hard at work on this feature. Soon, receive daily email summaries covering signals, watchlist updates, and key news. Personalized Risk Management: Tailor signals to match your unique risk management strategy. Your investments, your way. AI Assistant: Our LLM integration is almost ready, allowing you to ask it straightforward questions about particular securities in plain English. It will provide you with real-time context on fundamentals, news, and all the metrics and data points we monitor.

AI-Powered Business Analyst Tool Looking for Feedback
reddit
LLM Vibe Score0
Human Vibe Score1
ondro949This week

AI-Powered Business Analyst Tool Looking for Feedback

Hey r/sideproject! I’m excited to share a project I’ve been working on called Bianalytiq, a next-gen business intelligence platform designed to transform the way businesses interact with data through the power of AI. The Problem: SME companies struggle with data overload and the significant time investment required to generate actionable insights. Traditional data analysis methods are not only slow but often require extensive manual effort and are prone to errors. This makes it difficult for businesses to react quickly to new information and make informed decisions efficiently. Not everybody can write SQL or create/understand data dashboards.... AND - one big opportunity on market - non of the AI tools available on market offer reusable contexts focused on you as a company and your products. The Solution: Bianalytiq aims to solve these issues by automating tedious data analysis tasks and providing real-time insights. Here’s how: Reusable contexts: Let Bianalytiq learn everything about your company, your products, business model etc. - your company is your unique context. Autonomous AI Agents: Deploy AI agents that not only react to queries but proactively analyze data to uncover opportunities, tailored specifically to your business context. Real-Time Insights: With the use of Retrieval-Augmented Generation (RAG) technology, our platform delivers immediate, context-rich insights by dynamically accessing and analyzing connected databases and data warehouses. Integration with Existing Tools: Bianalytiq integrates seamlessly with popular tech stacks and communication platforms like Slack and Microsoft Teams, making it incredibly user-friendly and reducing the switch cost between applications. Why I’m Here: Before investing significant time and money I want to validate the product first and do pre-sale before releasing the MVP. I’ve developed a landing page for Bianalytiq and would love your feedback on both the service itself and the effectiveness of the landing page. Are the features presented clearly? Does the platform address the pain points you might experience in data analysis and decision-making processes? Here’s the link to the landing page: https://bianalytiq.com/ I appreciate any feedback or questions you have! Whether it's about the UI/UX of the site, the technical aspects of the service, or even the business model, I’m all ears. Your input will be invaluable :) Thanks for checking it out! https://preview.redd.it/t1dvp2q05dzc1.png?width=798&format=png&auto=webp&s=c7365b418abfc4d4260d9a23305ed3398e83c87b

I built an AI Stock Analysis Tool
reddit
LLM Vibe Score0
Human Vibe Score1
HenryObjThis week

I built an AI Stock Analysis Tool

Hi Reddit, TL;DR: I am sharing the tool I built to assist me with my investments I have been investing for over a decade, and I have always struggled with: Putting the time to do actual research Trusting analyst’s recommendations To \ invest \ we want to make sure that the price is right and the company will keep improving. How do we know that “the price is right?” How can we predict that the company will perform better in the future? To answer the above, we have to look at the company’s financials and their trends. We have to compare the company with its peers/competitors. We should understand its business model, the sector and geography the company evolves in and the perspective of the economy in general. We can also look at additional signals like insiders selling or buying. Just for one investment, this is already a lot of work. And a work that we need to repeat every time there is a significant change - for example, a significant price change from our last analysis or new quarterly results, etc. To automate all the above, I have built a stock analysis tool and have been using it the past years for my own investments. I have been adding LLMs agents (GPT 4o & Claude 3.5) to perform the qualitative analysis. Recently, I decided to share it and keep on building it in public. In this initial version, you can get the summary of the stock analysis my model generates. For now, it covers most of the S&P and Nasdaq stocks. Here is the link 👉 https://undervalued.ai If you are into investing yourself, please feel free to reach out. I would love to get your feedback and know more about your methodology.

[D] What is your honest experience with reinforcement learning?
reddit
LLM Vibe Score0
Human Vibe Score1
Starks-TechnologyThis week

[D] What is your honest experience with reinforcement learning?

In my personal experience, SOTA RL algorithms simply don't work. I've tried working with reinforcement learning for over 5 years. I remember when Alpha Go defeated the world famous Go player, Lee Sedol, and everybody thought RL would take the ML community by storm. Yet, outside of toy problems, I've personally never found a practical use-case of RL. What is your experience with it? Aside from Ad recommendation systems and RLHF, are there legitimate use-cases of RL? Or, was it all hype? Edit: I know a lot about AI. I built NexusTrade, an AI-Powered automated investing tool that lets non-technical users create, update, and deploy their trading strategies. I’m not an idiot nor a noob; RL is just ridiculously hard. Edit 2: Since my comments are being downvoted, here is a link to my article that better describes my position. It's not that I don't understand RL. I released my open-source code and wrote a paper on it. It's the fact that it's EXTREMELY difficult to understand. Other deep learning algorithms like CNNs (including ResNets), RNNs (including GRUs and LSTMs), Transformers, and GANs are not hard to understand. These algorithms work and have practical use-cases outside of the lab. Traditional SOTA RL algorithms like PPO, DDPG, and TD3 are just very hard. You need to do a bunch of research to even implement a toy problem. In contrast, the decision transformer is something anybody can implement, and it seems to match or surpass the SOTA. You don't need two networks battling each other. You don't have to go through hell to debug your network. It just naturally learns the best set of actions in an auto-regressive manner. I also didn't mean to come off as arrogant or imply that RL is not worth learning. I just haven't seen any real-world, practical use-cases of it. I simply wanted to start a discussion, not claim that I know everything. Edit 3: There's a shockingly number of people calling me an idiot for not fully understanding RL. You guys are wayyy too comfortable calling people you disagree with names. News-flash, not everybody has a PhD in ML. My undergraduate degree is in biology. I self-taught myself the high-level maths to understand ML. I'm very passionate about the field; I just have VERY disappointing experiences with RL. Funny enough, there are very few people refuting my actual points. To summarize: Lack of real-world applications Extremely complex and inaccessible to 99% of the population Much harder than traditional DL algorithms like CNNs, RNNs, and GANs Sample inefficiency and instability Difficult to debug Better alternatives, such as the Decision Transformer Are these not legitimate criticisms? Is the purpose of this sub not to have discussions related to Machine Learning? To the few commenters that aren't calling me an idiot...thank you! Remember, it costs you nothing to be nice! Edit 4: Lots of people seem to agree that RL is over-hyped. Unfortunately those comments are downvoted. To clear up some things: We've invested HEAVILY into reinforcement learning. All we got from this investment is a robot that can be super-human at (some) video games. AlphaFold did not use any reinforcement learning. SpaceX doesn't either. I concede that it can be useful for robotics, but still argue that it's use-cases outside the lab are extremely limited. If you're stumbling on this thread and curious about an RL alternative, check out the Decision Transformer. It can be used in any situation that a traditional RL algorithm can be used. Final Edit: To those who contributed more recently, thank you for the thoughtful discussion! From what I learned, model-based models like Dreamer and IRIS MIGHT have a future. But everybody who has actually used model-free models like DDPG unanimously agree that they suck and don’t work.

[D] What is your honest experience with reinforcement learning?
reddit
LLM Vibe Score0
Human Vibe Score1
Starks-TechnologyThis week

[D] What is your honest experience with reinforcement learning?

In my personal experience, SOTA RL algorithms simply don't work. I've tried working with reinforcement learning for over 5 years. I remember when Alpha Go defeated the world famous Go player, Lee Sedol, and everybody thought RL would take the ML community by storm. Yet, outside of toy problems, I've personally never found a practical use-case of RL. What is your experience with it? Aside from Ad recommendation systems and RLHF, are there legitimate use-cases of RL? Or, was it all hype? Edit: I know a lot about AI. I built NexusTrade, an AI-Powered automated investing tool that lets non-technical users create, update, and deploy their trading strategies. I’m not an idiot nor a noob; RL is just ridiculously hard. Edit 2: Since my comments are being downvoted, here is a link to my article that better describes my position. It's not that I don't understand RL. I released my open-source code and wrote a paper on it. It's the fact that it's EXTREMELY difficult to understand. Other deep learning algorithms like CNNs (including ResNets), RNNs (including GRUs and LSTMs), Transformers, and GANs are not hard to understand. These algorithms work and have practical use-cases outside of the lab. Traditional SOTA RL algorithms like PPO, DDPG, and TD3 are just very hard. You need to do a bunch of research to even implement a toy problem. In contrast, the decision transformer is something anybody can implement, and it seems to match or surpass the SOTA. You don't need two networks battling each other. You don't have to go through hell to debug your network. It just naturally learns the best set of actions in an auto-regressive manner. I also didn't mean to come off as arrogant or imply that RL is not worth learning. I just haven't seen any real-world, practical use-cases of it. I simply wanted to start a discussion, not claim that I know everything. Edit 3: There's a shockingly number of people calling me an idiot for not fully understanding RL. You guys are wayyy too comfortable calling people you disagree with names. News-flash, not everybody has a PhD in ML. My undergraduate degree is in biology. I self-taught myself the high-level maths to understand ML. I'm very passionate about the field; I just have VERY disappointing experiences with RL. Funny enough, there are very few people refuting my actual points. To summarize: Lack of real-world applications Extremely complex and inaccessible to 99% of the population Much harder than traditional DL algorithms like CNNs, RNNs, and GANs Sample inefficiency and instability Difficult to debug Better alternatives, such as the Decision Transformer Are these not legitimate criticisms? Is the purpose of this sub not to have discussions related to Machine Learning? To the few commenters that aren't calling me an idiot...thank you! Remember, it costs you nothing to be nice! Edit 4: Lots of people seem to agree that RL is over-hyped. Unfortunately those comments are downvoted. To clear up some things: We've invested HEAVILY into reinforcement learning. All we got from this investment is a robot that can be super-human at (some) video games. AlphaFold did not use any reinforcement learning. SpaceX doesn't either. I concede that it can be useful for robotics, but still argue that it's use-cases outside the lab are extremely limited. If you're stumbling on this thread and curious about an RL alternative, check out the Decision Transformer. It can be used in any situation that a traditional RL algorithm can be used. Final Edit: To those who contributed more recently, thank you for the thoughtful discussion! From what I learned, model-based models like Dreamer and IRIS MIGHT have a future. But everybody who has actually used model-free models like DDPG unanimously agree that they suck and don’t work.

[P]MMML | Deploy HuggingFace training model rapidly based on MetaSpore
reddit
LLM Vibe Score0
Human Vibe Score1
qazmkoppThis week

[P]MMML | Deploy HuggingFace training model rapidly based on MetaSpore

A few days ago, HuggingFace announced a $100 million Series C funding round, which was big news in open source machine learning and could be a sign of where the industry is headed. Two days before the HuggingFace funding announcement, open-source machine learning platform MetaSpore released a demo based on the HuggingFace Rapid deployment pre-training model. As deep learning technology makes innovative breakthroughs in computer vision, natural language processing, speech understanding, and other fields, more and more unstructured data are perceived, understood, and processed by machines. These advances are mainly due to the powerful learning ability of deep learning. Through pre-training of deep models on massive data, the models can capture the internal data patterns, thus helping many downstream tasks. With the industry and academia investing more and more energy in the research of pre-training technology, the distribution warehouses of pre-training models such as HuggingFace and Timm have emerged one after another. The open-source community release pre-training significant model dividends at an unprecedented speed. In recent years, the data form of machine modeling and understanding has gradually evolved from single-mode to multi-mode, and the semantic gap between different modes is being eliminated, making it possible to retrieve data across modes. Take CLIP, OpenAI’s open-source work, as an example, to pre-train the twin towers of images and texts on a dataset of 400 million pictures and texts and connect the semantics between pictures and texts. Many researchers in the academic world have been solving multimodal problems such as image generation and retrieval based on this technology. Although the frontier technology through the semantic gap between modal data, there is still a heavy and complicated model tuning, offline data processing, high performance online reasoning architecture design, heterogeneous computing, and online algorithm be born multiple processes and challenges, hindering the frontier multimodal retrieval technologies fall to the ground and pratt &whitney. DMetaSoul aims at the above technical pain points, abstracting and uniting many links such as model training optimization, online reasoning, and algorithm experiment, forming a set of solutions that can quickly apply offline pre-training model to online. This paper will introduce how to use the HuggingFace community pre-training model to conduct online reasoning and algorithm experiments based on MetaSpore technology ecology so that the benefits of the pre-training model can be fully released to the specific business or industry and small and medium-sized enterprises. And we will give the text search text and text search graph two multimodal retrieval demonstration examples for your reference. Multimodal semantic retrieval The sample architecture of multimodal retrieval is as follows: Our multimodal retrieval system supports both text search and text search application scenarios, including offline processing, model reasoning, online services, and other core modules: ​ https://preview.redd.it/w4v4c7vcez291.png?width=1834&format=png&auto=webp&s=0687efb1fddb26e8e30cb844d398ec712b947f31 Offline processing, including offline data processing processes for different application scenarios of text search and text search, including model tuning, model export, data index database construction, data push, etc. Model inference. After the offline model training, we deployed our NLP and CV large models based on the MetaSpore Serving framework. MetaSpore Serving helps us conveniently perform online inference, elastic scheduling, load balancing, and resource scheduling in heterogeneous environments. Online services. Based on MetaSpore’s online algorithm application framework, MetaSpore has a complete set of reusable online search services, including Front-end retrieval UI, multimodal data preprocessing, vector recall and sorting algorithm, AB experimental framework, etc. MetaSpore also supports text search by text and image scene search by text and can be migrated to other application scenarios at a low cost. The HuggingFace open source community has provided several excellent baseline models for similar multimodal retrieval problems, which are often the starting point for actual optimization in the industry. MetaSpore also uses the pre-training model of the HuggingFace community in its online services of searching words by words and images by words. Searching words by words is based on the semantic similarity model of the question and answer field optimized by MetaSpore, and searching images by words is based on the community pre-training model. These community open source pre-training models are exported to the general ONNX format and loaded into MetaSpore Serving for online reasoning. The following sections will provide a detailed description of the model export and online retrieval algorithm services. The reasoning part of the model is standardized SAAS services with low coupling with the business. Interested readers can refer to my previous post: The design concept of MetaSpore, a new generation of the one-stop machine learning platform. 1.1 Offline Processing Offline processing mainly involves the export and loading of online models and index building and pushing of the document library. You can follow the step-by-step instructions below to complete the offline processing of text search and image search and see how the offline pre-training model achieves reasoning at MetaSpore. 1.1.1 Search text by text Traditional text retrieval systems are based on literal matching algorithms such as BM25. Due to users’ diverse query words, a semantic gap between query words and documents is often encountered. For example, users misspell “iPhone” as “Phone,” and search terms are incredibly long, such as “1 \~ 3 months old baby autumn small size bag pants”. Traditional text retrieval systems will use spelling correction, synonym expansion, search terms rewriting, and other means to alleviate the semantic gap but fundamentally fail to solve this problem. Only when the retrieval system fully understands users’ query terms and documents can it meet users’ retrieval demands at the semantic level. With the continuous progress of pre-training and representational learning technology, some commercial search engines continue to integrate semantic vector retrieval methods based on symbolic learning into the retrieval ecology. Semantic retrieval model This paper introduces a set of semantic vector retrieval applications. MetaSpore built a set of semantic retrieval systems based on encyclopedia question and answer data. MetaSpore adopted the Sentence-Bert model as the semantic vector representation model, which fine-tunes the twin tower BERT in supervised or unsupervised ways to make the model more suitable for retrieval tasks. The model structure is as follows: The query-Doc symmetric two-tower model is used in text search and question and answer retrieval. The vector representation of online Query and offline DOC share the same vector representation model, so it is necessary to ensure the consistency of the offline DOC library building model and online Query inference model. The case uses MetaSpore’s text representation model Sbert-Chinese-QMC-domain-V1, optimized in the open-source semantically similar data set. This model will express the question and answer data as a vector in offline database construction. The user query will be expressed as a vector by this model in online retrieval, ensuring that query-doc in the same semantic space, users’ semantic retrieval demands can be guaranteed by vector similarity metric calculation. Since the text presentation model does vector encoding for Query online, we need to export the model for use by the online service. Go to the q&A data library code directory and export the model concerning the documentation. In the script, Pytorch Tracing is used to export the model. The models are exported to the “./export “directory. The exported models are mainly ONNX models used for wired reasoning, Tokenizer, and related configuration files. The exported models are loaded into MetaSpore Serving by the online Serving system described below for model reasoning. Since the exported model will be copied to the cloud storage, you need to configure related variables in env.sh. \Build library based on text search \ The retrieval database is built on the million-level encyclopedia question and answer data set. According to the description document, you need to download the data and complete the database construction. The question and answer data will be coded as a vector by the offline model, and then the database construction data will be pushed to the service component. The whole process of database construction is described as follows: Preprocessing, converting the original data into a more general JSonline format for database construction; Build index, use the same model as online “sbert-Chinese-qmc-domain-v1” to index documents (one document object per line); Push inverted (vector) and forward (document field) data to each component server. The following is an example of the database data format. After offline database construction is completed, various data are pushed to corresponding service components, such as Milvus storing vector representation of documents and MongoDB storing summary information of documents. Online retrieval algorithm services will use these service components to obtain relevant data. 1.1.2 Search by text Text and images are easy for humans to relate semantically but difficult for machines. First of all, from the perspective of data form, the text is the discrete ID type of one-dimensional data based on words and words. At the same time, images are continuous two-dimensional or three-dimensional data. Secondly, the text is a subjective creation of human beings, and its expressive ability is vibrant, including various turning points, metaphors, and other expressions, while images are machine representations of the objective world. In short, bridging the semantic gap between text and image data is much more complex than searching text by text. The traditional text search image retrieval technology generally relies on the external text description data of the image or the nearest neighbor retrieval technology and carries out the retrieval through the image associated text, which in essence degrades the problem to text search. However, it will also face many issues, such as obtaining the associated text of pictures and whether the accuracy of text search by text is high enough. The depth model has gradually evolved from single-mode to multi-mode in recent years. Taking the open-source project of OpenAI, CLIP, as an example, train the model through the massive image and text data of the Internet and map the text and image data into the same semantic space, making it possible to implement the text and image search technology based on semantic vector. CLIP graphic model The text search pictures introduced in this paper are implemented based on semantic vector retrieval, and the CLIP pre-training model is used as the two-tower retrieval architecture. Because the CLIP model has trained the semantic alignment of the twin towers’ text and image side models on the massive graphic and text data, it is particularly suitable for the text search graph scene. Due to the different image and text data forms, the Query-Doc asymmetric twin towers model is used for text search image retrieval. The image-side model of the twin towers is used for offline database construction, and the text-side model is used for the online return. In the final online retrieval, the database data of the image side model will be searched after the text side model encodes Query, and the CLIP pre-training model guarantees the semantic correlation between images and texts. The model can draw the graphic pairs closer in vector space by pre-training on a large amount of visual data. Here we need to export the text-side model for online MetaSpore Serving inference. Since the retrieval scene is based on Chinese, the CLIP model supporting Chinese understanding is selected. The exported content includes the ONNX model used for online reasoning and Tokenizer, similar to the text search. MetaSpore Serving can load model reasoning through the exported content. Build library on Image search You need to download the Unsplash Lite library data and complete the construction according to the instructions. The whole process of database construction is described as follows: Preprocessing, specify the image directory, and then generate a more general JSOnline file for library construction; Build index, use OpenAI/Clip-Vit-BASE-Patch32 pre-training model to index the gallery, and output one document object for each line of index data; Push inverted (vector) and forward (document field) data to each component server. Like text search, after offline database construction, relevant data will be pushed to service components, called by online retrieval algorithm services to obtain relevant data. 1.2 Online Services The overall online service architecture diagram is as follows: https://preview.redd.it/jfsl8hdfez291.png?width=1280&format=png&auto=webp&s=a858e2304a0c93e78ba5429612ca08cbee69b35a Multi-mode search online service system supports application scenarios such as text search and text search. The whole online service consists of the following parts: Query preprocessing service: encapsulate preprocessing logic (including text/image, etc.) of pre-training model, and provide services through gRPC interface; Retrieval algorithm service: the whole algorithm processing link includes AB experiment tangent flow configuration, MetaSpore Serving call, vector recall, sorting, document summary, etc.; User entry service: provides a Web UI interface for users to debug and track down problems in the retrieval service. From a user request perspective, these services form invocation dependencies from back to front, so to build up a multimodal sample, you need to run each service from front to back first. Before doing this, remember to export the offline model, put it online and build the library first. This article will introduce the various parts of the online service system and make the whole service system step by step according to the following guidance. See the ReadME at the end of this article for more details. 1.2.1 Query preprocessing service Deep learning models tend to be based on tensors, but NLP/CV models often have a preprocessing part that translates raw text and images into tensors that deep learning models can accept. For example, NLP class models often have a pre-tokenizer to transform text data of string type into discrete tensor data. CV class models also have similar processing logic to complete the cropping, scaling, transformation, and other processing of input images through preprocessing. On the one hand, considering that this part of preprocessing logic is decoupled from tensor reasoning of the depth model, on the other hand, the reason of the depth model has an independent technical system based on ONNX, so MetaSpore disassembled this part of preprocessing logic. NLP pretreatment Tokenizer has been integrated into the Query pretreatment service. MetaSpore dismantlement with a relatively general convention. Users only need to provide preprocessing logic files to realize the loading and prediction interface and export the necessary data and configuration files loaded into the preprocessing service. Subsequent CV preprocessing logic will also be integrated in this manner. The preprocessing service currently provides the gRPC interface invocation externally and is dependent on the Query preprocessing (QP) module in the retrieval algorithm service. After the user request reaches the retrieval algorithm service, it will be forwarded to the service to complete the data preprocessing and continue the subsequent processing. The ReadMe provides details on how the preprocessing service is started, how the preprocessing model exported offline to cloud storage enters the service, and how to debug the service. To further improve the efficiency and stability of model reasoning, MetaSpore Serving implements a Python preprocessing submodule. So MetaSpore can provide gRPC services through user-specified preprocessor.py, complete Tokenizer or CV-related preprocessing in NLP, and translate requests into a Tensor that deep models can handle. Finally, the model inference is carried out by MetaSpore, Serving subsequent sub-modules. Presented here on the lot code: https://github.com/meta-soul/MetaSpore/compare/add\python\preprocessor 1.2.2 Retrieval algorithm services Retrieval algorithm service is the core of the whole online service system, which is responsible for the triage of experiments, the assembly of algorithm chains such as preprocessing, recall, sorting, and the invocation of dependent component services. The whole retrieval algorithm service is developed based on the Java Spring framework and supports multi-mode retrieval scenarios of text search and text search graph. Due to good internal abstraction and modular design, it has high flexibility and can be migrated to similar application scenarios at a low cost. Here’s a quick guide to configuring the environment to set up the retrieval algorithm service. See ReadME for more details: Install dependent components. Use Maven to install the online-Serving component Search for service configurations. Copy the template configuration file and replace the MongoDB, Milvus, and other configurations based on the development/production environment. Install and configure Consul. Consul allows you to synchronize the search service configuration in real-time, including cutting the flow of experiments, recall parameters, and sorting parameters. The project’s configuration file shows the current configuration parameters of text search and text search. The parameter modelName in the stage of pretreatment and recall is the corresponding model exported in offline processing. Start the service. Once the above configuration is complete, the retrieval service can be started from the entry script. Once the service is started, you can test it! For example, for a user with userId=10 who wants to query “How to renew ID card,” access the text search service. 1.2.3 User Entry Service Considering that the retrieval algorithm service is in the form of the API interface, it is difficult to locate and trace the problem, especially for the text search image scene can intuitively display the retrieval results to facilitate the iterative optimization of the retrieval algorithm. This paper provides a lightweight Web UI interface for text search and image search, a search input box, and results in a display page for users. Developed by Flask, the service can be easily integrated with other retrieval applications. The service calls the retrieval algorithm service and displays the returned results on the page. It’s also easy to install and start the service. Once you’re done, go to http://127.0.0.1:8090 to see if the search UI service is working correctly. See the ReadME at the end of this article for details. Multimodal system demonstration The multimodal retrieval service can be started when offline processing and online service environment configuration have been completed following the above instructions. Examples of textual searches are shown below. Enter the entry of the text search map application, enter “cat” first, and you can see that the first three digits of the returned result are cats: https://preview.redd.it/0n5nuyvhez291.png?width=1280&format=png&auto=webp&s=1e9c054f541d53381674b8d6001b4bf524506bd2 If you add a color constraint to “cat” to retrieve “black cat,” you can see that it does return a black cat: https://preview.redd.it/rzc0qjyjez291.png?width=1280&format=png&auto=webp&s=d5bcc503ef0fb3360c7740e60e295cf372dcad47 Further, strengthen the constraint on the search term, change it to “black cat on the bed,” and return results containing pictures of a black cat climbing on the bed: ​ https://preview.redd.it/c4b2q8olez291.png?width=1280&format=png&auto=webp&s=4f3817b0b9f07e1e68d1d4a8281702ba3834a00a The cat can still be found through the text search system after the color and scene modification in the above example. Conclusion The cutting-edge pre-training technology can bridge the semantic gap between different modes, and the HuggingFace community can greatly reduce the cost for developers to use the pre-training model. Combined with the technological ecology of MetaSpore online reasoning and online microservices provided by DMetaSpore, the pre-training model is no longer mere offline dabbling. Instead, it can truly achieve end-to-end implementation from cutting-edge technology to industrial scenarios, fully releasing the dividends of the pre-training large model. In the future, DMetaSoul will continue to improve and optimize the MetaSpore technology ecosystem: More automated and wider access to HuggingFace community ecology. MetaSpore will soon release a common model rollout mechanism to make HuggingFace ecologically accessible and will later integrate preprocessing services into online services. Multi-mode retrieval offline algorithm optimization. For multimodal retrieval scenarios, MetaSpore will continuously iteratively optimize offline algorithm components, including text recall/sort model, graphic recall/sort model, etc., to improve the accuracy and efficiency of the retrieval algorithm. For related code and reference documentation in this article, please visit: https://github.com/meta-soul/MetaSpore/tree/main/demo/multimodal/online Some images source: https://github.com/openai/CLIP/raw/main/CLIP.png https://www.sbert.net/examples/training/sts/README.html

[P]MMML | Deploy HuggingFace training model rapidly based on MetaSpore
reddit
LLM Vibe Score0
Human Vibe Score1
qazmkoppThis week

[P]MMML | Deploy HuggingFace training model rapidly based on MetaSpore

A few days ago, HuggingFace announced a $100 million Series C funding round, which was big news in open source machine learning and could be a sign of where the industry is headed. Two days before the HuggingFace funding announcement, open-source machine learning platform MetaSpore released a demo based on the HuggingFace Rapid deployment pre-training model. As deep learning technology makes innovative breakthroughs in computer vision, natural language processing, speech understanding, and other fields, more and more unstructured data are perceived, understood, and processed by machines. These advances are mainly due to the powerful learning ability of deep learning. Through pre-training of deep models on massive data, the models can capture the internal data patterns, thus helping many downstream tasks. With the industry and academia investing more and more energy in the research of pre-training technology, the distribution warehouses of pre-training models such as HuggingFace and Timm have emerged one after another. The open-source community release pre-training significant model dividends at an unprecedented speed. In recent years, the data form of machine modeling and understanding has gradually evolved from single-mode to multi-mode, and the semantic gap between different modes is being eliminated, making it possible to retrieve data across modes. Take CLIP, OpenAI’s open-source work, as an example, to pre-train the twin towers of images and texts on a dataset of 400 million pictures and texts and connect the semantics between pictures and texts. Many researchers in the academic world have been solving multimodal problems such as image generation and retrieval based on this technology. Although the frontier technology through the semantic gap between modal data, there is still a heavy and complicated model tuning, offline data processing, high performance online reasoning architecture design, heterogeneous computing, and online algorithm be born multiple processes and challenges, hindering the frontier multimodal retrieval technologies fall to the ground and pratt &whitney. DMetaSoul aims at the above technical pain points, abstracting and uniting many links such as model training optimization, online reasoning, and algorithm experiment, forming a set of solutions that can quickly apply offline pre-training model to online. This paper will introduce how to use the HuggingFace community pre-training model to conduct online reasoning and algorithm experiments based on MetaSpore technology ecology so that the benefits of the pre-training model can be fully released to the specific business or industry and small and medium-sized enterprises. And we will give the text search text and text search graph two multimodal retrieval demonstration examples for your reference. Multimodal semantic retrieval The sample architecture of multimodal retrieval is as follows: Our multimodal retrieval system supports both text search and text search application scenarios, including offline processing, model reasoning, online services, and other core modules: ​ https://preview.redd.it/w4v4c7vcez291.png?width=1834&format=png&auto=webp&s=0687efb1fddb26e8e30cb844d398ec712b947f31 Offline processing, including offline data processing processes for different application scenarios of text search and text search, including model tuning, model export, data index database construction, data push, etc. Model inference. After the offline model training, we deployed our NLP and CV large models based on the MetaSpore Serving framework. MetaSpore Serving helps us conveniently perform online inference, elastic scheduling, load balancing, and resource scheduling in heterogeneous environments. Online services. Based on MetaSpore’s online algorithm application framework, MetaSpore has a complete set of reusable online search services, including Front-end retrieval UI, multimodal data preprocessing, vector recall and sorting algorithm, AB experimental framework, etc. MetaSpore also supports text search by text and image scene search by text and can be migrated to other application scenarios at a low cost. The HuggingFace open source community has provided several excellent baseline models for similar multimodal retrieval problems, which are often the starting point for actual optimization in the industry. MetaSpore also uses the pre-training model of the HuggingFace community in its online services of searching words by words and images by words. Searching words by words is based on the semantic similarity model of the question and answer field optimized by MetaSpore, and searching images by words is based on the community pre-training model. These community open source pre-training models are exported to the general ONNX format and loaded into MetaSpore Serving for online reasoning. The following sections will provide a detailed description of the model export and online retrieval algorithm services. The reasoning part of the model is standardized SAAS services with low coupling with the business. Interested readers can refer to my previous post: The design concept of MetaSpore, a new generation of the one-stop machine learning platform. 1.1 Offline Processing Offline processing mainly involves the export and loading of online models and index building and pushing of the document library. You can follow the step-by-step instructions below to complete the offline processing of text search and image search and see how the offline pre-training model achieves reasoning at MetaSpore. 1.1.1 Search text by text Traditional text retrieval systems are based on literal matching algorithms such as BM25. Due to users’ diverse query words, a semantic gap between query words and documents is often encountered. For example, users misspell “iPhone” as “Phone,” and search terms are incredibly long, such as “1 \~ 3 months old baby autumn small size bag pants”. Traditional text retrieval systems will use spelling correction, synonym expansion, search terms rewriting, and other means to alleviate the semantic gap but fundamentally fail to solve this problem. Only when the retrieval system fully understands users’ query terms and documents can it meet users’ retrieval demands at the semantic level. With the continuous progress of pre-training and representational learning technology, some commercial search engines continue to integrate semantic vector retrieval methods based on symbolic learning into the retrieval ecology. Semantic retrieval model This paper introduces a set of semantic vector retrieval applications. MetaSpore built a set of semantic retrieval systems based on encyclopedia question and answer data. MetaSpore adopted the Sentence-Bert model as the semantic vector representation model, which fine-tunes the twin tower BERT in supervised or unsupervised ways to make the model more suitable for retrieval tasks. The model structure is as follows: The query-Doc symmetric two-tower model is used in text search and question and answer retrieval. The vector representation of online Query and offline DOC share the same vector representation model, so it is necessary to ensure the consistency of the offline DOC library building model and online Query inference model. The case uses MetaSpore’s text representation model Sbert-Chinese-QMC-domain-V1, optimized in the open-source semantically similar data set. This model will express the question and answer data as a vector in offline database construction. The user query will be expressed as a vector by this model in online retrieval, ensuring that query-doc in the same semantic space, users’ semantic retrieval demands can be guaranteed by vector similarity metric calculation. Since the text presentation model does vector encoding for Query online, we need to export the model for use by the online service. Go to the q&A data library code directory and export the model concerning the documentation. In the script, Pytorch Tracing is used to export the model. The models are exported to the “./export “directory. The exported models are mainly ONNX models used for wired reasoning, Tokenizer, and related configuration files. The exported models are loaded into MetaSpore Serving by the online Serving system described below for model reasoning. Since the exported model will be copied to the cloud storage, you need to configure related variables in env.sh. \Build library based on text search \ The retrieval database is built on the million-level encyclopedia question and answer data set. According to the description document, you need to download the data and complete the database construction. The question and answer data will be coded as a vector by the offline model, and then the database construction data will be pushed to the service component. The whole process of database construction is described as follows: Preprocessing, converting the original data into a more general JSonline format for database construction; Build index, use the same model as online “sbert-Chinese-qmc-domain-v1” to index documents (one document object per line); Push inverted (vector) and forward (document field) data to each component server. The following is an example of the database data format. After offline database construction is completed, various data are pushed to corresponding service components, such as Milvus storing vector representation of documents and MongoDB storing summary information of documents. Online retrieval algorithm services will use these service components to obtain relevant data. 1.1.2 Search by text Text and images are easy for humans to relate semantically but difficult for machines. First of all, from the perspective of data form, the text is the discrete ID type of one-dimensional data based on words and words. At the same time, images are continuous two-dimensional or three-dimensional data. Secondly, the text is a subjective creation of human beings, and its expressive ability is vibrant, including various turning points, metaphors, and other expressions, while images are machine representations of the objective world. In short, bridging the semantic gap between text and image data is much more complex than searching text by text. The traditional text search image retrieval technology generally relies on the external text description data of the image or the nearest neighbor retrieval technology and carries out the retrieval through the image associated text, which in essence degrades the problem to text search. However, it will also face many issues, such as obtaining the associated text of pictures and whether the accuracy of text search by text is high enough. The depth model has gradually evolved from single-mode to multi-mode in recent years. Taking the open-source project of OpenAI, CLIP, as an example, train the model through the massive image and text data of the Internet and map the text and image data into the same semantic space, making it possible to implement the text and image search technology based on semantic vector. CLIP graphic model The text search pictures introduced in this paper are implemented based on semantic vector retrieval, and the CLIP pre-training model is used as the two-tower retrieval architecture. Because the CLIP model has trained the semantic alignment of the twin towers’ text and image side models on the massive graphic and text data, it is particularly suitable for the text search graph scene. Due to the different image and text data forms, the Query-Doc asymmetric twin towers model is used for text search image retrieval. The image-side model of the twin towers is used for offline database construction, and the text-side model is used for the online return. In the final online retrieval, the database data of the image side model will be searched after the text side model encodes Query, and the CLIP pre-training model guarantees the semantic correlation between images and texts. The model can draw the graphic pairs closer in vector space by pre-training on a large amount of visual data. Here we need to export the text-side model for online MetaSpore Serving inference. Since the retrieval scene is based on Chinese, the CLIP model supporting Chinese understanding is selected. The exported content includes the ONNX model used for online reasoning and Tokenizer, similar to the text search. MetaSpore Serving can load model reasoning through the exported content. Build library on Image search You need to download the Unsplash Lite library data and complete the construction according to the instructions. The whole process of database construction is described as follows: Preprocessing, specify the image directory, and then generate a more general JSOnline file for library construction; Build index, use OpenAI/Clip-Vit-BASE-Patch32 pre-training model to index the gallery, and output one document object for each line of index data; Push inverted (vector) and forward (document field) data to each component server. Like text search, after offline database construction, relevant data will be pushed to service components, called by online retrieval algorithm services to obtain relevant data. 1.2 Online Services The overall online service architecture diagram is as follows: https://preview.redd.it/jfsl8hdfez291.png?width=1280&format=png&auto=webp&s=a858e2304a0c93e78ba5429612ca08cbee69b35a Multi-mode search online service system supports application scenarios such as text search and text search. The whole online service consists of the following parts: Query preprocessing service: encapsulate preprocessing logic (including text/image, etc.) of pre-training model, and provide services through gRPC interface; Retrieval algorithm service: the whole algorithm processing link includes AB experiment tangent flow configuration, MetaSpore Serving call, vector recall, sorting, document summary, etc.; User entry service: provides a Web UI interface for users to debug and track down problems in the retrieval service. From a user request perspective, these services form invocation dependencies from back to front, so to build up a multimodal sample, you need to run each service from front to back first. Before doing this, remember to export the offline model, put it online and build the library first. This article will introduce the various parts of the online service system and make the whole service system step by step according to the following guidance. See the ReadME at the end of this article for more details. 1.2.1 Query preprocessing service Deep learning models tend to be based on tensors, but NLP/CV models often have a preprocessing part that translates raw text and images into tensors that deep learning models can accept. For example, NLP class models often have a pre-tokenizer to transform text data of string type into discrete tensor data. CV class models also have similar processing logic to complete the cropping, scaling, transformation, and other processing of input images through preprocessing. On the one hand, considering that this part of preprocessing logic is decoupled from tensor reasoning of the depth model, on the other hand, the reason of the depth model has an independent technical system based on ONNX, so MetaSpore disassembled this part of preprocessing logic. NLP pretreatment Tokenizer has been integrated into the Query pretreatment service. MetaSpore dismantlement with a relatively general convention. Users only need to provide preprocessing logic files to realize the loading and prediction interface and export the necessary data and configuration files loaded into the preprocessing service. Subsequent CV preprocessing logic will also be integrated in this manner. The preprocessing service currently provides the gRPC interface invocation externally and is dependent on the Query preprocessing (QP) module in the retrieval algorithm service. After the user request reaches the retrieval algorithm service, it will be forwarded to the service to complete the data preprocessing and continue the subsequent processing. The ReadMe provides details on how the preprocessing service is started, how the preprocessing model exported offline to cloud storage enters the service, and how to debug the service. To further improve the efficiency and stability of model reasoning, MetaSpore Serving implements a Python preprocessing submodule. So MetaSpore can provide gRPC services through user-specified preprocessor.py, complete Tokenizer or CV-related preprocessing in NLP, and translate requests into a Tensor that deep models can handle. Finally, the model inference is carried out by MetaSpore, Serving subsequent sub-modules. Presented here on the lot code: https://github.com/meta-soul/MetaSpore/compare/add\python\preprocessor 1.2.2 Retrieval algorithm services Retrieval algorithm service is the core of the whole online service system, which is responsible for the triage of experiments, the assembly of algorithm chains such as preprocessing, recall, sorting, and the invocation of dependent component services. The whole retrieval algorithm service is developed based on the Java Spring framework and supports multi-mode retrieval scenarios of text search and text search graph. Due to good internal abstraction and modular design, it has high flexibility and can be migrated to similar application scenarios at a low cost. Here’s a quick guide to configuring the environment to set up the retrieval algorithm service. See ReadME for more details: Install dependent components. Use Maven to install the online-Serving component Search for service configurations. Copy the template configuration file and replace the MongoDB, Milvus, and other configurations based on the development/production environment. Install and configure Consul. Consul allows you to synchronize the search service configuration in real-time, including cutting the flow of experiments, recall parameters, and sorting parameters. The project’s configuration file shows the current configuration parameters of text search and text search. The parameter modelName in the stage of pretreatment and recall is the corresponding model exported in offline processing. Start the service. Once the above configuration is complete, the retrieval service can be started from the entry script. Once the service is started, you can test it! For example, for a user with userId=10 who wants to query “How to renew ID card,” access the text search service. 1.2.3 User Entry Service Considering that the retrieval algorithm service is in the form of the API interface, it is difficult to locate and trace the problem, especially for the text search image scene can intuitively display the retrieval results to facilitate the iterative optimization of the retrieval algorithm. This paper provides a lightweight Web UI interface for text search and image search, a search input box, and results in a display page for users. Developed by Flask, the service can be easily integrated with other retrieval applications. The service calls the retrieval algorithm service and displays the returned results on the page. It’s also easy to install and start the service. Once you’re done, go to http://127.0.0.1:8090 to see if the search UI service is working correctly. See the ReadME at the end of this article for details. Multimodal system demonstration The multimodal retrieval service can be started when offline processing and online service environment configuration have been completed following the above instructions. Examples of textual searches are shown below. Enter the entry of the text search map application, enter “cat” first, and you can see that the first three digits of the returned result are cats: https://preview.redd.it/0n5nuyvhez291.png?width=1280&format=png&auto=webp&s=1e9c054f541d53381674b8d6001b4bf524506bd2 If you add a color constraint to “cat” to retrieve “black cat,” you can see that it does return a black cat: https://preview.redd.it/rzc0qjyjez291.png?width=1280&format=png&auto=webp&s=d5bcc503ef0fb3360c7740e60e295cf372dcad47 Further, strengthen the constraint on the search term, change it to “black cat on the bed,” and return results containing pictures of a black cat climbing on the bed: ​ https://preview.redd.it/c4b2q8olez291.png?width=1280&format=png&auto=webp&s=4f3817b0b9f07e1e68d1d4a8281702ba3834a00a The cat can still be found through the text search system after the color and scene modification in the above example. Conclusion The cutting-edge pre-training technology can bridge the semantic gap between different modes, and the HuggingFace community can greatly reduce the cost for developers to use the pre-training model. Combined with the technological ecology of MetaSpore online reasoning and online microservices provided by DMetaSpore, the pre-training model is no longer mere offline dabbling. Instead, it can truly achieve end-to-end implementation from cutting-edge technology to industrial scenarios, fully releasing the dividends of the pre-training large model. In the future, DMetaSoul will continue to improve and optimize the MetaSpore technology ecosystem: More automated and wider access to HuggingFace community ecology. MetaSpore will soon release a common model rollout mechanism to make HuggingFace ecologically accessible and will later integrate preprocessing services into online services. Multi-mode retrieval offline algorithm optimization. For multimodal retrieval scenarios, MetaSpore will continuously iteratively optimize offline algorithm components, including text recall/sort model, graphic recall/sort model, etc., to improve the accuracy and efficiency of the retrieval algorithm. For related code and reference documentation in this article, please visit: https://github.com/meta-soul/MetaSpore/tree/main/demo/multimodal/online Some images source: https://github.com/openai/CLIP/raw/main/CLIP.png https://www.sbert.net/examples/training/sts/README.html

[D] What is your honest experience with reinforcement learning?
reddit
LLM Vibe Score0
Human Vibe Score1
Starks-TechnologyThis week

[D] What is your honest experience with reinforcement learning?

In my personal experience, SOTA RL algorithms simply don't work. I've tried working with reinforcement learning for over 5 years. I remember when Alpha Go defeated the world famous Go player, Lee Sedol, and everybody thought RL would take the ML community by storm. Yet, outside of toy problems, I've personally never found a practical use-case of RL. What is your experience with it? Aside from Ad recommendation systems and RLHF, are there legitimate use-cases of RL? Or, was it all hype? Edit: I know a lot about AI. I built NexusTrade, an AI-Powered automated investing tool that lets non-technical users create, update, and deploy their trading strategies. I’m not an idiot nor a noob; RL is just ridiculously hard. Edit 2: Since my comments are being downvoted, here is a link to my article that better describes my position. It's not that I don't understand RL. I released my open-source code and wrote a paper on it. It's the fact that it's EXTREMELY difficult to understand. Other deep learning algorithms like CNNs (including ResNets), RNNs (including GRUs and LSTMs), Transformers, and GANs are not hard to understand. These algorithms work and have practical use-cases outside of the lab. Traditional SOTA RL algorithms like PPO, DDPG, and TD3 are just very hard. You need to do a bunch of research to even implement a toy problem. In contrast, the decision transformer is something anybody can implement, and it seems to match or surpass the SOTA. You don't need two networks battling each other. You don't have to go through hell to debug your network. It just naturally learns the best set of actions in an auto-regressive manner. I also didn't mean to come off as arrogant or imply that RL is not worth learning. I just haven't seen any real-world, practical use-cases of it. I simply wanted to start a discussion, not claim that I know everything. Edit 3: There's a shockingly number of people calling me an idiot for not fully understanding RL. You guys are wayyy too comfortable calling people you disagree with names. News-flash, not everybody has a PhD in ML. My undergraduate degree is in biology. I self-taught myself the high-level maths to understand ML. I'm very passionate about the field; I just have VERY disappointing experiences with RL. Funny enough, there are very few people refuting my actual points. To summarize: Lack of real-world applications Extremely complex and inaccessible to 99% of the population Much harder than traditional DL algorithms like CNNs, RNNs, and GANs Sample inefficiency and instability Difficult to debug Better alternatives, such as the Decision Transformer Are these not legitimate criticisms? Is the purpose of this sub not to have discussions related to Machine Learning? To the few commenters that aren't calling me an idiot...thank you! Remember, it costs you nothing to be nice! Edit 4: Lots of people seem to agree that RL is over-hyped. Unfortunately those comments are downvoted. To clear up some things: We've invested HEAVILY into reinforcement learning. All we got from this investment is a robot that can be super-human at (some) video games. AlphaFold did not use any reinforcement learning. SpaceX doesn't either. I concede that it can be useful for robotics, but still argue that it's use-cases outside the lab are extremely limited. If you're stumbling on this thread and curious about an RL alternative, check out the Decision Transformer. It can be used in any situation that a traditional RL algorithm can be used. Final Edit: To those who contributed more recently, thank you for the thoughtful discussion! From what I learned, model-based models like Dreamer and IRIS MIGHT have a future. But everybody who has actually used model-free models like DDPG unanimously agree that they suck and don’t work.

If only someone told me this before my first startup
reddit
LLM Vibe Score0
Human Vibe Score0.625
johnrushxThis week

If only someone told me this before my first startup

If only someone told me this before my first startup: Validate idea first. I wasted a decade building stuff nobody needed. Incubators and VCs served to me as a validation, but I was so wrong. Kill my EGO. It’s not about me, but the user. I must want what the user wants, not what I want. My taste isn't important. The user has expectations, and I must fulfill them. Don’t chaise investors. Chase users, and then investors will be chasing me. I've never had more incoming interest from VC than now when I'm the least interested in them. Never hire managers. Only hire doers until PMF. So many people know how to manage people and so few can actually get sh\*t done barehand. Landing page is the least important thing in a startup. Pick a simple template, edit texts with a no-code website builder in less than an hour and that's it! At the early stage, I win traffic outside of my website, people are already interested, so don't make them search for the signup button among the texts! Focus on conversion optimization only when the traffic is consistent. Keep it to one page. Nobody gonna browse this website. Hire only fullstack devs. There is nothing less productive in this world than a team of developers for an early-stage product. One full stack dev building the whole product. That’s it. Chase global market from day 1. If the product and marketing are good, it will work on the global market too, if it’s bad, it won’t work on the local market too. So better go global from day 1, so that if it works, the upside is 100x bigger. I launched all startups for the Norwegian market, hoping we will scale to international at some point. I wish I launched to international from day 1 as I do now. The size of the market is 10000x bigger. I can validate and grow products in days, not in years as it used to be. Do SEO from day 2. As early as I can. I ignored this for 14 years. It’s my biggest regret. It takes just 5 minutes to get it done on my landing page. I go to Google Keyword Planner, enter a few keywords around my product, sort them by traffic, filter out high competition kws, pick the top 10, and place them natively on my home page and meta tags. Add one blog article every week. Either manually or by paying for an AI blogging tool. Sell features, before building them. Ask existing users if they want this feature. I run DMs with 10-20 users every day, where I chat about all my ideas and features I wanna add. I clearly see what resonates with me most and only go build those. If I don't have followers, try HN, Reddit, or just search on X for posts and ask it in the replies. People are helpful, they will reply if the question is easy to understand. Hire only people I would wanna hug. My cofounder, an old Danish man said this to me in 2015. And it was a big shift. I realized that if I don’t wanna hug the person, it means I dislike them on a chemical/animal level. Even if I can’t say why, but that’s the fact. Sooner or later, we would have a conflict and eventually break up. It takes up to 10 years to build a startup, make sure I do it with people I have this connection with. Invest all money into my startups and friends. Not crypt0, not stockmarket, not properties. I did some math, if I kept investing all my money into all my friends’ startups, that would be about 70 investments. 3 of them turned into unicorns eventually. Even 1 would have made the bank. Since 2022, I have invested all my money into my products, friends, and network. If I don't have friends who do startups, invest it in myself. Post on Twitter daily. I started posting here in March last year. It’s my primary source of new connections and growth. I could have started it earlier, I don't know why I didn't. Don’t work/partner with corporates. Corporations always seem like an amazing opportunity. They’re big and rich, they promise huge stuff, millions of users, etc. But every single time none of this happens. Because I talk to a regular employees there. They waste my time, destroy focus, shift priorities, and eventually bring in no users/money. Don’t get ever distracted by hype e.g. crypt0. I lost 1.5 years of my life this way. I met the worst people along the way. Fricks, scammers, thieves. Some of my close friends turned into thieves along the way, just because it was so common in that space. I wish this didn’t happen to me. I wish I was stronger and stayed on my mission. Don’t build consumer apps. Only b2b. Consumer apps are so hard, like a lottery. It’s just 0.00001% who make it big. The rest don’t. Even if I got many users, then there is a monetization challenge. I’ve spent 4 years in consumer apps and regret it. Don’t hold on bad project for too long, max 1 year. Some projects just don’t work. In most cases, it’s either the idea that’s so wrong that I can’t even pivot it or it’s a team that is good one by one but can’t make it as a team. Don’t drag this out for years. Tech conferences are a waste of time. They cost money, take energy, and time and I never really meet anyone there. Most people there are the “good” employees of corporations who were sent there as a perk for being loyal to the corporation. Very few fellow makers. Scrum is a Scam. For small teams and bootstrapped teams. If I had a team that had to be nagged every morning with questions as if they were children in kindergarten, then things would eventually fail. The only good stuff I managed to do happened with people who were grownups and could manage their stuff on their own. We would just do everything over chat as a sync on goals and plans. Outsource nothing at all until PMF. In a startup, almost everything needs to be done in a slightly different way, more creative, and more integrated into the vision. When outsourcing, the external members get no love and no case for the product. It’s just yet another assignment in their boring job. Instead of coming up with great ideas for my project they will be just focusing on ramping up their skills to get a promotion or a better job offer. Bootstrap. I spent way too much time raising money. I raised more than 10 times, preseed, seeded, and series A. But each time it was a 3-9 month project, meetings every week, and lots of destruction. I could afford to bootstrap, but I still went the VC-funded way, I don’t know why. To be honest, I didn’t know bootstrapping was a thing I could do or anyone does. It may take a decade. When I was 20, I was convinced it takes a few years to build and succeed with a startup. So I kept pushing my plans forward, to do it once I exited. Family, kids. I wish I married earlier. I wish I had kids earlier. No Free Tier. I'd launch a tool with a free tier, and it'd get sign-ups, but very few would convert. I'd treat free sign-ups as KPIs and run on it for years. I'd brag about signups and visitors. I'd even raise VC money with these stats. But eventually, I would fail to reach PMF. Because my main feedback would come from free users and the product turned into a perfect free product. Once I switched to "paid only" until I validated the product, things went really well. Free and paid users often need different products. Don't fall into this trap as I did. Being To Cheap. I always started by checking all competitors and setting the lowest price. I thought this would be one of the key advantages of my product. But no, I was wrong. The audience on $5 and $50 are totally different. $5: pain in the \*ss, never happy, never recommend me to a friend, leave in 4 months. $50: polite, give genuine feedback, happy, share with friends, become my big fan if I solve their request. I will fail. When I started my first startup. I thought if I did everything right, it would work out. But it turned out that almost every startup fails. I wish I knew that and I tried to fail faster, to get to the second iteration, then to the third, and keep going on, until I either find out nothing works or make it work. Use boilerplates. I wasted years of dev time and millions of VC money to pay for basic things. To build yet another sidebar, yet another dashboard, and payment integration... I had too much pride, I couldn't see myself taking someone else code as a basis for my product. I wanted it to be 100% mine, original, from scratch. Because my product seems special to me. Spend more time with Family & Friends. I missed the weddings of all my best friends and family. I was so busy. I thought if I didn't do it on time, the world would end. Looking back today, it was so wrong. I meet my friends and can't share those memories with them, which makes me very sad. I realized now, that spending 10% of my time with family and friends would practically make no negative impact on my startups. Build Products For Audiences I Love. I never thought of this. I'd often build products either for corporates, consumers, or for developers. It turns out I have no love for all 3. But I deeply love indie founders. Because they are risk-takers and partly kids in their hearts. Once I switched the focus to indie makers on my products, my level of joy increased by 100x for me. Ignore Badges and Awards I was chasing those awards just like everyone else. Going to ceremonies, signing up for events and stuff. I've won tons of awards, but none of those were eventually useful to my business. I better focused on my business and users. Write Every Single Day. When I was a kid, I loved writing stories. In school, they would give an assignment, and I'd often write a long story for it, however, the teacher would put an F on it. The reason was simple, I had an issue with the direction of the letters and the sequence of letters in the words. I still have it, it's just the Grammarly app helping me to correct these issues. So the teacher would fail my stories because almost every sentence had a spelling mistake that I couldn't even see. It made me think I'm made at writing. So I stopped, for 15 years. But I kept telling stories all these years. Recently I realized that in any group, the setup ends up turning into me telling stories to everyone. So I tried it all again, here on X 10 months ago. I love it, the process, the feedback from people. I write every day. I wish I had done it all these years. The End. \ this is an updated version of my post on the same topic from 2 months ago. I've edited some of the points and added 9 new ones.* \\ This is not advice, it's my self-reflection that might help you avoid same mistakes if you think those were mistakes

Raised $450k for my startup, here are the lessons I've learned along the way
reddit
LLM Vibe Score0
Human Vibe Score1
marin_smiljanicThis week

Raised $450k for my startup, here are the lessons I've learned along the way

2021 has been a pretty amazing year for Omnisearch. Having started initial work on Omnisearch at the end of 2020, we entered the new year with a working MVP yet no revenue, no significant partnerships, and no funding. Fast forward to the end of 2021, and we now have fantastic revenue growth, a partnership with a public company, and a far more powerful, complete and polished product. But one milestone really changed Omnisearch’s trajectory: our $450,000 USD pre-seed round by GoAhead Ventures. In this post I want to share the story of how it came about and offer a couple of takeaways to keep in mind when preparing for fundraising. ​ The story Contrary to most advice, my co-founder Matej and I didn’t allocate a specific time to switch to “fundraising mode” but rather talked to investors on an ongoing basis. It was a bit of a distraction from working on the product, but on the positive side we were able to constantly get feedback on the idea, pitch, go-to-market strategy and hiring, as well as hearing investors’ major concerns sooner rather than later. That being said, our six-month long fundraising efforts weren’t yielding results - we talked to about twenty investors, mostly angels or smaller funds, with no success. The feedback was generally of the “too early for us” variety (since we were still pre-revenue), with additional questions about our go-to-market strategy and ideal customer persona. The introduction to our eventual investors, California-based GoAhead Ventures, came through a friend who had pitched them previously. We wrote a simple blurb and sent our pitch deck. We then went through GoAhead’s hyper-efficient screening process, consisting of a 30-minute call, a recorded three-minute pitch, and filling out a simple Google doc. Throughout the whole process, the GoAhead team left an awesome impression thanks to their knowledge of enterprise software and their responsiveness. They ended up investing and the whole deal was closed within two weeks, which is super fast even by Silicon Valley standards. While our fundraising experience is a single data point and your case might be different, here are the key takeaways from our journey. ​ Perseverance wins: Like I said above, we talked to about twenty investors before we closed our round. Getting a series of “no”s sucks, but we took the feedback seriously and tried to prepare better for questions that caught us off guard. But we persevered, keeping in mind that from a bird’s eye perspective it’s an amazing time to be building startups and raising funds. Focus on traction: Sounds pretty obvious, right? The truth is, though, that even a small amount of revenue is infinitely better than none at all. One of the major differences between our eventual successful investor pitch and the earlier ones was that we had actual paying customers, though our MRR was low. This allows you to talk about customers in the present tense, showing there’s actual demand for your product and making the use cases more tangible. And ideally, highlight a couple of customer testimonials to boost your credibility. Have a demo ready: In Omnisearch’s case, the demo was oftentimes the best received part of the pitch or call. We’d show investors the live demo, and for bonus points even asked them to choose a video from YouTube and then try searching through it. This always had a “wow” effect on prospective investors and made the subsequent conversation more exciting and positive. Accelerators: Accelerators like Y Combinator or Techstars can add enormous value to a startup, especially in the early stages. And while it’s a great idea to apply, don’t rely on them too heavily. Applications happen only a few times a year, and you should have a foolproof fundraising plan in case you don’t get in. In our case, we just constantly looked for investors who were interested in our space (defined as enterprise SaaS more broadly), using LinkedIn, AngelList, and intros from our own network. Practice the pitch ad nauseam: Pitching is tough to get right even for seasoned pros, so it pays to practice as often as possible. We took every opportunity to perfect the pitch: attending meetups and giving the thirty-second elevator pitch to other attendees over beer and pizza, participating in startup competitions, going to conferences and exhibiting at our own booth, attending pre-accelerator programs, and pitching to friends who are in the startup world. Show an understanding of the competition: Frankly, this was one of the strongest parts of our pitch and investor conversations. If you’re in a similar space to ours, Gartner Magic Quadrants and Forrester Waves are an awesome resource, as well as sites like AlternativeTo or Capterra and G2. By thoroughly studying these resources we gained a great understanding of the industry landscape and were able to articulate our differentiation more clearly and succinctly. Presenting this visually in a coordinate system or a feature grid is, from our experience, even more effective. Remember it’s just the beginning! Getting your first round of funding is just the beginning of the journey, so it’s important to avoid euphoria and get back to building and selling the product as soon as possible. While securing funding enables you to scale the team, and is a particular relief if the founders had worked without a salary, the end goal is still to build a big, profitable, and overall awesome startup.

Follow Along as I Flip this Website - Case Study
reddit
LLM Vibe Score0
Human Vibe Score1
jshogren10This week

Follow Along as I Flip this Website - Case Study

I am starting a new case study where I will be documenting my attempt to flip a website that I just purchased from Flippa. However, unlike most case studies where people hide certain parts and details from the public I will instead be sharing everything. That means you will know the exact URL of the site that I purchased and I will share everything with you all as I progress.I know that case studies are lot more interesting and you can learn better when you can see real examples of what I am talking about. Enough of the chatting, let's jump straight into this new case study and I will explain what this is all about. Before you get into the case study I want to give you the option of reading this one my website where all of the images can be seen within the post and it is easier to read. I also want to say that I have nothing to sell you or anything close to it. So if you want to read it there you can do so here ##Introductory Video I have put together a video that talks about many of the things that I cover in this article. So if you would rather watch a video you can watch that here - https://www.youtube.com/watch?v=EE3SxtNnqts However, I go into more detail in the actual article FYI. Also, I plan on using Youtube very frequently in this case study so be on the lookout for new videos.There is going to be a video that will accompany every single case study post because I like having it being presented in two different mediums. ##The Website I Just Bought Around a week ago I made a new website purchase from Flippa and you can view the website's Flippa listing here - https://flippa.com/6439965-hvactraining101-com Screenshot of the Homepage - http://imgur.com/T6Iv1QN I paid $1,250 for the site and you will soon see that I got a really good deal. As you might be able to tell from the URL, this site is focused around training and education for becoming a HVAC technician. This is a lucrative niche to be in and Adsense pays very well. I do not have control of the site yet due to the transfer process not being completed. However, I am hoping within a few days everything will be finalized and I will take full control of the site. In the meantime, I figured it would be a good time to put together the introduction post for this new case study! ##Why I Bought this Website Now that you have a general idea of the website that I purchased, I now want to explain the reasoning behind the purchase. There are 3 major reasons for this purchase and I will explain each one of them below. GREAT Price As I mentioned earlier, I bought this website for $1,250. However, that doesn't mean a whole lot unless you know how much the site is making each month. Screenshot of the earnings for the last 12 months - http://imgur.com/NptxCHy Average Monthly Profits: 3 Month = $126 6 Month = $128 12 Month = $229.50 Let's use the 6 month average of $128/month as our baseline average. Since it is making on average $128/month and it was sold for $1,250 then that means I bought this site at a multiple of 9.76x! Most sites in today's market go for 20x-30x multiples. As you can see, I got a great deal on this site. Although the great price was the biggest reason for me buying this site there are other factors that persuaded me as well. You need to remember that just because you can get a website for a good price it doesn't mean it is a good deal. There are other factors that you need to look at as well. Extremely Under Optimized This site is currently being monetized mainly by Adsense and a very small amount from Quinstreet. From my experience with testing and optimizing Adsense layouts for my site in my Website Investing case study I know the common ad layouts that work best for maximizing Adsense revenue. With that being said, I can quickly determine if a website is being under optimized in terms of the ad layout. One of the first things I did when analyzing this site was examine the ad layout it was using. Screenshot of the website with the ad layout the previous owner was using - http://imgur.com/wqleLVA There is only ONE ad per page being used, that's it. Google allows up to 6 total ads to be used per page and you can imagine how much money is being left on the table because of this. I am estimating that I can probably double the earnings for the site practically overnight once I add more ads to the site. Adding more ads in combination with my favorite Adsense plugin, AmpedSense, I will be able to easily boost the earnings for this site quickly. It is also worth mentioning how lucrative this niche is and how much advertisers are willing to spend on a per click basis. The average CPC for the top keywords this site is currently ranking for in Google - http://imgur.com/ifxiy8B Look at those average CPC numbers, they are insanely high! I could be making up to $25 per click for some of those keywords, which is so absurd to me. Combine these extremely high CPC with the fact that the site currently only has one ad per page and you can start to understand just how under optimized this site truly is. I also plan on utilizing other ad networks such as Quinstreet and Campus Explorer more as well. These two networks are targeted at the education niche which works very well with my site. I will be testing to see if these convert better than normal Adsense ads. Goldmine of Untapped Keywords One of the biggest opportunities I see for growing this site is to target local keywords related to HVAC training. As of right now, the site has only scratched the surface when it comes to trying to rank for state/city keywords. Currently there are only two pages on the entire website which go after local keywords, those two pages target Texas and Florida HVAC search terms. These two pages are two of the more popular pages in terms of total amount of traffic. See the screenshot of the Google Analytics - http://imgur.com/NB0xJ4G Two out of the top five most popular pages for the entire website are focused on local search terms. However, these are the ONLY two pages that target local search terms on the whole site! There are 48 other states, although there may not be search volume for all states, and countless cities that are not being targeted. Why do I think this is such a good opportunity? For a few reasons: Local keywords are a lot easier to rank for in Google than more general keywords This site has been able to rank for two states successfully already and it proves it is possible Traffic going to these local pages is WAY more targeted and will convert at a much higher rate, which means more commissions for me There are so many more states and cities that get a good amount of searches that I can target To give you an idea of the type of keywords these local pages rank for, you can see the top keywords that the Florida page is ranking for in Google: Top ranking keywords for the Florida page - http://imgur.com/j7uKzl2 As you can see these keywords don't get a ton of searches each month, but ranking 1st for a keyword getting 90 searches a month is better than being ranked 10th for a keyword getting 1,000 searches a month. I have started to do some keyword research for other states and I am liking what I am finding so far. Keywords that I have found which I will be targeting with future articles - http://imgur.com/8CCCCWU I will go into more detail about my keyword research in future articles, but I wanted to give you an idea of what my strategy will be! I also wanted to share why I am super excited about the future potential to grow this site by targeting local keywords. ##Risks Yes, there are many good things about this website, but there are always risks involved no matter what the investment is. The same thing goes for this site. Below are some of the risks that I currently see. HTML Site This website is a HTML site and I will need to transfer it to Wordpress ASAP. I have been doing some research on this process and it shouldn't be too hard to get this over to Wordpress. In doing so it will make adding content, managing the back end and just about everything else easier. Also, I am hoping that when I transfer it to Wordpress that it will become more optimized for Google which will increase keyword rankings. Declining Earnings Looking at the last 12 months of earnings you will notice a drop off from last year till now. Earnings from the last 12 months - http://imgur.com/WsotZsj In May of 2015 it looks like the site earned right around $500, which is much higher than the $128 that it is earning now. However, the last 7 or so months have been consistent which is a good sign. Even though the earnings are much lower now then they were a year ago it is good to know that this site has the potential to earn $500/month because it has done it before. Slightly Declining Traffic In the last 12 months the site's traffic has declined, however, it looks like it is picking back up. Traffic from the last 12 months - http://imgur.com/aiYZW9W The decline is nothing serious, but there is a drop on traffic. Let's take a look at the complete history of this site's traffic so we can get a better idea of what is going on here: Complete traffic history - http://imgur.com/tYmboVn The above screenshot is from 2012 all the way up to right now. In the grand scheme of things you can see that the traffic is still doing well and it looks like it is on the upswing now. Those three risks mentioned above are the three biggest risks with this site at this point. It is always good to note the risks and do everything you can to prevent them from causing a problem. ##My Growth Strategy Whenever I purchase a new site I always create an outline or plan on how I will grow the site. Right now, I have some basic ideas on how I will grow this site, but as I go on I will continue to change and optimize my strategies to be more effective. Below I have outlined my current plans to grow: Add more Adsense Ads The very first thing I will do once I get control of the site is add more ads per page. I am predicting that by just adding a few more ads per page I will be able to more than likely double the earnings. I will touch on exactly how I will be optimizing the ad layouts in future posts. Test other Ad Networks I will be doing a lot of testing and experimenting when it comes to the ad networks. I plan on trying out Adsense, Media.net, Quinstreet, Campus Explorer and finding the combination of those 4 which produces the most revenue. The Adsense and Media.net ads will perform well on the more general pages while Quinstreet and Campus Explorer ads will be geared towards the local search terms. There will probably be other ad networks I will try out but these are the four which I will be using right away. If you are aware of any other ad networks out there which are geared towards the education niche please let me know in the comments below! Target Local Keywords with new Content I have already touched on this, but I will starting to produce content targeting these local keywords ASAP. The sooner I add the content to the site the sooner it will start to rank and bring in traffic. I will not be writing my own content and instead I will be outsourcing all of it via Upwork. I will show you all how I go about outsourcing content production and you can see my process for doing that. ##Goals for this Website My goal for the website is to have it valued at $10,000+ within 12 months. Let's break down this larger goal into smaller chunks which will make achieving it easier and more attainable. Earnings - $500/month To get the site valued at $10,000 the site will need to be making $500/month using a 20x monthly multiple. Right now, the site is making around $130/month so it has a ways to before it reaches the $500 a month mark. However, after doing some Adsense optimization I think we could push the earnings to around $300/month without much work. From there, it will come down to trying to bring in more traffic! Traffic - 5,000 Visitors per Month Why 5,000 visitors? Because that is how much traffic it is going to take to get to the $500/month goal. Let me explain how I came to this conclusion: The average RPM for this site is currently $50, which means for every 1,000 page views the site earns $50. After I optimize the Adsense layout for the site and add more ads per page I think I will be able to double the RPM to $100. Using the RPM of $100 the site will need to have 5,000 monthly visitors to earn $500. So 5,000 monthly visitors is the traffic goal I have set and aiming for! The site is currently getting around 3,000 visitors per month so I will need to add an extra 2,000 visitors to get to this goal. ##Want to Follow this Case Study? I will be using Youtube a lot in this case study so make sure to follow my Youtube channel here - www.youtube.com/c/joshshogren Other than that, I think that is going to bring us to the end of the introductory post for this new case study. I hope that you enjoyed reading and that you are excited to follow along! If you have any suggestions to make this case study better PLEASE let me know in the comment below. I want to make this case study the best one I have done yet. Talk to you all in the comment section.

I've been building stuff for years, I can build your idea
reddit
LLM Vibe Score0
Human Vibe Score1
Minute-Line2712This week

I've been building stuff for years, I can build your idea

If anyone is interested at all, I'm a college student whose been massively passionate for entrepreneur and business stuff for years and can build anything from a marketplace to a social network, a booking app, a live streaming app, AI app*, what not. I'm really flexible in general and I'm also very reasonable with pricing. I like this community so I'd like to work on something I enjoy... (indeed... get out of my tabs please). Im a little passionate in general let's say, and I build all the time for fun. But I need money. So I'm here. Viola! Since I'm a college student, and this is my first time ever offering my services as well, chances are I won't expect anything past $500. I might even do it for cheaper if it looks doable for me. If you're generous that's up to you (lol). I'm happy to share my work and live previews you can interact with before we start. And, I'm happy to build things before you pay, so you can see it literally working :) THE "NEGATIVE"/catch: There will be no code from scratch as we'd build using no-code (www.bubble.io) and implement code as needed. This means you will not have the source code ever (even if we wanted to) and if you ever want your own from scratch platform/app you will have to find someone/a team to do it from scratch, as there's no option to export source code out. If you plan to grow past 1 million users, you may consider migrating to something built from scratch at this point scaling wise (and you can't get your source code - so you'll have to pay an entire dev team separately. If you have an app, you'll need a website developed, and Android and iOS developers for the app). For an MVP however, I personally think it's a faster, easier and cheaper way to get things running without investing a lot. THE GOOD: No-code/low-code will be a thousand times more easy, cheap, and fast to maintain. And if you're a startup, chances are you WILL need to tweak things... possibly a lot (100s vs. 1,000s difference in my opinion..). We can still build almost anything and also turn it into a mobile app for iOS / Android, though I'm more comfortable with getting a web app up first and the main posting here. But it depends on the complexity so just ask. Maintenance is likely to end up FAR cheaper and you can even do it free yourself if you decided to learn (which you could totally do in some days, and master in some weeks/months fully). I can build and document everything to be as beginner friendly as possible for you to be able to maintain it yourself if you wanted to learn. Send me a message describing what you need if you're interested. I will give you an estimate on price, time, and can send you some live previews. If we agree, I will start to build before any down payments to a (reasonable) point!:)

Started a content marketing agency 8 years ago - $0 to $7,863,052 (2025 update)
reddit
LLM Vibe Score0
Human Vibe Score0.882
mr_t_forhireThis week

Started a content marketing agency 8 years ago - $0 to $7,863,052 (2025 update)

Hey friends, My name is Tyler and for the past 8 years, I’ve been documenting my experience building a content marketing agency called Optimist. Year 1 — 0 to $500k ARR Year 2 — $500k to $1MM ARR Year 3 — $1MM ARR to $1.5MM(ish) ARR Year 4 — $3,333,686 Revenue Year 5 — $4,539,659 Revenue Year 6 — $5,974,324 Revenue Year 7 - $6,815,503 Revenue (Edit: Seems like links are banned now. You can check my post history for all of my previous updates with lessons and learnings.) How Optimist Works First, an overview/recap of the Optimist business model: We operate as a “collective” of full time/professional freelancers Everyone aside from me is a contractor Entirely remote/distributed team We pay freelancers a flat fee for most work, working out to roughly $65-100/hour. Clients pay us a flat monthly fee for full-service content marketing (research, strategy, writing, editing, design/photography, reporting and analytics, targeted linkbuilding, and more)\ Packages range in price from \~$10-20k/mo \This is something we are revisiting now* The Financials In 2024, we posted $1,032,035.34 in revenue. This brings our lifetime revenue to $7,863,052. Here’s our monthly revenue from January 2017 to December of 2024. (Edit: Seems like I'm not allowed to link to the chart.) The good news: Revenue is up 23% YoY. EBITDA in Q4 trending up 1-2 points. We hosted our first retreat in 4 years, going to Ireland with about half the team. The bad news: Our revenue is still historically low. At $1MM for the year, we’re down about 33% from our previous years over $1.5MM. Revenue has been rocky. It doesn’t feel like we’ve really “recovered” from the bumps last year. The trend doesn’t really look great. Even though, anecdotally, it feels like we are moving in a good direction. EBITDA is still hovering at around 7%. Would love to get that closer to 20%. (For those who may ask: I’m calculating EBITDA after paying taxes and W2 portion of my income.) — Almost every year, my update starts the same way: This has been a year of growth and change. Both for my business—and me personally. 2024 was no different. I guess that tells you something about entrepreneurship. It’s a lot more like sailing a ship than driving a car. You’re constantly adapting, tides are shifting, and any blip of calm is usually just a moment before the next storm. As with past years, there’s a lot to unpack from the last 12 months. Here we go again. Everything is Burning In the last 2 years, everything has turned upside down in the world of content and SEO. Back in 2020, we made a big decision to re-position the agency. (See post history) We decided to narrow our focus to our most successful, profitable, and consistent segment of clients and re-work our entire operation to focus on serving them. We defined our ICP as: \~Series A ($10mm+ funding) with 6-12 months runway to scale organic as a channel Product-led company with “simple” sales cycle involving fewer stakeholders Demonstrable opportunity to use SEO to drive business growth Our services: Content focused on growing organic search (SEO) Full-service engagements that included research, planning, writing, design, reporting And our engagement structure: Engaged directly with an executive; ownership over strategy and day-to-day execution 1-2 points of contact or stakeholders Strategic partner that drives business growth (not a service vendor who makes content) Most importantly, we decided that we were no longer going to offer a broader range of content that we used to sell. That included everything from thought leadership content to case studies and ebooks. We doubled-down on “SEO content” for product-led SaaS companies. And this worked phenomenally for us. We started bringing on more clients than ever. We developed a lot of internal system and processes that helped us scale and take on more work than we’ve ever had and drive great outcomes for our ideal clients. But in 2023 and 2024, things started going awry. One big change, of course, was the rise of AI. Many companies and executives (and writers) feel that AI can write content just as well as an agency like ours. That made it a lot harder to sell a $10,000 per month engagement when they feel like the bulk of the work could be “done for free.” (Lots of thoughts on this if you want my opinions.) But it wasn’t just that. Google also started tinkering with their algorithm, introducing new features like AI Overviews, and generally changing the rules of the game. This created 3 big shifts in our world: The perceived value of content (especially “SEO content”) dropped dramatically in many people’s minds because of AI’s writing capabilities SEO became less predictable as a source of traffic and revenue It’s harder than ever for startups and smaller companies to rank for valuable keywords (let alone generate any meaningful traffic or revenue from them) The effect? The middle of the content market has hollowed out. People—like us—providing good, human-crafted content aimed on driving SEO growth saw a dramatic decline in demand. We felt it all year. Fewer and fewer leads. The leads we did see usually scoffed at our prices. They were indexing us against the cost of content mills and mass-produced AI articles. It was a time of soul-searching and looking for a way forward. I spent the first half of the year convinced that the only way to survive was to run toward the fire. We have to build our own AI workflows. We have to cut our rates internally. We have to get faster and cheaper to stay competitive with the agencies offering the same number of deliverables for a fraction of our rates. It’s the only way forward. But then I asked myself a question… Is this the game I actually want to play? As an entrepreneur, do I want to run a business where I’m competing mostly on price and efficiency rather than quality and value? Do I want to hop into a race toward cheaper and cheaper content? Do I want to help people chase a dwindling amount of organic traffic that’s shrinking in value? No. That’s not the game I want to play. That’s not a business I want to run. I don’t want to be in the content mill business. So I decided to turn the wheel—again. Repositioning Part II: Electric Boogaloo What do you do when the whole world shifts around you and the things that used to work aren’t working anymore? You pivot. You re-position the business and move in another direction. So that’s what we decided to do. Again. There was only one problem: I honestly wasn’t sure what opportunities existed in the content marketing industry outside of what we were already doing. We lived in a little echo chamber of startups and SEO. It felt like the whole market was on fire and I had fight through the smoke to find an escape hatch. So I started making calls. Good ol’ fashioned market research. I reached out to a few dozen marketing and content leaders at a bunch of different companies. I got on the phone and just asked lots of questions about their content programs, their goals, and their pain points. I wanted to understand what was happening in the market and how we could be valuable. And, luckily, this process really paid off. I learned a lot about the fragmentation happening across content and how views were shifting. I noticed key trends and how our old target market really wasn’t buying what we were selling. Startups and small companies are no longer willing to invest in an agency like ours. If they were doing content and SEO at all, they were focused entirely on using AI to scale output and minimize costs. VC money is still scarce and venture-backed companies are more focused on profitability than pure growth and raising another round. Larger companies (\~500+ employees) are doing more content than ever and drowning in content production. They want to focus on strategy but can barely tread water keeping up with content requests from sales, demand gen, the CEO, and everyone else. Many of the companies still investing in content are looking at channels and formats outside of SEO. Things like thought leadership, data reports, interview-driven content, and more. They see it as a way to stand out from the crowd of “bland SEO content.” Content needs are constantly in flux. They range from data reports and blog posts to product one-pagers. The idea of a fixed-scope retainer is a total mismatch for the needs of most companies. All of this led to the logical conclusion: We were talking to the wrong people about the wrong things\.\ Many companies came to one of two logical conclusions: SEO is a risky bet, so it’s gotta be a moonshot—super-low cost with a possibility for a big upside (i.e., use AI to crank out lots of content. If it works, great. If it doesn’t, then at least we aren’t out much money.) SEO is a risky bet, so we should diversify into other strategies and channels to drive growth (i.e., shift our budget from SEO and keyword-focused content to video, podcasts, thought leadership, social, etc) Unless we were going to lean into AI and dramatically cut our costs and rates, our old buyers weren’t interested. And the segment of the market that needs our help most are looking primarily for production support across a big range of content types. They’re not looking for a team to run a full-blown program focused entirely on SEO. So we had to go back to the drawing board. I’ve written before about our basic approach to repositioning the business. But, ultimately it comes down to identifying our unique strengths as a team and then connecting them to needs in the market. After reviewing the insights from my discussions and taking another hard look at our business and our strengths, I decided on a new direction: Move upmarket: Serve mid-size to enterprise businesses with \~500-5,000 employees instead of startups Focus on content that supports a broader range of business goals instead of solely on SEO and organic growth (e.g., sales, demand gen, brand, etc) Shift back to our broader playbook of content deliverables, including thought leadership, data studies, and more Focus on content execution and production to support an internally-directed content strategy across multiple functions In a way, it’s sort of a reverse-niche move. Rather than zooming in specifically on driving organic growth for startups, we want to be more of an end-to-end content production partner that solves issues of execution and operations for all kinds of content teams. It’s early days, but the response here has been promising. We’ve seen an uptick in leads through Q4. And more companies in our pipeline fit the new ICP. They’re bigger, often have more budget. (But they move more slowly). We should know by the end of the quarter if this maneuver is truly paying off. Hopefully, this will work out. Hopefully our research and strategy are right and we’ll find a soft landing serving a different type of client. If it doesn’t? Then it will be time to make some harder decisions. As I already mentioned, I’m not interested in the race to the bottom of AI content. And if that’s the only game left in town, then it might be time to think hard about a much bigger change. — To be done: Build new content playbooks for expanded deliverables Build new showcase page for expanded deliverables Retooling the Operation It’s easy to say we’re doing something new. It’s a lot harder to actually do it—and do it well. Beyond just changing our positioning, we have to do open-heart surgery on the entire content operation behind the scenes. We need to create new systems that work for a broader range of content types, formats, and goals. Here’s the first rub: All of our workflows are tooled specifically for SEO-focused content. Every template, worksheet, and process that we’ve built and scaled in the last 5 years assumes that the primary goal of every piece of content is SEO. Even something as simple as requiring a target keyword is a blocker in a world where we’re not entirely focused on SEO. This is relatively easy to fix, but it requires several key changes: Update content calendars to make keywords optional Update workflows to determine whether we need an optimization report for each deliverable Next, we need to break down the deliverables into parts rather than a single line item. In our old system, we would plan content as a single row in a Content Calendar spreadsheet. It was a really wide sheet with lots of fields where we’d define the dimensions of each individual article. This was very efficient and simple to follow. But every article had the same overall scope when it came to the workflow. In Asana (our project management tool), all of the steps in the creation were strung together in a single task. We would create a few basic templates for each client, and then each piece would flow through the same steps: Briefing Writing Editing Design etc. If we had anything that didn’t fit into the “standard” workflow, we’d just tag it in the calendar with an unofficial notation \[USING BRACKETS\]. It worked. But it wasn’t ideal. Now we need the steps to be more modular. Imagine, for example, a client asks us to create a mix of deliverables: 1 article with writing + design 1 content brief 1 long-form ebook with an interview + writing + design Each of these would require its own steps and its own workflow. We need to break down the work to accommodate for a wider variety of workflows and variables. This means we need to update the fields and structure of our calendar to accommodate for the new dimensions—while also keeping the planning process simple and manageable. This leads to the next challenge: The number of “products” that we’re offering could be almost infinite. Just looking at the example scope above, you can mix and match all of these different building blocks to create a huge variety of different types of work, each requiring its own workflow. This is part of the reason we pivoted away from this model to focus on a productized, SEO-focused content service back in 2020. Take something as simple as a case study. On the surface, it seems like one deliverable that can be easily scoped and priced, right? Well, unpack what goes into a case study: Is there already source material from the customer or do we need to conduct an interview? How long is it? Is it a short overview case study or a long-form narrative? Does it need images and graphics? How many? Each of these variables opens up 2-3 possibilities. And when you combine them, we end up with something like 10 possible permutations for this single type of deliverable. It gets a bit messy. But not only do we have to figure out how to scope and price all for all of these variables, we also have to figure out how to account for these variables in the execution. We have to specify—for every deliverable—what type it is, how long, which steps are involved and not involved, the timeline for delivery, and all of the other factors. We’re approaching infinite complexity, here. We have to figure out a system that allows for a high level of flexibility to serve the diverse needs of our clients but is also productized enough that we can build workflows, process, and templates to deliver the work. I’ve spent the last few months designing that system. Failed Attempt #1: Ultra-Productization In my first pass, I tried to make it as straight forward as possible. Just sit down, make a list of all of the possible deliverables we could provide and then assign them specific scopes and services. Want a case study? Okay that’ll include an interview, up to 2,000 words of content, and 5 custom graphics. It costs $X. But this solution quickly fell apart when we started testing it against real-world scenarios. What if the client provided the brief instead of us creating one? What if they didn’t want graphics? What if this particular case study really needs to be 3,000 words but all of the others should be 2,000? In order for this system to work, we’d need to individual scope and price all of these permutations of each productized service. Then we’d need to somehow keep track of all of these and make sure that we accurately scope, price, and deliver them across dozens of clients. It’s sort of like a restaurant handling food allergies by creating separate versions of every single dish to account for every individual type of allergy. Most restaurants have figured out that it makes way more sense to have a “standard” and an “allergy-free” version. Then you only need 2 options to cover 100% of the cases. Onto the next option. Failed Attempt #2: Deliverable-Agnostic Services Next, I sat down with my head of Ops, Katy, to try to map it out. We took a big step back and said: Why does the deliverable itself even matter? At the end of the day, what we’re selling is just a few types of work (research, writing, editing, design, etc) that can be packaged up in an infinite number of ways. Rather than try to define deliverables, shouldn’t we leave it open ended for maximum flexibility? From there, we decided to break down everything into ultra-modular building blocks. We started working on this super complex system of modular deliverables where we would have services like writing, design, editing, etc—plus a sliding scale for different scopes like the length of writing or the number of images. In theory, it would allow us to mix and match any combination of services to create custom deliverables for the client. In fact, we wanted the work to be deliverable-agnostic. That way we could mold it to fit any client’s needs and deliver any type of content, regardless of the format or goal. Want a 5,000-word case study with 15 custom graphics? That’ll be $X. Want a 2,000-word blog post with an interview and no visuals? $Y. Just want us to create 10 briefs, you handle the writing, and we do design? It’s $Z. Again, this feels like a reasonable solution. But it quickly spiraled out of amuck. (That’s an Office reference.) For this to work, we need to have incredibly precise scoping process for every single deliverable. Before we can begin work (or even quote a price), we need to know pretty much the exact word count of the final article, for example. In the real world? This almost never happens. The content is as long as the content needs to be. Clients rarely know if the blog post should be 2,000 words or 3,000 words. They just want good content. We have a general ballpark, but we can rarely dial it in within just 1,000 words until we’ve done enough research to create the brief. Plus, from a packaging and pricing perspective, it introduces all kind of weird scenarios where clients will owe exactly $10,321 for this ultra-specific combination of services. We were building an open system that could accommodate any and all types of potential deliverables. On the face that seems great because it makes us incredibly flexible. In reality, the ambiguity actually works against us. It makes it harder for us to communicate to clients clearly about what they’ll get, how much it will cost, and how long it will take. That, of course, also means that it hurts our client relationships. (This actually kind of goes back to my personal learnings, which I’ll mention in a bit. I tend to be a “let’s leave things vague so we don’t have to limit our options” kind of person. But I’m working on fixing this to be more precise, specific, and clear in everything that we do.) Dialing It In: Building a Closed System We were trying to build an open system. We need to build a closed system. We need to force clarity and get specific about what we do, what we don’t do, and how much it all costs. Then we need a system to expand on that closed system—add new types of deliverables, new content playbooks, and new workflows if and when the need arises. With that in mind, we can start by mapping out the key dimensions of any type of deliverable that we would ever want to deliver. These are the universal dimensions that determine the scope, workflow, and price of any deliverable—regardless of the specific type output. Dimensions are: Brief scope Writing + editing scope Design scope Interview scope Revision (rounds) Scope, essentially, just tells us how many words, graphics, interviews, etc are required for the content we’re creating. In our first crack at the system, we got super granular with these scopes. But to help force a more manageable system, we realized that we didn’t need tiny increments for most of this work. Instead, we just need boundaries—you pay $X for up to Y words. We still need some variability around the scope of these articles. Obviously, most clients won’t be willing to pay the same price for a 1,000-word article as a 10,000-word article. But we can be smarter about the realistic break points. We boiled it down to the most common ranges: (Up to) 250 words 1,000 words 3,000 words 6,000 words 10,000 words This gives us a much more manageable number of variables. But we still haven’t exactly closed the system. We need one final dimension: Deliverable type. This tells us what we’re actually building with these building blocks. This is how we’ll put a cap on the potentially infinite number of combinations we could offer. The deliverable type will define what the final product should look like (e.g., blog post, case study, ebook, etc). And it will also give us a way to put standards and expectations around different types of deliverables that we want to offer. Then we can expand on this list of deliverables to offer new services. In the mean time, only the deliverables that we have already defined are, “on the menu,” so to speak. If a client comes to us and asks for something like a podcast summary article (which we don’t currently offer), we’ll have to either say we can’t provide that work or create a new deliverable type and define the dimensions of that specific piece. But here’s the kicker: No matter the deliverable type, it has to still fit within the scopes we’ve already defined. And the pricing will be the same. This means that if you’re looking for our team to write up to 1,000 words of content, it costs the same amount—whether it’s a blog post, an ebook, a LinkedIn post, or anything else. Rather than trying to retool our entire system to offer this new podcast summary article deliverable, we’ll just create the new deliverable type, add it to the list of options, and it’s ready to sell with the pre-defined dimensions we’ve already identified. To do: Update onboarding workflow Update contracts and scope documents Dial in new briefing process Know Thyself For the last year, I’ve been going through personal therapy. (Huge shout out to my wife, Laura, for her support and encouragement throughout the process.) It’s taught me a lot about myself and my tendencies. It’s helped me find some of my weaknesses and think about how I can improve as a person, as a partner, and as an entrepreneur. And it’s forced me to face a lot of hard truths. For example, consider some of the critical decisions I’ve made for my business: Unconventional freelance “collective” model No formal management structure Open-ended retainers with near-infinite flexibility General contracts without defined scope “Take it or leave it” approach to sales and marketing Over the years, I’ve talked about almost everything on this list as a huge advantage. I saw these things as a reflection of how I wanted to do things differently and better than other companies. But now, I see them more as a reflection of my fears and insecurities. Why did I design my business like this? Why do I want so much “flexibility” and why do I want things left open-ended rather than clearly defined? One reason that could clearly explain it: I’m avoidant. If you’re not steeped in the world of therapy, this basically means that my fight or flight response gets turned all the way to “flight.” If I’m unhappy or uncomfortable, my gut reaction is usually to withdraw from the situation. I see commitment and specificity as a prelude to future conflict. And I avoid conflict whenever possible. So I built my business to minimize it. If I don’t have a specific schedule of work that I’m accountable for delivering, then we can fudge the numbers a bit and hope they even out in the end. If I don’t set a specific standard for the length of an article, then I don’t have to let the client know when their request exceeds that limit. Conflict….avoided? Now, that’s not to say that everything I’ve built was wrong or bad. There is a lot of value in having flexibility in your business. For example, I would say that our flexible retainers are, overall, an advantage. Clients have changing needs. Having flexibility to quickly adapt to those needs can be a huge value add. And not everything can be clearly defined upfront (at least not without a massive amount of time and work just to decide how long to write an article). Overly-rigid structures and processes can be just as problematic as loosey-goosey ones. But, on the whole, I realized that my avoidant tendencies and laissez faire approach to management have left a vacuum in many areas. The places where I avoided specificity were often the places where there was the most confusion, uncertainty, and frustration from the team and from clients. People simply didn’t know what to expect or what was expected of them. Ironically, this often creates the conflict I’m trying to avoid. For example, if I don’t give feedback to people on my team, then they feel uneasy about their work. Or they make assumptions about expectations that don’t match what I’m actually expecting. Then the client might get upset, I might get upset, and our team members may be upset. Conflict definitely not avoided. This happens on the client side, too. If we don’t define a specific timeline when something will be delivered, the client might expect it sooner than we can deliver—creating frustration when we don’t meet their expectation. This conflict actually would have been avoided if we set clearer expectations upfront. But we didn’t do that. I didn’t do that. So it’s time to step up and close the gaps. Stepping Up and Closing the Gaps If I’m going to address these gaps and create more clarity and stability, I have to step up. Both personally and professionally. I have to actually face the fear and uncertainty that drives me to be avoidant. And then apply that to my business in meaningful ways that aren’t cop-out ways of kinda-sorta providing structure without really doing it. I’ve gotta be all in. This means: Fill the gaps where I rely on other people to do things that aren’t really their job but I haven’t put someone in place to do it Set and maintain expectations about our internal work processes, policies, and standards Define clear boundaries on things like roles, timelines, budgets, and scopes Now, this isn’t going to happen overnight. And just because I say that I need to step up to close these gaps doesn’t mean that I need to be the one who’s responsible for them (at least not forever). It just means that, as the business leader, I need to make sure the gaps get filled—by me or by someone else who has been specifically charged with owning that part of the operation. So, this is probably my #1 focus over the coming quarter. And it starts by identifying the gaps that exist. Then, step into those gaps myself, pay someone else to fill that role, or figure out how to eliminate the gap another way. This means going all the way back to the most basic decisions in our business. One of the foundational things about Optimist is being a “different kind” of agency. I always wanted to build something that solved for the bureaucracy, hierarchy, and siloed structure of agencies. If a client has feedback, they should be able to talk directly to the person doing the work rather than going through 3 layers of account management and creative directors. So I tried to be clever. I tried to design all kinds of systems and processes that eliminated these middle rungs. (In retrospect, what I was actually doing was designing a system that played into my avoidant tendencies and made it easy to abdicate responsibility for lots of things.) Since we didn’t want to create hierarchy, we never implemented things like Junior and Senior roles. We never hired someone to manage or direct the individual creatives. We didn’t have Directors or VPs. (Hell, we barely had a project manager for the first several years of existence.) This aversion to hierarchy aligned with our values around elevating ownership and collective contribution. I still believe in the value a flat structure. But a flat structure doesn’t eliminate the complexity of a growing business. No one to review writers and give them 1:1 feedback? I guess I’ll just have to do that….when I have some spare time. No Content Director? Okay, well someone needs to manage our content playbooks and roll out new ones. Just add it to my task list. Our flat structure didn’t eliminate the need for these roles. It just eliminated the people to do them. All of those unfilled roles ultimately fell back on me or our ops person, Katy. Of course, this isn’t the first time we’ve recognized this. We’ve known there were growing holes in our business as it’s gotten bigger and more complex. Over the years, we’ve experimented with different ways to solve for it. The Old Solution: Distributed Ops One system we designed was a “distributed ops” framework. Basically, we had one person who was the head of ops (at the time, we considered anything that was non-client-facing to be “ops”). They’d plan and organize all of the various things that needed to happen around Optimist. Then they’d assign out the work to whoever was able to help. We had a whole system for tying this into the our profit share and even gave people “Partner” status based on their contributions to ops. It worked—kinda. One big downfall is that all of the tasks and projects were ad hoc. People would pick up jobs, but they didn’t have much context or expertise to apply. So the output often varied. Since we were trying to maintain a flat structure, there was minimal oversight or management of the work. In other words, we didn’t always get the best results. But, more importantly, we still didn’t close all of the gaps entirely. Because everything was an ad-hoc list of tasks and projects, we never really had the “big picture” view of everything that needed to be done across the business. This also meant we rarely had clarity on what was important, what was trivial, and what was critical. We need a better system. Stop Reinventing the Wheel (And Create a Damn Org Chart) It’s time to get serious about filling the gaps in our business. It can’t be a half-fix or an ad hoc set of projects and tasks. We need clarity on the roles that need to be filled and then fill them. The first step here is to create an org chart. A real one. Map out all of the jobs that need to be done for Optimist to be successful besides just writers and designers. Roles like: Content director Design director SEO manager Reporting Finance Account management Business development Sales Marketing Project management It feels a bit laughable listing all of these roles. Because most are either empty or have my name attached to them. And that’s the problem. I can’t do everything. And all of the empty roles are gaps in our structure—places where people aren’t getting the direction, feedback, or guidance they need to do their best work. Or where things just aren’t being done consistently. Content director, for example, should be responsible for steering the output of our content strategists, writers, and editors. They’re not micromanaging every deliverable. But they give feedback, set overall policy, and help our team identify opportunities to get better. Right now we don’t have anyone in that role. Which means it’s my job—when I have time. Looking at the org chart (a real org chart that I actually built to help with this), it’s plain as day how many roles look like this. Even if we aren’t going to implement a traditional agency structure and a strict hierarchy, we still need to address these gaps. And the only way for that to happen is face the reality and then create a plan to close the gaps. Now that we have a list of theoretical roles, we need to clearly define the responsibilities and boundaries of those roles to make sure they cover everything that actually needs to happen. Then we can begin the process of delegating, assigning, hiring, and otherwise addressing each one. So that’s what I need to do. To be done: Create job descriptions for all of the roles we need to fill Hire Biz Dev role Hire Account Lead role(s) Hire Head of Content Playing Offense As we move into Q1 of 2025 and I reflect on the tumultuous few years we’ve had, one thought keeps running through my head. We need to play offense. Most of the last 1-2 years was reacting to changes that were happening around us. Trying to make sense and chart a new path forward. Reeling. But what I really want—as a person and as an entrepreneur—is to be proactive. I want to think and plan ahead. Figure out where we want to go before we’re forced to change course by something that’s out of our control. So my overarching focus for Q1 is playing offense. Thinking longer term. Getting ahead of the daily deluge and creating space to be more proactive, innovative, and forward thinking. To do: Pilot new content formats Audit and update our own content strategy Improve feedback workflows Build out long-term roadmap for 1-2 years for Optimist Final Note on Follow-Through and Cadence In my reflection this year, one of the things I’ve realized is how helpful these posts are for me. I process by writing. So I actually end up making a lot of decisions and seeing things more clearly each time I sit down to reflect and write my yearly recap. It also gives me a space to hold myself accountable for the things I said I would do. So, I’m doing two things a bit differently from here on out. First: I’m identifying clear action items that I’m holding myself accountable for getting done in the next 3 months (listed in the above sections). In each future update, I’ll do an accounting of what I got done and what wasn’t finished (and why). Second: I’m going to start writing shorter quarterly updates. This will gives me more chances each year to reflect, process, and make decisions. Plus it gives me a shorter feedback loop for the action items that I identified above. (See—playing offense.) — Okay friends, enemies, and frenemies. This is my first update for 2025. Glad to share with y’all. And thanks to everyone who’s read, commented, reached out, and shared their own experiences over the years. We are all the accumulation of our connections and our experiences. As always, I will pop in to respond to comments and answer questions. Feel free to share your thoughts, questions, and general disdain down below. Cheers, Tyler

Changing Careers, changing products? Age 38, Direction needed, investment advice too.
reddit
LLM Vibe Score0
Human Vibe Score0.667
Salad-BanditThis week

Changing Careers, changing products? Age 38, Direction needed, investment advice too.

Hello, At one point in my life I had a set plan that I had been following in which to design a life that fit my values, but during 2020 the viability was called into question and I have been on bad footing, unable to find stability, since. Though I currently have stable housing without roommate, and enough in savings for a year without any income and three more years in a mutual fund. The question I need help with is about utilizing approximately $40,000 that I would like to invest into a new or existing business venture, or possibly start investing my own hand in selecting stocks. To give context about the parameters of concepts that pertain to me, back in 2005 I graduated highschool and immediately was an entrepreneur, started a sports clothing company, was selling WoW bot accounts, ghillie suits on ebay, and graphic design commissions, and I was proficient in MX Flash. Although the first part of my life plan to start farming three years before 2012 for what I thought would be a peak oil economic collapse, and while watching 2008 unfold, along with my career in MX Flash falling flat, I started farming 2009. From that point I spent a total of 15 years farming, the majority of that was for my own LLC, where I was situated with leases on million dollar properties as Ag tax write off, on an elite island outside a major city, serving local high price wholesale, mainly salad mix and mushrooms, because they are fast turn around. That was truly the best 20s I could have asked for, working mainly for myself, very healthy and was putting away $10-20k in savings/investments per year, plus was earning about $3-5k more per year, while living in a cargo trailer on dirt cheap leases. But it all came to a slow end starting in 2020 when I lost all of my wholesale overnight, and my retail exploded, which burnt me out to the point I couldnt walk, as the sole worker in my LLC. So I do not fully trust the volatility of the wholesale food industry, from a small grower’s perspective, since i don't own land. SO now I am trying to figure out a way forward, because I can always farm in the future, and have taught myself hydroponics, and flat packed farm equipment, so my business is very agile and now I can grow in parking lots closer to the city for more sales opportunities, but I am not sure that is what I want to do in this current moment, because tech is exploding, and we have never had so much information available to us, it's a shame not to spend a moment in life to discover what new opportunities might be out there. I was laid off twice last year, so I've been out of work the past four months, doing thriftstore routes twice a week while making about $500+/wk, really just trying to understand what people still buy and break even, while I continue to study 3d design blender, as well as 2d digital art in the hopes that I can reconnect with my tech art past, because that is what I told myself when I was 18, that I would put off art and computers until I was past 30 and needed to do less with my body. But over the past three years, the better I get at digital art, the better Ai has been getting too. I have some mentors who might give me work and a foot in the door, but most of them are laid off, and scrounging for work if they are not on their own funded indie project. I've thought about continuing to learn 3d modeling despite Ai, and despite seeing Flash, computer program I was proficient in get removed from existence before I could really earn my money back. I assume there will always be a need for Ai models to get cleaned up, mapped and rigged, especially with AR technology coming to consumers soon, but more over it would help if I decided to go to a community college to do CNC certificates, so I can have that as a backup job on CAD at a machining warehouse and do my farm and digital art on the side, but CNC mechanics don't make a crazy amount of money and have a boss. BUT I am an inventor, and have two inventions so far, plus my ultimate goal is to one day have automated hydroponic greenhouses, using all CNC+3d printed parts to create a low time investment agriculture income, with Ai monitored greenhouse, seed to salad product that i can sell to other people, which would tie into my desire to teach people about farming too, as well as do something I enjoy, but it is not a proven concept yet. Anyways if you've read this far I appreciate it, I ultimately would like 3rd party feedback about how I should spend my $40k surplus cash. I originally had it saved and accessible in case I was going to lease land and start my full farm business again from scratch, but I think using the equipment and space I have, and exploring non-perishable products is a smart move for me right now. Should I invest in inventory of products to arbitrage online? Should I invest in the top index funds? Should I buy Silver? Should I invest in inventory of a new product line? Should I spend some money insuring and equipment for a landscaping company? I want to future proof myself the best I can as Ai unfolds, I am pretty set with an income for the rest of my life as long as I can grow food and sell it, but there are currently so many changing opportunities, I want to cast out my net and see what works with my temperment. I’ve thought about getting into cyber security, or maybe be an electrician, or less staple jobs like Landscape Architech (can use art/modeling) and CNC engineer/modeler, but honestly I prefer to make a product and sell it without client service related interaction, and particularly no boss. Thank you for reading

100 best ai sustainable business ideas in 2025
reddit
LLM Vibe Score0
Human Vibe Score1
Low_Philosopher1792This week

100 best ai sustainable business ideas in 2025

AI in Renewable Energy AI-powered smart solar panel optimization Predictive maintenance for wind turbines AI-driven energy storage management AI-based microgrid optimization Smart grid energy forecasting AI-powered water desalination efficiency AI-driven carbon footprint reduction software AI-powered hydropower efficiency monitoring AI for geothermal energy exploration AI-driven green hydrogen production optimization AI in Waste Management & Recycling AI-based waste sorting robots Smart recycling bins with AI recognition AI-powered food waste management AI-driven upcycling marketplace AI-enabled e-waste management solutions AI-powered sustainable packaging optimization AI-driven landfill management systems AI-powered plastic waste tracking and reduction AI-based waste-to-energy conversion AI-driven composting automation AI in Water Conservation AI-powered leak detection and water conservation AI-driven smart irrigation systems AI-based flood prediction and mitigation AI-powered ocean plastic cleanup robots AI-driven rainwater harvesting optimization AI-based groundwater level monitoring AI-powered desalination energy efficiency AI-driven smart water meters AI-powered wastewater treatment optimization AI-based water pollution monitoring AI in Sustainable Agriculture AI-driven precision farming AI-powered vertical farming automation AI-based pest and disease prediction AI-powered livestock health monitoring AI-driven soil health analysis AI-powered regenerative agriculture analytics AI-driven smart greenhouses AI-powered crop rotation optimization AI-based carbon farming solutions AI-powered sustainable aquaculture AI in Transportation & Mobility AI-powered electric vehicle (EV) battery optimization AI-driven smart traffic management AI-powered EV charging station optimization AI-based sustainable urban mobility planning AI-powered drone delivery for carbon reduction AI-driven logistics and supply chain sustainability AI-powered smart public transport systems AI-driven sustainable aviation fuel optimization AI-powered bicycle-sharing optimization AI-driven AI carpooling and ride-sharing efficiency AI in Green Manufacturing AI-powered energy-efficient manufacturing AI-driven supply chain sustainability analytics AI-based material waste reduction AI-powered sustainable fashion production AI-driven predictive demand to reduce overproduction AI-powered eco-friendly textile manufacturing AI-driven 3D printing for sustainable manufacturing AI-powered emission reduction in factories AI-driven green construction material optimization AI-based lifecycle assessment for eco-products AI in Carbon Offsetting & Climate Action AI-powered carbon credit marketplaces AI-driven tree planting optimization AI-based carbon capture efficiency enhancement AI-powered reforestation tracking and monitoring AI-driven climate risk prediction AI-powered environmental compliance software AI-driven sustainable investment analysis AI-based corporate sustainability tracking AI-powered carbon accounting and reporting AI-driven decarbonization roadmaps for businesses AI in Sustainable Smart Cities AI-powered urban energy efficiency monitoring AI-driven AI-powered smart lighting for cities AI-based pollution monitoring and reduction AI-driven green building automation AI-powered smart HVAC energy optimization AI-driven urban tree canopy management AI-powered digital twins for sustainable city planning AI-based urban noise pollution monitoring AI-powered public waste management optimization AI-driven citizen engagement for sustainability AI in Eco-Friendly Consumer Solutions AI-powered sustainable shopping assistant AI-driven personal carbon footprint tracking app AI-powered second-hand marketplace optimization AI-driven sustainable food delivery services AI-powered ethical supply chain transparency AI-driven zero-waste grocery stores AI-powered green subscription services AI-driven sustainable tourism planning AI-powered smart home energy efficiency optimization AI-driven personal finance for sustainability investments AI in Sustainable Healthcare & Well-being AI-powered climate impact on health analytics AI-driven sustainable hospital management AI-based predictive disease outbreak prevention AI-powered mental health solutions for eco-anxiety AI-driven green pharmaceutical production AI-powered sustainable medical waste management AI-based air quality health impact monitoring AI-driven climate-friendly diet and nutrition planning AI-powered fitness and well-being optimization for sustainability AI-driven telemedicine to reduce healthcare emissions These AI-driven sustainable business ideas offer high growth potential while making a positive impact on the planet. Let me know if you want details on a specific idea or need help with implementation strategies!

AI Content Campaign Got 4M impressions, Thousands of Website Views, Hundreds of Customers for About $100 — This is the future of marketing
reddit
LLM Vibe Score0
Human Vibe Score0.857
adamkstinsonThis week

AI Content Campaign Got 4M impressions, Thousands of Website Views, Hundreds of Customers for About $100 — This is the future of marketing

Alright. So, a few months ago I tested a marketing strategy for a client that I’ve sense dedicated my life to developing on. The Idea was to take the clients Pillar content (their YouTube videos) and use AI to rewrite the content for all the viable earned media channels (mainly Reddit). The campaign itself was moderately successful. To be specific, after one month it became their 2nd cheapest customer acquisition cost (behind their organic YouTube content). But there is a lot to be done to improve the concept. I will say, having been in growth marketing for a decade, I felt like I had hit something big with the concept. I’m going to detail how I built that AI system, and what worked well and what didn’t here. Hopefully you guys will let me know what you think and whether or not there is something here to keep working on. DEFINING THE GOAL Like any good startup, their marketing budget was minimal. They wanted to see results, fast and cheap. Usually, marketers like me hate to be in this situation because getting results usually either takes time or it takes money. But you can get results fast and cheap if you focus on an earned media strategy - basically getting featured in other people’s publication. The thing is these strategies are pretty hard to scale or grow over time. That was a problem for future me though. I looked through their analytics and saw they were getting referral traffic from Reddit - it was their 5th or 6th largest source of traffic - and they weren’t doing any marketing on the platform. It was all digital word of mouth there. It kind of clicked for me there, that Reddit might be the place to start laying the ground work. So with these considerations in mind the goal became pretty clear: Create content for relevant niche communities on Reddit with the intent of essentially increasing brand awareness. Use an AI system to repurpose their YouTube videos to keep the cost of producing unique content for each subreddit really low. THE HIGH-LEVEL STRATEGY I knew that there are huge amounts of potential customers on Reddit (About 12M people in all the relevant communities combined) AND that most marketers have a really tough time with the platform. I also knew that any earned media strategy, Reddit or not, means Click Through Rates on our content would be extremely low. A lot of people see this as a Reddit specific problem because you can’t self-promote on the platform, but really you have to keep self-promotion to a minimum with any and all earned media. This basically meant we had to get a lot of impressions to make up for it. The thing about Reddit is if your post absolutely crushes it, it can get millions of views. But crushing it is very specific to what the expectations are of that particular subreddit. So we needed to make content that was specifically written for that Subreddit. With that I was able to essentially design how this campaign would work: We would put together a list of channels (specifically subreddits to start) that we wanted to create content for. For each channel, we would write a content guideline that details out how to write great content for this subreddit. These assets would be stored in an AirTable base, along with the transcripts of the YouTube videos that were the base of our content. We would write and optimize different AI Prompts that generated different kinds of posts (discussion starters about a stock, 4-5 paragraph stock analysis, Stock update and what it means, etc…) We would build an automation that took the YouTube transcripts, ran each prompt on it, and then edited each result to match the channel writing guidelines. And then we would find a very contextual way to leave a breadcrumb back to the client. Always as part of the story of the content. At least, this is how I originally thought things would go. CHOOSING THE RIGHT SUBREDDITS Picking the right communities was vital. Here’s the basic rubric we used to pick and prioritize them: • Relevance: We needed communities interested in stock analysis, personal finance, or investing. • Subreddit Size vs. Engagement: Large subreddits offer more potential impressions but can be less focused. Smaller subreddits often have higher engagement rates. • Content Feasibility: We had to ensure we could consistently create high-value posts for each chosen subreddit. We started with about 40 possibilities, then narrowed it down to four or five that consistently delivered upvotes and user signups. CREATING CHANNEL-SPECIFIC GUIDES By the end, creating channel specific writing guidelines looked like a genius decision. Here’s how we approached it and used AI to get it done quickly: Grabbed Top Posts: We filtered the subreddit’s top posts (change filter to “Top” and then “All Time”) of all time to see the kinds of content that performed best Compiled The Relevant Posts: We took the most relevant posts to what we were trying to do and put them all on one document (basically created one document per subreddit that just had the top 10 posts in that subreddit). Had AI Create Writing Guideline Based On Posts: For each channel, we fed the document with the 10 posts with the instructions “Create a writing guideline for this subreddit based on these high performing posts. I had to do some editing on each guideline but this worked pretty well and saved a lot of time. Each subreddit got a custom guideline, and we put these inside the “Channels” table of the AirTable base we were developing with these assets. BUILDING THE AI PROMPTS THAT GENERATED CONTENT Alright this is probably the most important section so I’ll be detailed. Essentially, we took all the assets we developed up until this point, and used them to create unique posts for each channel. This mean each AI prompt was about 2,000 words of context and produced about a 500-word draft. There was a table in our AirTable where we stored the prompts, as I alluded to earlier. And these were basically the instructions for each prompt. More specifically, they detailed out our expectations for the post. In other words, there were different kinds of posts that performed well on each channel. For example, you can write a post that’s a list of resources (5 tools we used to…), or a how to guide (How we built…), etc.. Those weren’t the specific ones we used, but just wanted to really explain what I meant there. That actual automation that generated the content worked as follows: New source content (YouTube video transcript) was added to the Source Content table. This triggered the Automation. The automation grabbed all the prompts in the prompt table. For each prompt in the prompt table, we sent a prompt to OpenAI (gpt-4o) that contained first the prompt and also the source content. Then, for each channel that content prompt could be used on, we sent another prompt to OpenAI that revised the result of the first prompt based on the specific channel guidelines. The output of that prompt was added to the Content table in AirTable. To be clear, our AirTable had 4 tables: Content Channels Prompts Source Content The Source Content, Prompts, and Channel Guidelines were all used in the prompt that generated content. And the output was put in the Content table. Each time the automation ran, the Source Content was turned into about 20 unique posts, each one a specific post type generated for a specific channel. In other words, we were create a ton of content. EDITING & REFINING CONTENT The AI drafts were never perfect. Getting them Reddit-ready took editing and revising The main things I had to go in and edit for were: • Tone Adjustments: We removed excessively cliche language. The AI would say silly things like “Hello fellow redditors!” which sound stupid. • Fact-Checking: Financial data can be tricky. We discovered AI often confused figures, so we fact check all stock related metrics. Probably something like 30-40% error rate here. Because the draft generation was automated, that made the editing and getting publish ready the human bottleneck. In other words, after creating the system I spent basically all my time reviewing the content. There were small things I could do to make this more efficient, but not too much. The bigger the model we used, the less editing the content needed. THE “BREADCRUMB” PROMOTION STRATEGY No where in my prompt to the AI did I mention that we were doing any marketing. I just wanted the AI to focus on creating content that would do well on the channel. So in the editing process I had to find a way to promote the client. I called it a breadcrumb strategy once and that stuck. Basically, the idea was to never overtly promote anything. Instead find a way to leave a breadcrumb that leads back to the client, and let the really interested people follow the trail. Note: this is supposed to be how we do all content marketing. Some examples of how we did this were: Shared Visuals with a Subtle Watermark: Because our client’s product offered stock data, we’d often include a chart or graph showing a company’s financial metric with the client’s branding in the corner. Added Supporting Data from Client’s Website: If we mentioned something like a company’s cash flow statement, we could link to that company’s cash flow statement on the client’s website. It worked only because there was a lot of data on the client’s website that wasn’t gated. These tactics were really specific to the client. Which is should be. For other companies I would rethink what tactics I use here. THE RESULTS I’m pretty happy with the results • Impressions: – Early on posts averaged \~30,000 apiece, but after about a month of optimization, we hit \~70,000 impressions average. Over about two months, we reached 4 million total impressions. • Signups: – In their signups process there was one of those “Where did you find us?” questions and the amount of people who put Reddit jumped into the few hundred a month. Precise tracking of this is impossible. • Cost Efficiency (This is based on what I charged, and not the actual cost of running the campaign which is about $100/mo): – CPM (cost per thousand impressions) was about $0.08, which is far better than most paid channels. – Cost per free user: \~$8-10. After about a 10% conversion rate to a paid plan, our cost per paying user was $80–$100—well below the client’s previous $300–$400. HIGHLIGHTS: WHAT WORKED Subreddit-Specific Content: – Tailoring each post’s format and length to the audience norms boosted engagement. Worked out really well. 1 post got over 1M views alone. We regularly had posts that had hundreds of thousands. Breadcrumbs: – We never had anyone call us out for promoting. And really we weren’t. Our first priority was writing content that would crush on that subreddit. Using the Founder’s Existing Material: – The YouTube transcripts grounded the AI’s content in content we already made. This was really why we were able to produce so much content. CHALLENGES: WHAT DIDN’T WORK AI is still off: – Maybe it’s expecting too much, but still I wish the AI had done a better job. I editing a lot of content. Human oversight was critical. Scheduling all the content was a pain: – Recently I automated this pretty well. But at first I was scheduling everything manually and scheduling a hundred or so posts was a hassle. Getting Data and Analytics: – Not only did we have not very good traffic data, but the data from reddit had to be collected manually. Will probably automate this in the future. COST & TIME INVESTMENT Setup: The setup originally took me a couple weeks. I’ve since figured out how to do much faster (about 1 week). AirTable Setup here was easy and the tools costs $24/mo so not bad. ChatGPT costs were pretty cheap. Less than $75 per month. I’ve sense switched to using o1 which is much more expensive but saves me a lot of editing time Human Editing: Because this is the human part of the process and everything else was automated it mean by default all my time was spent editing content. Still this was a lot better than creating content from scratch probably by a factor of 5 or 10. The main expense was paying an editor (or using your own time) to refine posts. Worth it? Yes even with the editing time I was able to generate way more content that I would have otherwise. LESSONS & ACTIONABLE TAKEAWAYS Reddit as a Growth Channel: – If you genuinely respect each subreddit’s culture, you can achieve massive reach on a tight budget. AI + Human Collaboration: – AI excels at first drafts, but human expertise is non-negotiable for polishing and ensuring factual integrity. Soft Promotion Wins: – The “breadcrumb” approach paid off. It might feel like too light a touch, but is crucial for Reddit communities. Create once, repurpose as many times as possible: – If you have blog posts, videos, podcasts, or transcripts, feed them into AI to keep your message accurate and brand-consistent. CONCLUSION & NEXT STEPS If you try a similar approach: • Begin with smaller tests in a few niches to learn what resonates. • Create a clear “channel guide” for each community. • Carefully fact-check AI-generated posts. • Keep brand mentions low-key until you’ve established credibility.

Only 2 months of cash in the Bank for my business but was able to save it with the help of AI.
reddit
LLM Vibe Score0
Human Vibe Score1
CALLIRDAN90This week

Only 2 months of cash in the Bank for my business but was able to save it with the help of AI.

Hi there! I’m excited to share something very personal with you. We needed to book at least 2 appointments per day in the next 60 days, or my business would fail. We were already trying two acquisition channels, LinkedIn and email. The problem with these channels was that the positive response rate was very low in both. So I decided to focus on LinkedIn and get the attention of the lead by sending videos directly to them via LinkedIn messages. (You can send videos to your connections on LinkedIn if you use your cell phone.) This wasn’t new, but I added a small twist to get the lead’s attention. All the covers of the videos had a picture of me holding a sign with the person’s name and an interesting phrase. This showed some okay results, but the rest of the video was not personalized. Only the picture on the cover was. I even developed a Chrome extension for this because I thought this would be the answer and that I would book tons of appointments.  But after more trial and outreach, my leads responded, telling me that because the video itself wasn’t personalized for them, they felt like I didn’t put enough effort in, so they would not book a call with me. So after investing time and effort into my “new bright idea” and getting developers to make the Chrome extension, I was back to square one with no results. A few weeks went by, and after researching online, I found an online course from a guy who promised to teach me how to book 30+ appointments per month, guaranteed (at the time, I was making 2 or 3 appointments per week, maximum). He promised that I would only pay if he actually booked appointments for me and even offered to give me money if his course didn’t work for me. I never paid attention to internet gurus, but the offer was actually not bad, so I looked into this guy’s website. I found out he had hundreds of reviews from people who had taken his course and were talking amazing things about it. The more I read, the more excited I got. I booked a call that day and talked to a salesperson. The call was very short, and he promised I would get at least 2 appointments per day, easily. He seemed a bit cocky and told me that I just needed to trust him and the 100+ reviews from people who had taken the course. He didn’t share details, a proposal, or anything. I asked the price, and he told me it was close to $10k. (Not kidding, this was the price.) Then he told me that I would make the money back in no time with the clients I would get following his course, and that if it didn’t work, he would give me the money back. But I needed to follow everything the course said for at least 6 months. I had never paid $10k for anything in my life; it was extremely expensive for me. Also, my salary from my business was not in dollars but in a currency that was worth much less than the dollar. I continued to research more and more, but no other course was close to the number of reviews and promises that this guy had. I got desperate and told myself that I would bet everything on this course. If it worked for so many others, surely it would work for me. I got a loan from the bank and paid for the course. You might read this and think it was the most stupid thing ever, but the reality is that after 2 months in the course (I did the course as fast as I could), I learned a lot. The course was not bad; it was very extensive—probably more than 200 hours or so—and they taught a lot of things. I don’t think it was worth $10k for me, but I can see how for other people it might be worth that. Now, to the question you’re all thinking: did it get me the 2 appointments I needed per day? The answer is no. Here’s the thing: most of the techniques they taught were innovative and disruptive, but the focus was always on personalization, and they didn’t teach any way to automate the personalization. (I think, at the time they made the course, the tools didn’t exist yet.) So they taught how to do everything manually, and it took a lot—a lot of time and effort. And most annoyingly: an incredible amount of time doing operational things. I did get 2 appointments on some days, but it wasn’t consistent, and I didn’t have the time to spend 14 hours a day doing everything manually or the money to hire someone to do this for me. (I needed to also spend time delivering our service to our current clients; otherwise, they would leave.) I told them this, and they were very reasonable. After some negotiation, they gave me part of the money back. (To be fair, there was a lot of value in the course, so asking for the full $10k back would have been excessive because, in the end, it really taught me a lot of things I didn’t know.) So in the end, I spent $10k and 200+ hours on an online course, spent time and effort developing a Chrome extension, and was still not able to hit the meetings I needed. Money in the business was running out, and I needed to do something fast, or I was doomed. After investing time and effort in tools, research, and spending $10k and over 200 hours on a course that didn’t deliver the consistent results I needed, I was at a crossroads. My businesses were running out of money, and I knew I needed to find a solution quickly, or everything I had worked for would collapse. It was during this time of desperation that I started exploring other options. One night, while scrolling through the internet, I stumbled upon a 2024 article about how AI was being used to revolutionize various industries. It wasn’t directly related to appointment booking, but it sparked an idea in my mind. What if I could use AI to automate the personalization process that I had learned in the course? It seemed like a long shot, but I had nothing to lose. I started researching AI tools and technologies—YouTube videos, podcasts, pretty much everything related to AI—desperate to find something that could help me scale my outreach without investing too much time, while still maintaining the personalization that was so important. After a lot of trial and error, I found a few tools that showed promise. All of these tools were extremely new. Some of them had just launched the versions I needed just weeks ago. I can say I researched and tested more than 50 AI startups, experimenting with them, testing different approaches, checking prices (the problem was that most of them were cheap but became very expensive when applying the volume I needed to get results), and gradually refining my process. It wasn’t an overnight success, but for the first time, I felt like I was onto something that could truly work. The idea of combining AI personalization with volume was something new, and it gave me hope that I could finally book the meetings I needed without burning out. One day, I sent a video of myself talking—completely AI-generated—to my family chat group and waited for their response. None of them noticed it wasn’t actually me. At that moment, I said to myself: “Okay, I am ready to test this in the real world and see if it works.” Like everything in life, focus is key. As I mentioned earlier, we were already trying outbound strategies on LinkedIn and email, but I decided to narrow my focus to LinkedIn and specifically to video outreach. My goal was to stand out from the crowd, where most people were using text or sending generic videos. I knew that if my videos were 100% personalized, it would make a strong impression on my leads. I focused on two key metrics during my tests: Time spent on manual personalized outreach vs. AI-generated personalized outreach. Positive reply rate for non-personalized manual outreach vs. AI-generated personalized outreach. I ran a test using a sample of 50 one-minute videos sent to 50 leads, and here are the results: Time Spent to Make the Videos: Manual Process: It took me up to 10 hours to create and send 50 personalized videos. This included looking good on camera, brushing my hair, choosing appropriate clothing, ensuring proper lighting, not messing up the script, using a camera holder, recharging the phone, pausing to drink water, avoiding external sounds, being in an appropriate room, downloading the videos, deleting the videos that were not good, and sending the final ones. On average, it took me at least 12.5 minutes per one-minute video. AI Process: With AI, it took me just 32 seconds to create the exact same one-minute personalized video—without saying a word or recording a second of footage. In total, I could make and send the same 50 personalized videos in just 27 minutes. Result: The AI process was 24 times faster. Completely crazy! Positive Reply Rate: Non-Personalized Script (Manual): Using a good script without personalization (no name, job title, city, company, etc.) resulted in a positive reply rate of 4-6% on LinkedIn, including follow-ups. Personalized Script (AI): Using the same script but adding personalized details like the lead's name, company, city, and job title resulted in a positive reply rate of 15-20%, including follow-ups. Result: AI personalization led to 3x (three times) more replies. The best part was the responses. Almost everyone who replied thanked me for taking the time to research them, congratulated me on my speech, and appreciated the personalization and eloquence of my message.  These metrics were a complete breakthrough for me. I researched online to see if anyone else had done something similar, but I couldn’t find anything close. After achieving these metrics, booking the two appointments I desperately needed became easy. In fact, in the last 10 weeks, I’ve been able to consistently book 3-4 appointments per day. This success allowed me to train someone in my company to handle the process, freeing me up to focus on other aspects of the business and ultimately saving it. With the AI appointment machine we built, I even have free time now—time that I’ve been using to develop a methodology and tech tools that I now teach to others. I named the methodology Clip2Lead as a reference to the first Chrome extension I developed that didn’t work but ended up being the first step toward everything that followed. I’ve condensed everything I learned and throughout my experiences into a simple and short FREE training where I cover the entire AI appointment booking process. This includes how to find leads, create scripts, set up follow-up sequences, generate AI videos, clone your voice, compare non-AI metrics with AI metrics, and even navigate AI safety controls. I also offer Chrome extensions that helped me automate the process even further, so you can spend your time closing deals or focusing on other acquisition channels, while your AI machine for booking appointments runs with minimal effort from you. If you’re interested please get in touch with me and thank you for taking the time to read my personal story.

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰
reddit
LLM Vibe Score0
Human Vibe Score1
benfromwhereThis week

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰

(Monthly income breakdown is in the end) 📌 Introduction Hey everyone! 👋 Before I dive into this month’s breakdown, I just want to be upfront—English isn’t my first language, so I’ve used ChatGPT to refine this post for better readability. That said, everything here is 100% real—my personal experiences, struggles, and earnings as someone running a full-time AI influencer business. Since I get a lot of DMs asking about my AI models, here are their Instagram links: 📷 Emma – https://www.instagram.com/emmalauireal 📷 Jade – https://www.instagram.com/jadelaui (jadecasual is the second account) Also, if you’ve been wondering about the community I run, where I teach others how to build AI influencers from scratch, here’s the link (I got approval from mods for this link): 🔗 AI Winners Now, let’s get into what happened this month. 🚀 \------- First, a huge thank you! 🎉 Three months ago, I shared my journey of building an AI influencer business, and I was blown away by the response. That post got 263K+ views and was shared over 2.7K times—way more than I ever expected. If you’re new here or want to check out the full story of how I started, you can read it here: 🔗 Click Here (Reddit link) \------- 🔹 What I Did in January After the holiday rush in December, I knew January would be a slow month—people had already spent most of their money at the end of the year. So instead of pushing harder on monetization, I shifted my focus to tech development and optimization. Flux Character Loras: I spent a lot of time refining and testing different Flux-based character Loras for my models. This is still a work in progress, but the goal is to improve long-term consistency and make my workflow even more efficient. NSFW Content Expansion: On Emma’s side, I expanded her content library using a real model body double, making her content look more organic and natural. Jade, however, remains 100% AI-generated, keeping her workflow entirely digital. Social Media Wipeout (Thanks, VA 🙃): I had handed off both Twitter accounts to a virtual assistant to help with engagement and DMs. Big mistake. He ended up spamming DMs, which got both accounts banned—Emma (80K followers) and Jade (20K followers). 🤦‍♂️ Right now, I’m rebuilding Emma’s account from scratch and taking a much more cautious approach. Jade’s account is still offline for now. New Platform: Threads – I hadn’t touched Threads before, but since engagement on Instagram can be unpredictable, I decided to start accounts for both models. So far, they’re performing well, and I’ll continue experimenting. Launched AI Winners Community: After getting flooded with DMs (both here and on Instagram), I realized there was a massive demand for structured learning around AI influencers. So, I launched AI Winners, a paid community where I break down everything I’ve learned. It’s still early, but I see it turning into a solid, long-term community. Investment & Acquisition Talks: I’m still evaluating potential investors and acquisition offers for my AI models. There’s growing interest in buying or investing in Emma & Jade, so I’ve been having conversations to explore different options. Overall, January was about tech, rebuilding, and long-term planning—not immediate revenue. But that’s what keeps this business sustainable. 🚀 \------- ⚠️ Biggest Challenges This Month Lost Both Twitter Accounts (Massive Traffic Hit) 🚨 The biggest blow this month was losing my models’ Twitter accounts. Twitter was responsible for about 40% of my total traffic, meaning both free and paid subs took a direct hit. While Emma’s revenue took a slight dip, Jade’s income dropped significantly—partly due to the account loss and partly because January is naturally slow. (Full revenue breakdown at the end of the post.) Jade’s Instagram Tanked (Possible Shadow Ban?) 🤔 Jade’s Instagram completely lost momentum in early January. Engagement and reach dropped by over 80%, and I still haven’t figured out why. It feels like a shadow ban, but I have no clear confirmation. To counter this, I launched a second backup account, and things are starting to recover. \------- 🚀 Potential Improvements & What’s Next Locking in a Stable Workflow 🔄 Right now, Emma & Jade’s workflow is still evolving, but I’m aiming to fully stabilize it. As I’m writing this, content is generating on my second monitor—a sign that I’m close to achieving full automation without compromising quality. Boosting Jade’s Fanvue Revenue 💰 Jade’s income took a hit this month, and it’s 100% a traffic issue. The solution? More content, more reach. I’ll be increasing social media output to drive consistent traffic back to Fanvue and restore her earnings. Patreon is Done. All Focus on Fanvue 🚫 I shut down both Emma & Jade’s Patreon accounts. The goal is not to split revenue—I want everything funneled into Fanvue for higher engagement and bigger paydays. \------- 💰 January 2025 Earnings Breakdown Despite January being one of the slowest months for online creators, Emma and Jade still brought in over $29K in revenue, with a net profit exceeding $20K after all expenses. Emma Laui generated $20,206.77, with around $6,000 in expenses (chatter payments, NSFW designer fees, and other operational costs). Jade Laui earned $8,939.05, with $2,000 in expenses. Considering Twitter account losses, Instagram setbacks, and the usual January spending slump, this is still a solid outcome. The focus now is on scaling traffic and maximizing Fanvue revenue heading into February. 🚀🔥 That’s the full breakdown for January! If you have questions, feel free to drop a comment, and I’ll answer when I can. Happy to help, just like others helped me when I was starting out! 🚀🔥

A lead generation agency using personalized physical outreach
reddit
LLM Vibe Score0
Human Vibe Score1
IAmRogueStarThis week

A lead generation agency using personalized physical outreach

Hey guys! I’ve been experimenting with different outbound marketing strategies to target digital marketing agencies, specifically CEOs and founders, to promote an AI software. In the message, I invite them to test it out for free. I ran two campaigns: one using only cold email and the other combining handwritten direct mail with email follow-ups. Here are the results: Campaign 1: Cold email (3-email sequence) 200 prospects 22 responses (11%) 7 meetings booked (3.5%) Campaign 2: Handwritten direct mail + 2 follow-up emails 33 prospects 3 responses (9%) 2 meetings booked (6%) The handwritten letter approach seems more personalized and leads to better conversion rates for booked meetings (6% vs. 3.5%), but the small sample size (33 prospects) makes it hard to draw solid conclusions, I guess. My Plan This experiment got me thinking: I’d like to launch a lead generation agency to help B2B companies get meetings with their dream clients. My focus would be on sending personalized physical objects—like handwritten letters—as the first touchpoint, followed by other outreach strategies. I’m wondering: Should I increase the number of prospects contacted with handwritten direct mail to 100 to validate the results? Do you think this approach is scalable and worth investing in compared to traditional email outreach? Have you ever tried using personalized physical objects for outbound marketing? If so, what worked for you? Your feedback would be very appreciated! Thank you :)

🛒7 Strategies to Increase Retail Store Footfall post-COVID | Ultimate Blueprint & Guide 📈
reddit
LLM Vibe Score0
Human Vibe Score1
bnk3r_This week

🛒7 Strategies to Increase Retail Store Footfall post-COVID | Ultimate Blueprint & Guide 📈

Hello fellow marketers/entrepreneurs! Covid has had a gobsmacking effect on all retail promotions and marketing efforts. For people with retail businesses that thrive on footfall, it has been an uphill battle, but markets of the world are slowly resuming action. Knowing the footfall to your retail store can help you decide how many products you need to stock, which days of the week are best for promotions, and what type of promotional offers work well. The pandemic has drastically impacted customer behavior and customer loyalty is plunging. People prefer shopping online to brick-and-mortar purchases, and consumers are limiting their spending on a range of items - investing only in essentials is the norm now (McKinsey). We found some companies like Target having programs like Cartwheel that offer 5% to 50% off specific items when customers shop in-store to increase foot traffic. Strategies like these ultimately add up, an ICSC report cites that 69% of customers who went to collect their orders eventually bought additional items. I've put together a detailed list of 7 strategies to boost footfall to stores post COVID, I hope they come in handy! Abide by COVID-19 Protocols for a Safer Environment Be well-informed of the COVID-19 protocols. Don't implement this merely under the government norms, instead take extra measures to show customers that you care! Have an automated entrance Deploy hygiene counters Fix thermal sensors in the entrance Have an isolation space for those showing symptoms of the coronavirus To see more check this link for the entire list! Run Catchy In-Store Promotions Discounts are a perfect way to attract new customers and retain existing ones. When you want to increase customer traffic in a brick-and-mortar store, give customers an offer that only works inside the store. Surprise your consumers with free samples of your products. This would allow them to try some new brands and products. If you’d want to reduce your excess stock post the quarantine time, try running a multi-buy campaign. Digital Signages - Enhance In-store Shopping Experience Digital signage is a type of advertising that uses a video screen to display marketing messages. They can be used for attracting customers, conveying information, and promoting merchandise. Retail outlets in malls that have fashion sections can display the latest trends on their screens so customers know what’s new. This helps them pick out something they might like quickly. Some restaurants showcase menus on screens while others even project live cooking shows! These displays help with menu navigation too; helping a diner decide between chicken tikka masala or steak tartare by showing pictures of both dishes at once. Leverage Beacon Notification to Attract Customers to Your Store The beacon technology is a way to implement a tracking system indoors. A beacon is an inaudible signal that can be tracked and act as the trigger for other events like sending notifications about deals, discounts, or new products. Beacon technology helps with driving footfalls by giving customers an indoor mapping experience of your store's inventory. This ensures they always know where they are going and what’s around them. The navigation reminds them of their proximity to items on display so there’s never any confusion over whether something is nearby or farther off. Train your Salespeople to Become the Shopper's Friend Educating your salespersons on how to be consumers’ friends is important. They should be knowledgeable about what products are popular and in-demand so that they can help the customers find exactly what they want while at the same time giving guidance on how to save money by telling them where discounts and deals can be found. Reconceptualize Checkout Counters Customers abandon their purchases because of long lines at the checkout. With the pandemic out there, this could be one of the reasons why the retail foot traffic is diminishing. Include contactless payments that can be automated or replace your existing POS setup. Encourage BOPIS (Buy Online Pick-up In-store) To implement BOPIS for your retail store, you need to have a centralized platform that allows you to manage orders, sales, and customers. This helps you to deliver a personalized customer experience. In combination with BOPIS, another way to promote footfall into the store and drive sales in retail is by bringing your website in-store. And this will be a good move if you have multiple stores and not all the stock in one place. This is because, when you know how to calculate footfall in retail it can help you with many retail metrics like: How to plan your store for peak footfall times? How much stock you need in the store and how often you'll need to restock it? What products are selling well on an hourly basis? This is so crucial information for retailers that will help inform decisions about where to place certain items or which ones may be more popular than others etc. When stores should have promotions (if they want), discounts, and raise weekend sales? We've put together an elaborate, research-based White Paper that covers these segments: How have pandemics catalyzed technological innovations Customer sentiment and behavior during COVID-19 An omnichannel customer engagement strategy to drive sales in retail and footfall The ultimate roadmap to increase retail footfalls How to build the perfect loyalty program to turn foot traffic into brand ambassadors? You can find the same over here, hope my team's effort comes in handy to some of y'all that could improve your store visits, cheers!

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰
reddit
LLM Vibe Score0
Human Vibe Score1
benfromwhereThis week

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰

(Monthly income breakdown is in the end) 📌 Introduction Hey everyone! 👋 Before I dive into this month’s breakdown, I just want to be upfront—English isn’t my first language, so I’ve used ChatGPT to refine this post for better readability. That said, everything here is 100% real—my personal experiences, struggles, and earnings as someone running a full-time AI influencer business. Since I get a lot of DMs asking about my AI models, here are their Instagram links: 📷 Emma – https://www.instagram.com/emmalauireal 📷 Jade – https://www.instagram.com/jadelaui (jadecasual is the second account) Also, if you’ve been wondering about the community I run, where I teach others how to build AI influencers from scratch, here’s the link (I got approval from mods for this link): 🔗 AI Winners Now, let’s get into what happened this month. 🚀 \------- First, a huge thank you! 🎉 Three months ago, I shared my journey of building an AI influencer business, and I was blown away by the response. That post got 263K+ views and was shared over 2.7K times—way more than I ever expected. If you’re new here or want to check out the full story of how I started, you can read it here: 🔗 Click Here (Reddit link) \------- 🔹 What I Did in January After the holiday rush in December, I knew January would be a slow month—people had already spent most of their money at the end of the year. So instead of pushing harder on monetization, I shifted my focus to tech development and optimization. Flux Character Loras: I spent a lot of time refining and testing different Flux-based character Loras for my models. This is still a work in progress, but the goal is to improve long-term consistency and make my workflow even more efficient. NSFW Content Expansion: On Emma’s side, I expanded her content library using a real model body double, making her content look more organic and natural. Jade, however, remains 100% AI-generated, keeping her workflow entirely digital. Social Media Wipeout (Thanks, VA 🙃): I had handed off both Twitter accounts to a virtual assistant to help with engagement and DMs. Big mistake. He ended up spamming DMs, which got both accounts banned—Emma (80K followers) and Jade (20K followers). 🤦‍♂️ Right now, I’m rebuilding Emma’s account from scratch and taking a much more cautious approach. Jade’s account is still offline for now. New Platform: Threads – I hadn’t touched Threads before, but since engagement on Instagram can be unpredictable, I decided to start accounts for both models. So far, they’re performing well, and I’ll continue experimenting. Launched AI Winners Community: After getting flooded with DMs (both here and on Instagram), I realized there was a massive demand for structured learning around AI influencers. So, I launched AI Winners, a paid community where I break down everything I’ve learned. It’s still early, but I see it turning into a solid, long-term community. Investment & Acquisition Talks: I’m still evaluating potential investors and acquisition offers for my AI models. There’s growing interest in buying or investing in Emma & Jade, so I’ve been having conversations to explore different options. Overall, January was about tech, rebuilding, and long-term planning—not immediate revenue. But that’s what keeps this business sustainable. 🚀 \------- ⚠️ Biggest Challenges This Month Lost Both Twitter Accounts (Massive Traffic Hit) 🚨 The biggest blow this month was losing my models’ Twitter accounts. Twitter was responsible for about 40% of my total traffic, meaning both free and paid subs took a direct hit. While Emma’s revenue took a slight dip, Jade’s income dropped significantly—partly due to the account loss and partly because January is naturally slow. (Full revenue breakdown at the end of the post.) Jade’s Instagram Tanked (Possible Shadow Ban?) 🤔 Jade’s Instagram completely lost momentum in early January. Engagement and reach dropped by over 80%, and I still haven’t figured out why. It feels like a shadow ban, but I have no clear confirmation. To counter this, I launched a second backup account, and things are starting to recover. \------- 🚀 Potential Improvements & What’s Next Locking in a Stable Workflow 🔄 Right now, Emma & Jade’s workflow is still evolving, but I’m aiming to fully stabilize it. As I’m writing this, content is generating on my second monitor—a sign that I’m close to achieving full automation without compromising quality. Boosting Jade’s Fanvue Revenue 💰 Jade’s income took a hit this month, and it’s 100% a traffic issue. The solution? More content, more reach. I’ll be increasing social media output to drive consistent traffic back to Fanvue and restore her earnings. Patreon is Done. All Focus on Fanvue 🚫 I shut down both Emma & Jade’s Patreon accounts. The goal is not to split revenue—I want everything funneled into Fanvue for higher engagement and bigger paydays. \------- 💰 January 2025 Earnings Breakdown Despite January being one of the slowest months for online creators, Emma and Jade still brought in over $29K in revenue, with a net profit exceeding $20K after all expenses. Emma Laui generated $20,206.77, with around $6,000 in expenses (chatter payments, NSFW designer fees, and other operational costs). Jade Laui earned $8,939.05, with $2,000 in expenses. Considering Twitter account losses, Instagram setbacks, and the usual January spending slump, this is still a solid outcome. The focus now is on scaling traffic and maximizing Fanvue revenue heading into February. 🚀🔥 That’s the full breakdown for January! If you have questions, feel free to drop a comment, and I’ll answer when I can. Happy to help, just like others helped me when I was starting out! 🚀🔥

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰
reddit
LLM Vibe Score0
Human Vibe Score1
benfromwhereThis week

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰

(Monthly income breakdown is in the end) 📌 Introduction Hey everyone! 👋 Before I dive into this month’s breakdown, I just want to be upfront—English isn’t my first language, so I’ve used ChatGPT to refine this post for better readability. That said, everything here is 100% real—my personal experiences, struggles, and earnings as someone running a full-time AI influencer business. Since I get a lot of DMs asking about my AI models, here are their Instagram links: 📷 Emma – https://www.instagram.com/emmalauireal 📷 Jade – https://www.instagram.com/jadelaui (jadecasual is the second account) Also, if you’ve been wondering about the community I run, where I teach others how to build AI influencers from scratch, here’s the link (I got approval from mods for this link): 🔗 AI Winners Now, let’s get into what happened this month. 🚀 \------- First, a huge thank you! 🎉 Three months ago, I shared my journey of building an AI influencer business, and I was blown away by the response. That post got 263K+ views and was shared over 2.7K times—way more than I ever expected. If you’re new here or want to check out the full story of how I started, you can read it here: 🔗 Click Here (Reddit link) \------- 🔹 What I Did in January After the holiday rush in December, I knew January would be a slow month—people had already spent most of their money at the end of the year. So instead of pushing harder on monetization, I shifted my focus to tech development and optimization. Flux Character Loras: I spent a lot of time refining and testing different Flux-based character Loras for my models. This is still a work in progress, but the goal is to improve long-term consistency and make my workflow even more efficient. NSFW Content Expansion: On Emma’s side, I expanded her content library using a real model body double, making her content look more organic and natural. Jade, however, remains 100% AI-generated, keeping her workflow entirely digital. Social Media Wipeout (Thanks, VA 🙃): I had handed off both Twitter accounts to a virtual assistant to help with engagement and DMs. Big mistake. He ended up spamming DMs, which got both accounts banned—Emma (80K followers) and Jade (20K followers). 🤦‍♂️ Right now, I’m rebuilding Emma’s account from scratch and taking a much more cautious approach. Jade’s account is still offline for now. New Platform: Threads – I hadn’t touched Threads before, but since engagement on Instagram can be unpredictable, I decided to start accounts for both models. So far, they’re performing well, and I’ll continue experimenting. Launched AI Winners Community: After getting flooded with DMs (both here and on Instagram), I realized there was a massive demand for structured learning around AI influencers. So, I launched AI Winners, a paid community where I break down everything I’ve learned. It’s still early, but I see it turning into a solid, long-term community. Investment & Acquisition Talks: I’m still evaluating potential investors and acquisition offers for my AI models. There’s growing interest in buying or investing in Emma & Jade, so I’ve been having conversations to explore different options. Overall, January was about tech, rebuilding, and long-term planning—not immediate revenue. But that’s what keeps this business sustainable. 🚀 \------- ⚠️ Biggest Challenges This Month Lost Both Twitter Accounts (Massive Traffic Hit) 🚨 The biggest blow this month was losing my models’ Twitter accounts. Twitter was responsible for about 40% of my total traffic, meaning both free and paid subs took a direct hit. While Emma’s revenue took a slight dip, Jade’s income dropped significantly—partly due to the account loss and partly because January is naturally slow. (Full revenue breakdown at the end of the post.) Jade’s Instagram Tanked (Possible Shadow Ban?) 🤔 Jade’s Instagram completely lost momentum in early January. Engagement and reach dropped by over 80%, and I still haven’t figured out why. It feels like a shadow ban, but I have no clear confirmation. To counter this, I launched a second backup account, and things are starting to recover. \------- 🚀 Potential Improvements & What’s Next Locking in a Stable Workflow 🔄 Right now, Emma & Jade’s workflow is still evolving, but I’m aiming to fully stabilize it. As I’m writing this, content is generating on my second monitor—a sign that I’m close to achieving full automation without compromising quality. Boosting Jade’s Fanvue Revenue 💰 Jade’s income took a hit this month, and it’s 100% a traffic issue. The solution? More content, more reach. I’ll be increasing social media output to drive consistent traffic back to Fanvue and restore her earnings. Patreon is Done. All Focus on Fanvue 🚫 I shut down both Emma & Jade’s Patreon accounts. The goal is not to split revenue—I want everything funneled into Fanvue for higher engagement and bigger paydays. \------- 💰 January 2025 Earnings Breakdown Despite January being one of the slowest months for online creators, Emma and Jade still brought in over $29K in revenue, with a net profit exceeding $20K after all expenses. Emma Laui generated $20,206.77, with around $6,000 in expenses (chatter payments, NSFW designer fees, and other operational costs). Jade Laui earned $8,939.05, with $2,000 in expenses. Considering Twitter account losses, Instagram setbacks, and the usual January spending slump, this is still a solid outcome. The focus now is on scaling traffic and maximizing Fanvue revenue heading into February. 🚀🔥 That’s the full breakdown for January! If you have questions, feel free to drop a comment, and I’ll answer when I can. Happy to help, just like others helped me when I was starting out! 🚀🔥

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰
reddit
LLM Vibe Score0
Human Vibe Score1
benfromwhereThis week

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰

(Monthly income breakdown is in the end) 📌 Introduction Hey everyone! 👋 Before I dive into this month’s breakdown, I just want to be upfront—English isn’t my first language, so I’ve used ChatGPT to refine this post for better readability. That said, everything here is 100% real—my personal experiences, struggles, and earnings as someone running a full-time AI influencer business. Since I get a lot of DMs asking about my AI models, here are their Instagram links: 📷 Emma – https://www.instagram.com/emmalauireal 📷 Jade – https://www.instagram.com/jadelaui (jadecasual is the second account) Also, if you’ve been wondering about the community I run, where I teach others how to build AI influencers from scratch, here’s the link (I got approval from mods for this link): 🔗 AI Winners Now, let’s get into what happened this month. 🚀 \------- First, a huge thank you! 🎉 Three months ago, I shared my journey of building an AI influencer business, and I was blown away by the response. That post got 263K+ views and was shared over 2.7K times—way more than I ever expected. If you’re new here or want to check out the full story of how I started, you can read it here: 🔗 Click Here (Reddit link) \------- 🔹 What I Did in January After the holiday rush in December, I knew January would be a slow month—people had already spent most of their money at the end of the year. So instead of pushing harder on monetization, I shifted my focus to tech development and optimization. Flux Character Loras: I spent a lot of time refining and testing different Flux-based character Loras for my models. This is still a work in progress, but the goal is to improve long-term consistency and make my workflow even more efficient. NSFW Content Expansion: On Emma’s side, I expanded her content library using a real model body double, making her content look more organic and natural. Jade, however, remains 100% AI-generated, keeping her workflow entirely digital. Social Media Wipeout (Thanks, VA 🙃): I had handed off both Twitter accounts to a virtual assistant to help with engagement and DMs. Big mistake. He ended up spamming DMs, which got both accounts banned—Emma (80K followers) and Jade (20K followers). 🤦‍♂️ Right now, I’m rebuilding Emma’s account from scratch and taking a much more cautious approach. Jade’s account is still offline for now. New Platform: Threads – I hadn’t touched Threads before, but since engagement on Instagram can be unpredictable, I decided to start accounts for both models. So far, they’re performing well, and I’ll continue experimenting. Launched AI Winners Community: After getting flooded with DMs (both here and on Instagram), I realized there was a massive demand for structured learning around AI influencers. So, I launched AI Winners, a paid community where I break down everything I’ve learned. It’s still early, but I see it turning into a solid, long-term community. Investment & Acquisition Talks: I’m still evaluating potential investors and acquisition offers for my AI models. There’s growing interest in buying or investing in Emma & Jade, so I’ve been having conversations to explore different options. Overall, January was about tech, rebuilding, and long-term planning—not immediate revenue. But that’s what keeps this business sustainable. 🚀 \------- ⚠️ Biggest Challenges This Month Lost Both Twitter Accounts (Massive Traffic Hit) 🚨 The biggest blow this month was losing my models’ Twitter accounts. Twitter was responsible for about 40% of my total traffic, meaning both free and paid subs took a direct hit. While Emma’s revenue took a slight dip, Jade’s income dropped significantly—partly due to the account loss and partly because January is naturally slow. (Full revenue breakdown at the end of the post.) Jade’s Instagram Tanked (Possible Shadow Ban?) 🤔 Jade’s Instagram completely lost momentum in early January. Engagement and reach dropped by over 80%, and I still haven’t figured out why. It feels like a shadow ban, but I have no clear confirmation. To counter this, I launched a second backup account, and things are starting to recover. \------- 🚀 Potential Improvements & What’s Next Locking in a Stable Workflow 🔄 Right now, Emma & Jade’s workflow is still evolving, but I’m aiming to fully stabilize it. As I’m writing this, content is generating on my second monitor—a sign that I’m close to achieving full automation without compromising quality. Boosting Jade’s Fanvue Revenue 💰 Jade’s income took a hit this month, and it’s 100% a traffic issue. The solution? More content, more reach. I’ll be increasing social media output to drive consistent traffic back to Fanvue and restore her earnings. Patreon is Done. All Focus on Fanvue 🚫 I shut down both Emma & Jade’s Patreon accounts. The goal is not to split revenue—I want everything funneled into Fanvue for higher engagement and bigger paydays. \------- 💰 January 2025 Earnings Breakdown Despite January being one of the slowest months for online creators, Emma and Jade still brought in over $29K in revenue, with a net profit exceeding $20K after all expenses. Emma Laui generated $20,206.77, with around $6,000 in expenses (chatter payments, NSFW designer fees, and other operational costs). Jade Laui earned $8,939.05, with $2,000 in expenses. Considering Twitter account losses, Instagram setbacks, and the usual January spending slump, this is still a solid outcome. The focus now is on scaling traffic and maximizing Fanvue revenue heading into February. 🚀🔥 That’s the full breakdown for January! If you have questions, feel free to drop a comment, and I’ll answer when I can. Happy to help, just like others helped me when I was starting out! 🚀🔥

AI-Strategies-StockMarket
github
LLM Vibe Score0.407
Human Vibe Score0.026251017937713218
Solano96Mar 28, 2025

AI-Strategies-StockMarket

Artificial Intelligence Trading App to test strategies based on artificial intelligence for investing in the stock market. The program has two simple investment strategies to compare results. One of these strategies is simply to buy and hold. The other is a classic strategy based on the crossing of Moving Averages and the use of the Relative Strength Index or RSI. At this moment the app has the following strategies based on artificial intelligence: Deep Neural Network: strategy that tries to predict the market trend with the use of neural networks that take different technical indicators as inputs. Strategy that combines in a weighted way buy-sell signals coming from moving average crosses. The weights are obtained through the PSO (Particle swarm optimization) algorithm. Getting Started 🚀 These instructions will get you a copy of the project up and running on your local machine for development and testing purposes. The local installation has been successfully tested in Ubuntu 18.04. Prerequisites 📋 Have installed Python3, you can check with the following command in your terminal: In case you did not have Python3 installed, you can use the following commands: To use the program with interface is necessary to intall tkinter with the following command: Installing 🔧 First clone the repository: Now we need to install some dependencies. To do this, execute the following command: Usage 📦 First we need to activate the virtual environment with: You can use the command line program, that can be execute as follow: You can also use the short options: Example: PD: You can use as data name any market abbreviation recognized by yahoo finance. After the execution you can find the results in 'reports' where you can find a PDF report with a summary of the execution. Author ✒️ Francisco Solano López Rodríguez

eiten
github
LLM Vibe Score0.549
Human Vibe Score0.754375921646308
tradyticsMar 27, 2025

eiten

Eiten - Algorithmic Investing Strategies for Everyone Eiten is an open source toolkit by Tradytics that implements various statistical and algorithmic investing strategies such as Eigen Portfolios, Minimum Variance Portfolios, Maximum Sharpe Ratio Portfolios, and Genetic Algorithms based Portfolios. It allows you to build your own portfolios with your own set of stocks that can beat the market. The rigorous testing framework included in Eiten enables you to have confidence in your portfolios. If you are looking to discuss these tools in depth and talk about more tools that we are working on, please feel free to join our Discord channel where we have a bunch of more tools too. Files Description | Path | Description | :--- | :---------- | eiten | Main folder. | └ figures | Figures for this github repositories. | └ stocks | Folder to keep your stock lists that you want to use to create your portfolios. | └ strategies | A bunch of strategies implemented in python. | backtester.py | Backtesting module that both backtests and forward tests all portfolios. | data_loader.py | Module for loading data from yahoo finance. | portfolio_manager.py | Main file that takes in a bunch of arguments and generates several portfolios for you. | simulator.py | Simulator that uses historical returns and monte carlo to simulate future prices for the portfolios. | strategy_manager.py | Manages the strategies implemented in the 'strategies' folder. Required Packages You will need to install the following package to train and test the models. Scikit-learn Numpy Tqdm Yfinance Pandas Scipy You can install all packages using the following command. Please note that the script was written using python3. Build your portfolios Let us see how we can use all the strategies given in the toolkit to build our portfolios. The first thing you need to do is modify the stocks.txt file in the stocks folder and add the stocks of your choice. It is recommended to keep the list small i.e anywhere between 5 to 50 stocks should be fine. We have already put a small stocks list containing a bunch of tech stocks like AAPL, MSFT, TSLA etc. Let us build our portfolios now. This is the main command that you need to run. This command will use last 5 years of daily data excluding the last 90 days and build several portfolios for you. Based on those portfolios, it will then test them on the out of sample data of 90 days and show you the performance of each portfolio. Finally, it will also compare the performance with your choice of market index which is QQQ here. Let's dive into each of the parameters in detail. istest: The value determined if the program is going to keep some separate data for future testing. When this is enabled, the value of futurebars should be larger than 5. future_bars: These are the bars that the tool will exclude during portfolio building and will forward test the portfolios on the excluded set. This is also called out of sample data. datagranularityminutes: How much granular data do you want to use to build your portfolios. For long term portfolios, you should use daily data but for short term, you can use hourly or minute level data. The possible values here are 3600, 60, 30, 15, 5, 1. 3600 means daily. historytouse: Whether to use a specific number of historical bars or use everything that we receive from yahoo finance. For minute level data, we only receive up to one month of historical data. For daily, we receive 5 years worth of historical data. If you want to use all available data, the value should be all but if you want to use smaller history, you can set it to an integer value e.g 100 which will only use the last 100 bars to build the portfolios. applynoisefiltering: This uses random matrix theory to filter out the covariance matrix from randomness thus yielding better portfolios. A value of 1 will enable it and 0 will disable it. market_index: Which index do you want to use to compare your portfolios. This should mostly be SPY but since we analyzed tech stocks, we used QQQ. only_long: Whether to use long only portfolio or enable short selling as well. Long only portfolios have shown to have better performance using algorithmic techniques. eigenportfolionumber: Which eigen portfolio to use. Any value between 1-5 should work. The first eigen portfolio (1) represents the market portfolio and should act just like the underlying index such as SPY or QQQ. The second one is orthogonal and uncorrelated to the market and poses the greatest risk and reward. The following ones have reduced risk and reward. Read more on eigen-portfolios. stocksfilepath: File that contains the list of stocks that you want to use to build your portfolio. Some Portfolio Building Examples Here are a few examples for building different types of portfolios. Both long and short portfolios by analyzing last 90 days data and keeping the last 30 days as testing data. This will give us 60 days of portfolio construction data and 30 days of testing. Only long portfolio on 60 minute bars of the last 30 days. No future testing. Compare the results with SPY index instead of QQQ. Do not apply noise filtering on the covariance matrix. Use the first eigen portfolio (market portfolio) and compare with SQQQ, Portfolio Strategies Four different portfolio strategies are currently supported by the toolkit. Eigen Portfolios These portfolios are orthogonal and uncorrelated to the market in general thus yielding high reward and alpha. However, since they are uncorrelated to the market, they can also provide great risk. The first eigen portfolio is considered to be a market portfolio which is often ignored. The second one is uncorrelated to the others and provides the highest risk and reward. As we go down the numbering, the risk as well as the reward are reduced. Minimum Variance Portfolio (MVP) MVP tries to minimize the variance of the portfolio. These portfolios are lowest risk and reward. Maximum Sharpe Ratio Portfolio (MSR) MSR solves an optimization problem that tries to maximize the sharpe ratio of the portfolio. It uses past returns during the optimization process which means if past returns are not the same as future returns, the results can vary in future. Genetic Algorithm (GA) based Portfolio This is our own implementation of a GA based portfolio that again tries to maximize the sharpe ratio but in a slightly more robust way. This usually provides more robust portfolios than the others. When you run the command above, our tool will generate portfolios from all these strategies and give them to you. Let us look at some resulting portfolios. Resulting Portfolios For the purpose these results, we will use the 9 stocks in the stocks/stocks.txt file. When we run the above command, we first get the portfolio weights for all four strategies. For testing purposes, the above command used last five years of daily data up till April 29th. The remaining data for this year was used for forward testing i.e the portfolio strategies had no access to it when building the portfolios. What if my portfolio needs different stocks?: All you need to do is change the stocks in the stocks.txt file and run the tool again. Here is the final command again that we run in order to get our portfolios: Portfolio Weights We can see that the eigen portfolio is giving a large weight to TSLA while the others are dividing their weights more uniformly. An interesting phenomena happening here is the hedging with SQQQ that all the strategies have learned automatically. Every tool is assigning some positive weight to SQQQ while also assigning positive weights to other stocks which indicates that the strategies are automatically trying to hedge the portfolios from risk. Obviously this is not perfect, but just the fact that it's happening is fascinating. Let us look at the backtest results on the last five years prior to April 29, 2020. Backtest Results The backtests look pretty encouraging. The black dotted line is the market index i.e QQQ. Other lines are the strategies. Our custom genetic algorithm implementation seems to have the best backtest results because it's an advanced version of other strategies. The eigen portfolio that weighed TSLA the most have the most volatility but its profits are also very high. Finally, as expected, the MVP has the minimum variance and ultimately the least profits. However, since the variance is extremely low, it is a good portfolio for those who want to stay safe. The most interesting part comes next, let us look at the forward or future test results for these portfolios. Forward Test Results These results are from April 29th, 2020 to September 4th, 2020. The eigen portfolio performed the best but it also had a lot of volatility. Moreover, most of those returns are due to TSLA rocketing in the last few months. After that, our GA algorithm worked quite effectively as it beat the market index. Again, as expected, the MVP had the lowest risk and reward and slowly went up in 4-5 months. This shows the effectiveness and power of these algorithmic portfolio optimization strategies where we've developed different portfolios for different kinds of risk and reward profiles. Conclusion and Discussion We are happy to share this toolkit with the trading community and hope that people will like and contribute to it. As is the case with everything in trading, these strategies are not perfect but they are based on rigorous theory and some great empirical results. Please take care when trading with these strategies and always manage your risk. The above results were not cherry picked but the market has been highly bullish in the last few months which has led to the strong results shown above. We would love for the community to try out different strategies and share them with us. Special Thanks Special thanks to Scott Rome's blog. The eigen portfolios and minimum variance portfolio concepts came from his blog posts. The code for filtering eigen values of the covariance matrix was also mostly obtained from one of his posts. License A product by Tradytics Copyright (c) 2020-present, Tradytics.com

5 Genius Ways to Make Money From Home (Using AI)
youtube
LLM Vibe Score0.419
Human Vibe Score0.77
Charlie ChangNov 15, 2023

5 Genius Ways to Make Money From Home (Using AI)

Check out Fundrise to get started with investing in pre-IPO blue-chip companies that are leading the AI industry: http://fundrise.com/charliechang #fundrisetestimonial #fundrisepartner In this video, I'm going to share 5 genius ways to make money online, using AI (that are all proven). I'll also give you a clear outline and show you exactly how to leverage these new AI opportunities to make money online. ► Daily advice and BTS on my Instagram: https://www.instagram.com/charliechang/ ► Get access to my FREE side hustle courses: https://www.sidehustlemastery.com My favorite business must-haves: 💳 Best business credit cards: https://yourbestcreditcards.com/card-finder/?ccid=2004 🏦 Novo (best business bank): https://startupwise.com/novo 🖥️ Best AI website builder ($3/month using code CHARLIECHANG): https://hostinger.com/charliechang ⚙️ Northwest (best $39 LLC formation service): https://startupwise.com/northwestLLC 🥇 Hire top 1% overseas talent: https://paired.so Whether it's optimizing businesses, doing social media management, or investing in pre-IPO tech companies, there are so many interesting opportunities that are out there for you guys to take advantage of. I highly encourage every aspiring entrepreneur out there to find a way to use AI because this can absolutely change the efficiency and output of your business. If you liked the video, and you want to see more videos on AI and making money, check out my videos: How To Use ChatGPT To Learn ANY Skill Quickly (Tutorial): https://www.youtube.com/watch?v=vYvOTGk7hOA 5 Passive Income Ideas - How I ACTUALLY Make $35K/Week in 2023 https://youtu.be/TVLgIKMOYJ0 I hope you guys found this video helpful, and if you did please share it with a friend or family member who you think could benefit and also LIKE and SUBSCRIBE for more videos like this in the future! Thank you for watching and I hope you have a wonderful rest of your day! – Charlie #AI #Money #SideHustle Timeline: 0:00 - Introduction 0:28 - Social Media Management Business 3:11 - AI Optimization Agency 5:41 - Investing in Pre-IPO AI Companies With Fundrise 8:12 - Building an E-commerce Business 10:00 - AI Automated Affiliate Marketing Business 11:12 - Conclusion 11:40 - Outro Disclaimer: Some of the links above may be affiliate links, which means that if you click on them I may receive a small commission. The retailers and financial services companies pay the commission at no cost to you, and this helps to support our channel and keep our videos free. Thank you! In addition, I am not a financial advisor. Charlie Chang does not provide tax, legal or accounting advice. The ideas presented in this video are for entertainment purposes only. Please do your own due diligence before making any financial decisions. ► My Instagram: https://www.instagram.com/charliechang/

The future of AI
youtube
LLM Vibe Score0.471
Human Vibe Score0.61
GaryVeeMay 9, 2023

The future of AI

When voice and ai hit scale … shits gonna get interesting… — Thanks for watching! Join My Discord!: https://www.garyvee.com/discord Check out another series on my channel: Keynotes: https://www.youtube.com/watch?v=6vCDlmhRmBo&list=PLfA33-E9P7FCEF1izpctGGoak841XYzrJ NFTs: https://www.youtube.com/watch?v=AwMJ6bScB2s&list=PLfA33-E9P7FAcvsVSFqzSuJhHu3SkW2Ma Business Meetings: https://www.youtube.com/watch?v=wILI_VV6z4Y&list=PLfA33-E9P7FCTIY62wkqZ-E1cwpc2hxBJ Gary Vaynerchuk Original Films: https://youtube.com/playlist?list=PLfA33-E9P7FAvnrOcgy4MvIcCXxoyjuku Trash Talk: https://youtube.com/playlist?list=PLfA33-E9P7FDelN4bXFgtJuczC9HHmm2- WeeklyVee: https://youtube.com/playlist?list=PLfA33-E9P7FBPjdQcF6uedz9fdk8XKn-b — Gary Vaynerchuk is a serial entrepreneur, and serves as the Chairman of VaynerX, the CEO of VaynerMedia and the Creator & CEO of VeeFriends. Gary is considered one of the leading global minds on what’s next in culture, relevance and the internet. Known as “GaryVee” he is described as one of the most forward thinkers in business – he acutely recognizes trends and patterns early to help others understand how these shifts impact markets and consumer behavior. Whether its emerging artists, esports, NFT investing or digital communications, Gary understands how to bring brand relevance to the forefront. He is a prolific angel investor with early investments in companies such as Facebook, Twitter, Tumblr, Venmo, Snapchat, Coinbase and Uber. Gary is an entrepreneur at heart — he builds businesses. Today, he helps Fortune 1000 brands leverage consumer attention through his full service advertising agency, VaynerMedia which has offices in NY, LA, London, Mexico City, LATAM and Singapore. VaynerMedia is part of the VaynerX holding company which also includes VaynerProductions, VaynerNFT, Gallery Media Group, The Sasha Group, Tracer, VaynerSpeakers, VaynerTalent, and VaynerCommerce. Gary is also the Co-Founder of VaynerSports, Resy and Empathy Wines. Gary guided both Resy and Empathy to successful exits — both were sold respectively to American Express and Constellation Brands. He’s also a Board Member at Candy Digital, Co-Founder of VCR Group, Co-Founder of ArtOfficial, and Creator & CEO of VeeFriends. Gary was recently named to the Fortune list of the Top 50 Influential people in the NFT industry. In addition to running multiple businesses, Gary documents his life daily as a CEO through his social media channels which has more than 34 million followers and garnishes over 272 million monthly impressions/views across all platforms. His podcast ‘The GaryVee Audio Experience’ ranks among the top podcasts globally. He is a five-time New York Times Best-Selling Author and one of the most highly sought after public speakers. Gary serves on the board of MikMak, Bojangles Restaurants, and Pencils of Promise. He is also a longtime Well Member of Charity:Water.