VibeBuilders.ai Logo
VibeBuilders.ai

Were

Explore resources related to were to help implement AI solutions for your business.

 Looking for beta testers for my AI-powered website builder - no templates, no coding required
reddit
LLM Vibe Score0
Human Vibe Score1
Interesting_Flow_342This week

Looking for beta testers for my AI-powered website builder - no templates, no coding required

Hey r/sideproject, I'm working on an exciting new project since 4 months- an AI-powered website builder that creates completely custom, professional-looking websites from scratch. No templates, no coding The key capabilities of this AI website builder are: Designing unique, mobile-responsive layouts based on your preferences and content Generating custom written content for each page, section, and element Ensuring best practices for things like typography, color schemes, and SEO But the real power comes in the customization. Once the AI generates your initial website, you can easily make changes to any part of it - from the design and layout to the text and images. Simply select the specific element you want to modify, and the AI will make the requested changes, whether that's tweaking the font and colors, rearranging the page structure, or rewriting the copy. It's a truly interactive, AI-driven web building experience. This is perfect for things like: Marketing/informational websites Landing pages Online resumes and portfolios Small business websites When you're ready, you can publish your AI-generated, fully customized website on a free subdomain or download the full code. I'm looking to get a few early users to try this out and provide feedback before the full public launch. If you're interested in being a beta tester, I'd love to hear from you! This could be especially useful for small business owners, freelancers, job seekers, or anyone who needs a professional web presence but doesn't have the time or skills for traditional web development. If you're interested, just leave a comment below or send me a DM. I'll be in touch to get you set up with early access. Thanks for checking it out! Muhammad Bilal Moten

Introducing Vest: Your AI-Powered Due Diligence Partner - Looking for feedback!
reddit
LLM Vibe Score0
Human Vibe Score1
nervousslinkyThis week

Introducing Vest: Your AI-Powered Due Diligence Partner - Looking for feedback!

TLDR; We are introducing Vest, an AI powered due-diligence and stock recommendation platform. We have bootstrapped ourselves so far and are wanting to get as much feedback from Reddit as we can to see where we can improve, but also what we are doing right. So please have a look around, give us feedback and if you like it, feel free to use it. Hi Reddit, My name is Drian and I'm one of the founders of Vest. We believe we are crafting something special at Vest and we want to get the word out and gather as much feedback as possible! Our major goal at Vest is to help new retail investors make sense of the investment landscape and get AI powered assistance, or even help experienced investors get confirmation of their potential moves. Overall, we want people to start their journey to financial freedom and not be daunted by the complexity of it. So how do we do this? Vest is a user-friendly service that harnesses fundamental metrics, social and news sentiment, and technical analysis, that we feed into some advanced AI models to generate clear buy, sell, or hold signals for US-based (for now!) stocks, offering our users transparent due-diligence for confident investing. The service is currently free with no ads - however, at some point we do plan on adding a paid tier. What's included: ​ Financial Metrics. Our financial metrics take all the potentially complex mathematical equations and present the fundamentals of a company to users in a simple 1 pager, with a score displaying if the metric is positive for a stock. We also provide publicly available analyst ratings from investment banks as well as price targets they have set. News Sentiment. We take publications about a specific stock from new articles, journals and socials and give these all a rating to determine if social sentiment is positive around a stock or not. Each article and its rating is visible to our users through through our dashboard. AI assisted Stock Signals. We have developed an algorithm to take all the metrics, sentiment and technical analysis we collate and analyze this with historic performance data for every stock to attempt to figure out if a stock is undervalued (great time to buy) or overvalued (great time to sell). 155 US stock tickers and counting. We currently have trained our models for around 155 US based stocks on the NASDAQ and NYSE exchanges. As we get more funding/runway we do plan on adding more, with the eventual goal to expand to more exchanges, countries and securities. Knowledge base and community. Our knowledge base & community contains explanations and articles for all metrics and the other good stuff behind Vest. We don’t want to just tell users what to do, but to also assist in their financial education. We hope our knowledge base can also become a thriving community where users can interact with us and each, ask questions around investing and keep gaining knowledge. Is it 100% accurate? Absolutely not. While we do a pretty great job at tracking and surfacing signals, we are not presenting a fool-proof, silver bullet with a guarantee here - rather a starting point for users to make more informed decisions, find potential new investment opportunities and hopefully learn about investing as they do so. We encourage our users to do their own research and due-diligence and not just take our signals as gospel - we know each and every person has a different risk appetite and goals, and we encourage you to use Vest in a way that fits with your own financial goals and risk appetite. We also display our win rates, average returns, and comparisons with buy and hold for each stock - and we are transparent about it when we’ve fallen short. Next steps: ​ Hope over to vestapp.ai and sign-up From the dashboard, play around, inspect our stock information and add some stocks to your watchlist. If you like what you see, and you’ve done your homework - use your favourite brokerage account to make an investment and watch Vest for changes in a stocks signals. If you don’t have one, we have a pop-up when you click buy/sell on any given stock with some non-affiliated brokerage options for the US, Australia and New Zealand - we don’t get a kickback from these brokerages, they are just what we’ve personally been using. FEEDBACK - We’re just getting started and we know the value of a fresh pair of eyes - our current mission is to get as much feedback as possible - anything you think of please send it through here or on the dedicated feedback form on our website in the sidebar on the left. Features we’re working on We're quietly thrilled about the direction Vest is headed, and we want to give you a sneak peek of what's in store for the next couple of quarters. Some of these may roll out as premium features, but we're diligently fine-tuning the details. Here's what you can expect: ​ Insider Trading Insights: Get daily reports on major stock moves by whales and company insiders. Institutional Holders: We're adding daily reports on institutional holders, keeping you informed about their moves. Lobbying Activity: We're actively working on daily updates about lobbying activities, so you can stay informed. Government Contracts Data: We'll provide a quarterly snapshot of government contract values for the companies you're tracking. US Congress Stock Activity: Keep an eye on daily trading actions of House and Senate members. Daily Summaries & Signal Alerts: We're currently hard at work on this feature. Soon, receive daily email summaries covering signals, watchlist updates, and key news. Personalized Risk Management: Tailor signals to match your unique risk management strategy. Your investments, your way. AI Assistant: Our LLM integration is almost ready, allowing you to ask it straightforward questions about particular securities in plain English. It will provide you with real-time context on fundamentals, news, and all the metrics and data points we monitor.

We've built an AI-powered business building platform, and we're looking for entrepreneurs to try out the MVP!
reddit
LLM Vibe Score0
Human Vibe Score1
UltraIngoThis week

We've built an AI-powered business building platform, and we're looking for entrepreneurs to try out the MVP!

Hey r/sideproject! I'm Felix, co-founder of Buildpad, and we're excited to share our latest project with you. https://reddit.com/link/1eve8n4/video/ahktfda2bgjd1/player Buildpad is an AI-powered (Claude Sonnet 3.5) business-building platform that guides entrepreneurs through every step of creating and growing a business. Here's what makes it unique: Idea validation: Leverage Reddit's API to get real-world data on your ideas through posts, comments and discussions. Structured process: Follow a clear roadmap from idea validation to launch and beyond. Team collaboration: Work with co-founders, all assisted by the same AI. Central context bank: Our AI remembers everything about your project for consistent, informed guidance. We're solving the common problem of entrepreneurs not knowing what to do next, especially during idea generation and validation phases. With Buildpad, you can validate your ideas by searching for relevant keywords across Reddit, helping you understand if people are actually experiencing the problems you're aiming to solve. We're in the MVP stage and looking for early adopters to test the platform and provide feedback. We'd love to hear from you: Does this solution resonate with your entrepreneurial challenges? What features would you find most valuable in a tool like this? Any thoughts or concerns about using AI for startup guidance? If you're interested in trying out Buildpad or have any questions, please comment below or DM me. Thanks for checking it out! buildpad.io

Finding domains for a business: The troubles faced and how they were solved.
reddit
LLM Vibe Score0
Human Vibe Score0.6
DrobushevskiyThis week

Finding domains for a business: The troubles faced and how they were solved.

Hey everyone! I’m sure some of you have experience searching for a domain name for your project or startup. And you know how hard it can be to find the right one. You want it to be short, memorable, SEO-friendly, free of a bad history, and relevant to your project’s meaning. As a solo entrepreneur, I’ve faced the same challenges. I tried using domain auctions and drop-catching platforms to find short and valuable domain names for my projects and for resale. But these platforms can be frustrating – there’s too much competition, bidding wars drive up prices, and waiting for a domain to become available takes forever. GoDaddy auctions can last up to 10 days, and placing a backorder doesn’t always guarantee success. This process can be stressful and time-consuming. I just wanted a way to quickly grab the right domain and start using it immediately – without all the waiting and worrying. One day, I found a great domain on Product Hunt. The product was abandoned, and the domain was available. I thought, "What if I could find more domains like this in the same niche from this site?" and "How can I automate this?” That’s how I ended up creating GoneDomains GoneDomains helps to find available domain names from popular websites like Product Hunt, Medium, Hacker News, Forbes, and others. It saves hours of searching and eliminates the stress of competing with other buyers. Recently, I added a Domain Rating (DR) metric for each domain, making it easier to find valuable domains for SEO. If you’re familiar with DR, you know that domains with high DR can boost SEO rankings. Dashboard of GoneDomains with the filter Now, I’m working on new features: A feature that shows the average price of domains across multiple sources. A tool to check how many domain extensions are already registered for a specific name. AI-powered analysis to determine a domain’s niche and keywords, plus a filter for one-, two-, or three-word domains. Today, GoneDomains has over 30,000 available domain names sourced from platforms like Product Hunt, Medium, Hacker News, Forbes, TechCrunch, and more. New domains are added daily. GoneDomains saves you from spending hours manually searching, dealing with bidding wars, waiting for auctions to end, and unnecessary stress.

Building an AI-Powered Cold Email Blocker for Business Leaders – Would You Use This?
reddit
LLM Vibe Score0
Human Vibe Score1
Pale-Examination4855This week

Building an AI-Powered Cold Email Blocker for Business Leaders – Would You Use This?

Hey everyone, I've been working on a side project that I'm really excited about, and I'd love to get some feedback from the community here. I noticed that CEOs, CMOs, CTOs, and other high-level executives often get bombarded with cold emails daily. Sorting through them can be a huge time sink, and sometimes important emails can get lost in the noise. To solve this, I'm developing a browser extension that uses AI to scan incoming emails and block unwanted cold emails before they even reach your inbox. The idea is to save time, reduce inbox clutter, and help business leaders focus on the emails that really matter. Here’s a bit more detail: AI-Powered Filtering: The extension will analyze the content of incoming emails to determine if they're likely cold emails Customizable Settings: Users can adjust the AI prompt to whatever they want to block Seamless Integration: It’ll work directly within your existing email client as a browser extension, so there’s no need to switch apps or platforms. Privacy First: All the processing happens locally, so your email data stays private and secure. I’m still in the development phase, and I'd love to hear your thoughts: Would this be something you'd use or find valuable? Are there any specific features or pain points you'd like to see addressed? Any suggestions on how to improve the concept? Thanks in advance for your feedback! Your insights will be invaluable as I continue to refine this project.

Introducing Novus – an AI-powered QA agent that automates testing for your web apps!
reddit
LLM Vibe Score0
Human Vibe Score1
namish800This week

Introducing Novus – an AI-powered QA agent that automates testing for your web apps!

Hello, I'm excited to introduce a project I've been working on—an AI-powered QA agent designed to streamline and enhance the testing process for web applications. Here's how it works: Key Features: Natural Language Test Definitions: You can define the behavior you want to validate using plain English. Automated Navigation and Validation: The agent autonomously navigates your web app and checks if the specified behavior functions as expected. Comprehensive Reporting: After execution, it provides detailed reports, including step-by-step actions, screenshots, and video recordings.​ How It Works: Define Behavior: Describe the functionality you want to test in simple English.​ Run Test: The agent interprets your description, interacts with your web app accordingly, and validates the outcomes. Review Results: Access detailed reports that include all actions taken, along with visual documentation like screenshots and videos.​ Current Capabilities: Dashboard for Test Management: Create and manage multiple test suites and individual tests through an intuitive interface.​ Visual Regression Analysis: Utilize visual artifacts to perform regression analysis and ensure UI consistency.​ Future Plans: Intelligent Reporting: Implement advanced reporting features to provide deeper insights and analytics. Enhanced Visual Regression: Develop more sophisticated tools for detecting and analyzing visual discrepancies.​ I'm eager to hear your thoughts and feedback. What challenges do you face in QA testing? How do you see AI tools fitting into your workflow? Let's discuss! Here's the demo of what I've built so far https://www.loom.com/share/11b1dd4d18124f9a8032ae81e9cbdab4?sid=56237f10-cffd-4394-b080-0a3fb5ef4b01 Note: This project is currently in development, and I'm actively seeking input to refine and enhance its features.

[P] Building a Code Search Engine for an AI-powered Junior Developer
reddit
LLM Vibe Score0
Human Vibe Score0
williamsweepThis week

[P] Building a Code Search Engine for an AI-powered Junior Developer

The last month building Sweep has been fun. We’ve dealt with countless formatting errors, irrelevant search results, and LLM hallucinations. Sweep is an open source AI-powered junior developer. We take your codebase and provide it as context to GPT to solve small requests related to your code. Code Search Code search is a key part of working with LLMs to automate programming. We used small language models to perform code retrieval(aka semantic search), which comes with several benefits (to be discussed in a later post!). However, one shortcoming of pure semantic search is distinguishing between two similar pieces of code in a vacuum. Example Take the following code snippets: Code Snippet A: accesstoken = os.environ.get("ACCESSTOKEN") g = Github(access_token) repo_name = "sweepai/bot-internal" issue_url = "github.com/sweepai/bot-internal/issues/28" username = "wwzeng1" repo_description = "A repo for Sweep" title = "Sweep: Use loguru.info to show the number of tokens in the anthropic call" summary = "" replies_text = "" Code Snippet B: g = getgithubclient(installation_id) if comment_id: logger.info(f"Replying to comment {comment_id}...") logger.info(f"Getting repo {repofullname}") repo = g.getrepo(repofull_name) currentissue = repo.getissue(number=issue_number) if current_issue.state == 'closed': posthog.capture(username, "issue_closed", properties=metadata) return {"success": False, "reason": "Issue is closed"} Explanation It might not be clear which file is more important, but Code Snippet A is from test\pr\diffs.py#L63-L71 (a test I wrote that’s no longer used), while B is from on\ticket.py#L87-L96 (our core logic for handling tickets). Since Code Snippet B is in an often used file, it is likely that this snippet will be more relevant as input to the LLM. Problem How can we differentiate between these two pieces of code when they’re both so similar? They both discuss issues, repositories, and some usernames. If the user asks “How can I change the username when creating an issue” it will be hard to differentiate between these two. Solution The trick is a ranking model. An important piece of ranking results is the concept of “quality”, i.e. what makes a file or snippet of code intrinsically valuable to the user. The results from our vector search model are a list of items (test\pr\diffs.py#L63-L71, on\ticket.py#L87C1-L96C63) and similarity scores (0.65, 0.63). By combining intuition and attention to the data, we can create a ranking model that is “personalized” for each repository we onboard. Ideas File Length Up to a point, longer files are generally more valuable for search. A 20-line file is probably not valuable unless the user specifically asks for it. However, 2000-line config files should not be ranked much higher either. linecountscore = min(line_count / 20, 10) Number of Commits The more commits a file has, the more valuable it is. This lets us distinguish between one off tests and core logic (which should receive the majority of commits). commitscore = numcommits + 1 Recency of changes The more recently a file was modified, the better. recencyscore = hourssincelastmodified + 1 Scoring To get the final score, we normalize and multiply these three scores together and add the similarity score. qualityscore = linecountscore * commitscore / recency_score finalscore = qualityscore/max(qualityscore) + similarityscore This solution usually worked fine, but we saw the same unexpected files showing up often. The max normalization was not enough. We fixed this by squashing the scores into percentiles, and then capping the increase at .25. In this case, the best result gets a .25 boost and the worst gets no boost. This lets us avoid fetching tests and configs which seem similar, and instead fetch business logic that actually helps Sweep write code! Sweep GitHub If this was interesting, take a look through our github repo (and give it a star!).https://github.com/sweepai/sweep

Looking for a Developer Co-Founder to Build an AI-Powered Film Budgeting Tool
reddit
LLM Vibe Score0
Human Vibe Score1
Boring_Elephant2767This week

Looking for a Developer Co-Founder to Build an AI-Powered Film Budgeting Tool

Hey everyone, I’m a seasoned producer/line producer with over 10 years in the film industry, specializing in budgeting and production strategy for films, commercials, and music videos. I’ve built over 150 budgets for projects ranging from indie features to large-scale commercials and have worked with major artists, brands, and studios. I’m looking for a developer or AI/ML engineer interested in co-founding a startup with me to build an AI-powered budgeting tool for the film industry. The Problem Creating a budget for a film, music video, or commercial is time-consuming and expensive (typically $3K–$5K per budget for films). Filmmakers, studios, agencies, and managers need a faster, more cost-effective way to estimate production costs without hiring a full-time producer for every project. The Solution The goal is to develop an AI-assisted budgeting tool that takes in scripts, creative decks, or project briefs and outputs a preliminary budget & production schedule. The vision is a hybrid service: • AI-powered script/deck breakdown to extract production elements • Smart reasoning based on real industry budgets • Producer oversight for accuracy before sending budgets to users • Flexible pricing model (lower cost than hiring a full-time producer) What I Bring to the Table Deep industry knowledge – I know how to build accurate budgets & schedules for any type of project. Proven demand – I already have early adopters in indie film, production companies, and agencies. Strong network – I work with studios, reps, and filmmakers who would use this tool. A unique approach – I haven’t seen an AI budgeting tool that truly understands production costs based on creative elements. What I’m Looking For I need a developer partner with experience in AI, automation, and/or SaaS development who can help build this. Ideally, someone interested in co-founding (equity-based, not just a freelance gig). If you have experience with GPT, machine learning, NLP, or building interactive SaaS products, that’s a plus. I’m keeping this low-key for now while I figure out the best path forward. If you’re interested, let’s chat! Even if you’re not a developer but have advice or ideas, I’d love to hear your thoughts. Drop a comment or DM me if this sounds interesting!

What are your thoughts on this AI-Powered Interest Rate Negotiation Service Business Model?
reddit
LLM Vibe Score0
Human Vibe Score1
Background_Value_610This week

What are your thoughts on this AI-Powered Interest Rate Negotiation Service Business Model?

Value Proposition: Helps homebuyers secure the best mortgage rates through AI-driven negotiation. Saves time and effort by automating communication with multiple lenders. Increases chances of approval at a favorable rate. Customer Segments: First-time homebuyers Homeowners refinancing their mortgages Investors seeking lower interest rates Revenue Streams: Subscription-based model (monthly/one-time fee for AI-powered negotiation) Success-based fee (small percentage of interest savings upon approval) Affiliate commissions from mortgage lenders for closed deals Channels: Website with a step-by-step AI-powered negotiation tool API integration with mortgage marketplaces Email and social media marketing targeting homebuyers Customer Relationships: AI-powered chatbot and live support for users Automated email sequences keeping users informed Personalized mortgage rate tracking & negotiation updates Key Activities: Developing AI models for lender negotiation Automating email and lender response handling Expanding partnerships with mortgage providers Key Resources: AI/ML engineers to refine the negotiation model CRM system for tracking lender-client interactions Email automation and lead generation tools Key Partners: Mortgage lenders willing to negotiate rates AI-powered email automation services Real estate and mortgage brokers Cost Structure: AI model training and maintenance Web platform hosting and development Compliance and legal expenses

AI-Powered Tool to Detect and Mask PII in Documents
reddit
LLM Vibe Score0
Human Vibe Score1
hoa_nguyen95This week

AI-Powered Tool to Detect and Mask PII in Documents

Hi, everyone! 👋 I’ve been working on an idea for an application that could be a game-changer for data privacy and compliance: The Concept Imagine an app where users can upload a PDF or DOC/DOCX file, and with the power of AI, it scans the document for Personally Identifiable Information (PII). Once identified, the app automatically masks (censors) all the PII and generates a new, sanitized version of the document. Why This Matters In today’s data-driven world, sharing documents is routine, but protecting sensitive information is critical. Businesses, freelancers, and even everyday users often need to redact PII for privacy reasons or compliance with regulations like GDPR, HIPAA, or CCPA. Current Challenges: Manual redaction is time-consuming and error-prone. The Solution: This app ensures quick, accurate, and automated PII redaction, saving time while enhancing data security. Potential Features File Support: PDF, DOC/DOCX, and maybe more formats in the future. AI-Powered Detection: Identify PII such as names, addresses, phone numbers, SSNs, and email addresses. Customization: Users could define additional sensitive terms to be masked. Audit Logs: For compliance, generate a report of what was redacted. Integration: Plug into cloud storage services like Google Drive or Dropbox for seamless workflows. My Questions for the Community Use Cases: What industries or professionals do you think would benefit most from this? Features: Are there additional features or considerations I’m overlooking? Competition: Do you know of similar tools already on the market, and how could this app differentiate itself? Challenges: What technical or market challenges should I anticipate when building and launching this product? I’d love to hear your thoughts, feedback, or ideas for collaboration. If you’re interested in discussing this further, let me know! Thanks in advance for your time and input.

What are your thoughts on this AI-Powered Interest Rate Negotiation Service Business Model?
reddit
LLM Vibe Score0
Human Vibe Score1
Background_Value_610This week

What are your thoughts on this AI-Powered Interest Rate Negotiation Service Business Model?

Value Proposition: Helps homebuyers secure the best mortgage rates through AI-driven negotiation. Saves time and effort by automating communication with multiple lenders. Increases chances of approval at a favorable rate. Customer Segments: First-time homebuyers Homeowners refinancing their mortgages Investors seeking lower interest rates Revenue Streams: Subscription-based model (monthly/one-time fee for AI-powered negotiation) Success-based fee (small percentage of interest savings upon approval) Affiliate commissions from mortgage lenders for closed deals Channels: Website with a step-by-step AI-powered negotiation tool API integration with mortgage marketplaces Email and social media marketing targeting homebuyers Customer Relationships: AI-powered chatbot and live support for users Automated email sequences keeping users informed Personalized mortgage rate tracking & negotiation updates Key Activities: Developing AI models for lender negotiation Automating email and lender response handling Expanding partnerships with mortgage providers Key Resources: AI/ML engineers to refine the negotiation model CRM system for tracking lender-client interactions Email automation and lead generation tools Key Partners: Mortgage lenders willing to negotiate rates AI-powered email automation services Real estate and mortgage brokers Cost Structure: AI model training and maintenance Web platform hosting and development Compliance and legal expenses

AI-Powered Tool to Detect and Mask PII in Documents
reddit
LLM Vibe Score0
Human Vibe Score1
hoa_nguyen95This week

AI-Powered Tool to Detect and Mask PII in Documents

Hi, everyone! 👋 I’ve been working on an idea for an application that could be a game-changer for data privacy and compliance: The Concept Imagine an app where users can upload a PDF or DOC/DOCX file, and with the power of AI, it scans the document for Personally Identifiable Information (PII). Once identified, the app automatically masks (censors) all the PII and generates a new, sanitized version of the document. Why This Matters In today’s data-driven world, sharing documents is routine, but protecting sensitive information is critical. Businesses, freelancers, and even everyday users often need to redact PII for privacy reasons or compliance with regulations like GDPR, HIPAA, or CCPA. Current Challenges: Manual redaction is time-consuming and error-prone. The Solution: This app ensures quick, accurate, and automated PII redaction, saving time while enhancing data security. Potential Features File Support: PDF, DOC/DOCX, and maybe more formats in the future. AI-Powered Detection: Identify PII such as names, addresses, phone numbers, SSNs, and email addresses. Customization: Users could define additional sensitive terms to be masked. Audit Logs: For compliance, generate a report of what was redacted. Integration: Plug into cloud storage services like Google Drive or Dropbox for seamless workflows. My Questions for the Community Use Cases: What industries or professionals do you think would benefit most from this? Features: Are there additional features or considerations I’m overlooking? Competition: Do you know of similar tools already on the market, and how could this app differentiate itself? Challenges: What technical or market challenges should I anticipate when building and launching this product? I’d love to hear your thoughts, feedback, or ideas for collaboration. If you’re interested in discussing this further, let me know! Thanks in advance for your time and input.

AI-Powered Business Analyst Tool Looking for Feedback
reddit
LLM Vibe Score0
Human Vibe Score1
ondro949This week

AI-Powered Business Analyst Tool Looking for Feedback

Hey r/sideproject! I’m excited to share a project I’ve been working on called Bianalytiq, a next-gen business intelligence platform designed to transform the way businesses interact with data through the power of AI. The Problem: SME companies struggle with data overload and the significant time investment required to generate actionable insights. Traditional data analysis methods are not only slow but often require extensive manual effort and are prone to errors. This makes it difficult for businesses to react quickly to new information and make informed decisions efficiently. Not everybody can write SQL or create/understand data dashboards.... AND - one big opportunity on market - non of the AI tools available on market offer reusable contexts focused on you as a company and your products. The Solution: Bianalytiq aims to solve these issues by automating tedious data analysis tasks and providing real-time insights. Here’s how: Reusable contexts: Let Bianalytiq learn everything about your company, your products, business model etc. - your company is your unique context. Autonomous AI Agents: Deploy AI agents that not only react to queries but proactively analyze data to uncover opportunities, tailored specifically to your business context. Real-Time Insights: With the use of Retrieval-Augmented Generation (RAG) technology, our platform delivers immediate, context-rich insights by dynamically accessing and analyzing connected databases and data warehouses. Integration with Existing Tools: Bianalytiq integrates seamlessly with popular tech stacks and communication platforms like Slack and Microsoft Teams, making it incredibly user-friendly and reducing the switch cost between applications. Why I’m Here: Before investing significant time and money I want to validate the product first and do pre-sale before releasing the MVP. I’ve developed a landing page for Bianalytiq and would love your feedback on both the service itself and the effectiveness of the landing page. Are the features presented clearly? Does the platform address the pain points you might experience in data analysis and decision-making processes? Here’s the link to the landing page: https://bianalytiq.com/ I appreciate any feedback or questions you have! Whether it's about the UI/UX of the site, the technical aspects of the service, or even the business model, I’m all ears. Your input will be invaluable :) Thanks for checking it out! https://preview.redd.it/t1dvp2q05dzc1.png?width=798&format=png&auto=webp&s=c7365b418abfc4d4260d9a23305ed3398e83c87b

Building and launching an AI-powered Product Strategy tool, or; a story of nights and weekends
reddit
LLM Vibe Score0
Human Vibe Score1
_raZeThis week

Building and launching an AI-powered Product Strategy tool, or; a story of nights and weekends

Speaking to peers in the software development sphere I learned of one constant that we had all personally experienced throughout our careers: a bloated product development process that feels like work for the sake of work, centred around the highest-paid person's opinion instead of its customers. We didn't like how current tools assume AI will provide the perfect answer on the first run. Instead, we wanted a tool that allows for manual refining and editing AI suggestions, keeping all previous ideas in context. This way, we can develop a solution step by step, instead of trying to get it perfect on the first try. An approach more similar to how you'd typically approach product discovery as a human. AI is then used to help save time and reduce admin, instead of replace the expert So, we got together and asked over 100 Product Managers questions about it, brought all that feedback goodness together, and started building Squad. We think we've created something really cool and hope you think so too. The ELI5 on what Squad does: 1) Creates alignment that empowers bottom up software development whilst keeping executive in the loop 2) Increases confidence that what you're building is what people actually want - data driven by default 2) Speeds up the time from idea --> execution by ideating with you on an experimentation approach 3) Helps gives PMs time back to focus on strategy (currently stats show they spend 75% of their time on admin, 25% on strategy) The team hustled hard on this as a passion project while working day jobs, and today have launched on Product Hunt. Check it out and see if the mission resonates with you, we'd appreciate the love! https://www.producthunt.com/posts/squad-8b75e29c-d767-4a8f-a60a-fd162e141a72 ​

Seeking Feedback on Business Idea: AI-Powered Business Partner Matching Platform
reddit
LLM Vibe Score0
Human Vibe Score1
torrentialdownpour34This week

Seeking Feedback on Business Idea: AI-Powered Business Partner Matching Platform

Hey everyone, I've been toying with an idea for a new business venture and I'd love to get some feedback and insights from this community. The Idea: I'm considering building a business platform that utilizes AI to match potential business partners. Whether you're a startup looking for a co-founder, a company seeking strategic partnerships, or an investor searching for promising ventures, this platform would help connect you with compatible partners based on your specific needs, goals, and preferences. How It Works: Users would create detailed profiles outlining their business objectives, industry expertise, skills, and what they're looking for in a partner. The AI algorithm would then analyze this data to identify compatible matches, taking into account factors like complementary skill sets, shared values, and mutual goals. The platform would provide users with a curated list of potential partners, along with insights and recommendations to facilitate meaningful connections. Key Features: Comprehensive Profiles: Users can create detailed profiles highlighting their background, experience, and what they bring to the table. AI Matching Algorithm: The platform's AI algorithm would use advanced data analysis techniques to generate accurate partner recommendations. Communication Tools: Built-in messaging and video conferencing tools would enable seamless communication between potential partners. Feedback and Ratings: Users can leave feedback and ratings for their matches, helping to build trust and credibility within the community. Resource Hub: Access to resources, articles, and guides on partnership development, negotiation strategies, and other relevant topics. Why It's Needed: Finding the right business partner can be a daunting task, often relying on personal networks or serendipitous encounters. By harnessing the power of AI, this platform aims to streamline the partner matching process, saving time and increasing the likelihood of finding compatible collaborators. Looking for Feedback: Before diving headfirst into this venture, I wanted to reach out to this community to gather some feedback: Does this idea resonate with you? Why or why not? Are there any existing platforms or services that offer similar functionalities? What features would be essential for you as a user? Any potential challenges or concerns you foresee with this concept? I'm eager to hear your thoughts and suggestions. Thanks in advance for your input!

I built an OCR powered by Mistral AI that extracts text, tables, formulas from docs (20+ languages & JSON output!)
reddit
LLM Vibe Score0
Human Vibe Score0
hhe_kkmThis week

I built an OCR powered by Mistral AI that extracts text, tables, formulas from docs (20+ languages & JSON output!)

Hi everyone 👋 Most OCR tools struggle with complex documents—crumbling tables, garbled formulas, or unstructured text. Need clean data for RAG or apps? Good luck. So I built Mistral OCR (https://www.mistralocr.app/) using Mistral AI’s document understanding models. It doesn’t just scan—it understands the document’s structure, and extracts: ✅ Text (plain/formatted) ✅ Tables (pixel-perfect JSON with headers 🧮) ✅ Math formulas (LaTeX-ready via Mistral’s ML pipeline) ✅ Images (preserved or extracted) Why Mistral AI? Their models nail context-aware parsing—unlike rigid OCRs, Mistral’s tech handles: Cursed PDFs(scanned/watermarked/warped text) Mixed layouts (research papers with tables + formulas) 20+ languages (English, Japanese, Mandarin, Spanish...) Structured JSON output (directly feeds into RAG/APIs) See examples → https://www.mistralocr.app/ Why build this? I needed an OCR that could extract RAG-ready data without regex nightmares. Mistral AI’s models finally made this possible—they preserve relationships between text, tables, and formulas, something traditional OCRs butcher. Who’s using it? Devs automating document workflows Researchers digitizing datasets from papers Teams processing multilingual forms/contracts Anyone frustrated by copying tables from PDFs Challenge me: Send your worst documents (scanned receipts? handwritten tables?) and I’ll run them through Mistral OCR live. Try it here → https://www.mistralocr.app/ Let me know what you think! 🙏 Let me know if bugs🐛!🙏

Vibe Coding For Non Coders - I built an online game in 30 seconds using AI
youtube
LLM Vibe Score0.371
Human Vibe Score0.5
AI BORDERMar 25, 2025

Vibe Coding For Non Coders - I built an online game in 30 seconds using AI

🚀 No coding skills? No problem! In this video, I show you how I built a working online game in just 30 seconds using AI-powered coding tools – perfect for beginners, creators, or anyone curious about AI development. 🔥 Try CodeLLM Teams FREE for 1 Month! 🎁👉 https://chatllm.abacus.ai/jTYLJgzFxy 👨‍💻 About CodeLLM Teams CodeLLM Teams is an advanced AI assistant that helps you write, optimize, and debug code across 10+ programming languages including Python, JavaScript, C++, PHP, and more. It works seamlessly with GitHub and all leading LLMs like Claude Sonnet 3.7, O3 Mini High, Quen, and others. 💻 Whether you're a solo developer or working in a team, CodeLLM makes your workflow faster and more efficient — even if you’ve never written a line of code before! #NoCode #AItools #GameDev #CodeLLM #AbacusAI #VibeCoding #LearnToCode #AIToolsForBeginners #CodingWithoutCode #BuildAGame #LLM #ChatGPT #Claude #GeminiAI #CodingTutorial #NonCoders #aifordevelopers ✨Contact AI Border: composition365@gmail.com✨ The videos use materials in a transformative and educational manner, following fair use guidelines and without any intention of copyright infringement. If you are the copyright owner or representative and have any concerns regarding the material used, please contact me at composition365@gmail.com, and we can address the issue. ✨Here are some more videos to watch 👍 ▶Top Free AI Video Generators: Image-to-Video and Text-to-Video Tools for 2025 https://youtu.be/VNDT2yA6zc0 ▶ Who Is the King of AI Video in 2025? Heygen vs Vozo AI vs Akool (Full Test) https://youtu.be/43up6iNj1wo ▶ GlobalGPT: The Ultimate All-in-One AI Tool for Writing, Proofreading, and Image Generation https://youtu.be/iPcFVC6Xz_8 ▶Uncensored AI Tool: Open Source Mimic PC Revolutionizes Content Creation https://youtu.be/4dvqDXQ09TY ▶AI Text-to-3D Animation: Effortlessly Create 3D Animated Videos from Text Prompts https://youtu.be/wzOCO8NYiLM ▶ Create Stunning Game & Film Concept Art with Shakker AI: AI Art Generation Tutorial https://youtu.be/OFv2CjWfq9U ▶ Create Viral Videos Using the Top AI Image and Video Generator https://youtu.be/1T3PxLdm2VY ▶ This video could help who are looking for: ai game builder,ai coding assistant,no code game development,code with ai,ai coding tutorial,build games with ai,image to game ai,html game with ai,free ai coding tools,how to build games with ai,ai game generator,learn coding with ai,ai tools for beginners,ai game development,ai for non coders,ai project tutorial,abacus ai,codeLLM tutorial,ai programming tools,ai powered coding,ai programming assistant,ai dev tools,build apps with ai,no code ai tools,code generator ai,ai video tutorial, #CodeWithAI #NoCodeTools #AIGameBuilder #AICodingAssistant #CodeLLM #AbacusAI #AIforBeginners #AIProjects #AIDevTools #LearnCodingWithAI #AITools2025 #AICodingTutorial #BuildWithAI #NoCodeDevelopment #AIProgramming #AIpowered #VibeCoding #CodingWithoutCode #CreateWithAI #HTMLGameWithAI #AIWorkflow #AIForEveryone #NonCodersWelcome #ShortVideoMaker #TextToCode #AIGeneratedCode #AIHack #AIForDevelopers #CreativeTools #ArtificialIntelligence #chatgpt #ClaudeAI

Experienced Software Developer looking for startup to help. I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
DB010112This week

Experienced Software Developer looking for startup to help. I will not promote

My passion for programming started at the age of 9 when I began playing video games. It was during this time that I first dived into programming, creating scripts for SA:MP (San Andreas Multiplayer) using the Pawn language. SA:MP is a modification for the popular game Grand Theft Auto: San Andreas, allowing players to experience multiplayer gameplay. My early experiences in programming were all about problem-solving—finding ways to enhance the game and improve the player experience. This was when I realized how satisfying it is to solve a problem through code, and that feeling has stayed with me throughout my career. I am a self-taught programmer, and everything I know today comes from my own initiative to learn and improve. After five years of working with local clients, I decided to expand my knowledge and started learning more widely applicable programming languages like Java and Python. I’ve always been the type of person who thrives on challenges. Whenever I encounter a problem, I don’t just look for a quick fix—I dive deep into researching and understanding the problem, and I find a solution that works in the long run. This is what drives me. The ability to solve problems, no matter how complex, and the satisfaction that comes with it is what fuels my passion for programming. My big break came when I had the opportunity to work at \\\\. There, I replaced two senior and two junior developers, which led to significant cost savings for the company. I completed all tasks ahead of schedule, focusing on Java-based applications that were multithreaded and communicated with embedded systems. This experience taught me how to work under pressure and how to manage and solve complex technical problems efficiently. Following my time at \\\\, I transitioned into freelance work as a FullStack Developer, working with technologies such as HTML, CSS, Bootstrap, JavaScript, Django, Spring, MySQL, and PostgreSQL. As a freelancer, I was responsible for finding solutions to a wide range of problems, often working independently and making decisions on the fly. I learned that self-reliance is key in this industry, and being resourceful is one of the most important qualities a developer can have. Later, I joined \\\\ elecom, where I worked on system integration with foreign teams, BPM process solutions, and the merging of complex systems in Oracle databases. I continued to solve challenges, often working with teams across borders and tackling technical obstacles that required creative and well-thought-out solutions. Eventually, I founded my own company, \\\\, where I focus on developing software solutions, Artificial Intelligence (AI), Cybersecurity, and Ethical Hacking. As an entrepreneur, I take pride in finding innovative solutions to problems, whether they come from clients or from technical obstacles I encounter along the way. I’ve also had the privilege of working with the Serbian Ministry of Defense and the police, handling sensitive projects that demand both technical expertise and trustworthiness. Being a self-taught programmer means that I have had to learn and adapt on my own, and I’ve learned to embrace challenges as opportunities for growth. I am constantly driven by the process of solving problems, and it is what keeps me engaged and fulfilled in my work. I am always open to new collaborations and am eager to take on new challenges that push my boundaries in technology, cybersecurity, and software development.

Hot Take: Not all your startups need AI forced into them
reddit
LLM Vibe Score0
Human Vibe Score1
bitorsicThis week

Hot Take: Not all your startups need AI forced into them

I'm a final year Computer Engineering student, hence applying for jobs all around. There's this particular trend I've noticed with startups that are coming up these days. That is, even for the absolute basic stuff they'll use 'AI', and they'll think they built something 'revolutionary'. No. You're breaking your product in ways you don't realise. An example, that even some well established companies are guilty of: AI Chatbots You absolutely don't need them and it's an entire gimmick. If you really wanna implement a chatbot, connect the user to an actual person on your end, which I think is not possible if you're at a 'startup' stage. You'll need employees who can handle user queries in real time. If the user really is stuck let them use the 'Contact Us' page. A really close relative of mine is very vocal about the frustration he faces whenever he tries to use the AI Chatbot on any well known e-com website. The only case for AI Chatbot that makes sense is when it's directing the customer to an actual customer support rep if none of the AI's solutions is working for the customer. Even then, implementing a search page for FAQ is extremely easy and user friendly. Another example: AI Interviewer I recently interviewed for a startup, and their whole interviewing process was AI'zed?!?! No real person at the other end, I was answering to their questions which were in video format. They even had a 'mascot' / 'AI interviewer' avatar designed by an AI (AI-ception???). This mascot just text-to-speech'ed all the questions for me to rewind and hear what I missed again. And I had to record video and audio to answer these questions on their platform itself. The entire interview process just could've been a questionnaire, or if you're really concerned on the integrity of the interviewee, just take a few minutes out of your oh-so-busy schedule as a startup owner. Atleast for hiring employees who would make the most impact on your product going ahead. I say the most impact, because (atleast as a developer) the work done by these employees would define how robust your product is, and/or how easily other features can be integrated into the codebase. Trust me, refactoring code later on would only cost you time and money. These resources would rather be more useful in other departments of your startup. The only use case for an AI Interviewer I see is for preparing for an actual interview, provided that feedback is given to the user at the earliest, which you don't need to worry about as a startup owner. So yeah, you're probably better off without integrating AI in your product. Thank you for reading. TLDR; The title; I know AI is the new thing and gets everyone drooling and all, but for the love of God, just focus on what your startup does best and put real people behind it; Integrating AI without human intervention is as good as a broken product; Do your hiring yourself, or through real people, emphasizing on the fact that the people you hire at an early stage will define your growth ahead;

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

Anyone finding that they just don't NEED to add more Employees anymore? (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score0.6
wilschroterThis week

Anyone finding that they just don't NEED to add more Employees anymore? (I will not promote)

A friend of mine who was looking for work asked me if we were hiring and I responded "You know, it's weird but all of our growth goals don't seem to map back to hiring people anymore." This isn't about the economy or growth goals. It's a really fascinating shift in focus and costs for startups. My gut reaction is that I HATE the idea of not creating more jobs. In my career I've hired thousands of people, and I've always prided myself on job creation. We just sold a company that employed 200 people last year, and I'm proud of the work we were able to create. What's interesting is that I simply don't feel like we NEED to like we used to. As we're looking at all of our growth goals, for the first time I'm not assigning FTEs to them. Nearly everything we're doing is actually reducing the need for more humans, not adding them - and we're not even trying to reduce the need. Obviously the timing of AI has had a major impact. Product - Our team is shipping more code than ever before, and even our designers who have never touched code are shipping final code. If we doubled the size of the team, it would make no difference (this is a big deal considering the historical cost here). Marketing - So many aspects of our marketing are getting automated and streamlined, to the point where even a single FTE can create a massive amount of reach across channels. Support - Our Success team is able to effectively respond to tickets in a fraction of the time, which essentially doubles their capacity without adding any more staff. Management - With less staff we need less managers, which are a big expense, but it also means reporting and decisions are more streamlined, which is a positive. But it also means those positions simply don't get created like they used to. I think this is a big deal for the younger startups because it translates into needing less capital (or none!) which provides for more ownership and agency. Clearly we still need some folks to build out the core team, but that's very different than a massive staffing line item. Anyone else here finding the same trend? Opposite? I don't have a strong opinion either way, but I'd love to hear how other Founders are processing this.

How I made a high tech salary in my first selling month
reddit
LLM Vibe Score0
Human Vibe Score1
Ok_Negotiation_2587This week

How I made a high tech salary in my first selling month

For over 7 years I worked as a full-stack developer, helping other companies bring their ideas to life. But one day, I thought “Why not try making my own dream come true?”. That’s when I decided to quit my job and start my own journey to becoming an entrepreneur. At first, it wasn’t easy. I didn’t make any money for months and had no idea where to start. I felt lost. Then, I decided to focus on something popular and trending. AI was everywhere, and ChatGPT was the most used AI platform. So I looked into it and I found the OpenAI community forum where people had been asking for features that weren’t being added. That gave me an idea. Why not build those features myself? I created a Chrome extension and I worked on some of the most requested features, like: Downloading the advanced voice mode and messages as MP3 Adding folders to organize chats Saving and reusing prompts Pinning important chats Exporting chats to TXT/JSON files Deleting or archiving multiple chats at once Making chat history searches faster and better It took me about a week to build the first version, and when I published it, the response was incredible. People loved it! Some even said things like, “You’re a lifesaver!” That’s when I realized I had something that could not only help people but also turn into a real business. I kept the first version free to see how people would respond. Many users have been downloading my extension, which prompted Chrome to review it to determine if it qualified for the featured badge. I received the badge, and it has significantly boosted traffic to my extension ever since. After all the positive feedback, I launched a paid version one month ago. A few minutes after publishing it, I made my first sale! That moment was so exciting, and it motivated me to keep going. I already have over 4,000 users and have made more than $4,500 in my first selling month. I’ve decided to release 1-2 new features every month to keep improving the extension based on what users ask for. I also created the same extension for Firefox and Edge users because many people have been asking for it! I also started a Reddit community, where I share updates, sales, discount codes, and ideas for new features. It’s been awesome to connect with users directly and get their feedback. Additionally, I’ve started working on another extension for Claude, which I’m hoping will be as successful as this one. My message to you is this: never give up on your dreams. It might feel impossible at first, but with patience, hard work, and some creativity, you can make it happen. I hope this inspires you to go after what you want. Good luck to all of us!

The Drawing of the Three - Once you look through the veil, nothing is the same again. (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score0.333
Tim-SylvesterThis week

The Drawing of the Three - Once you look through the veil, nothing is the same again. (I will not promote)

Originally published Nov 5, 2024 In my last post, I talked about assembling a series of filters to use to view the startup landscape, which led me to a few conclusions about what opportunities I should pursue. What did I see through those filters? What I saw through the moire pattern of those two lists overlaid by one another is what I think will be the third great monetization strategy for the internet, matching the pattern of: web1 => Ad monetization web2 => Subscription monetization web3 => For AI, neither of those work anymore, which demands something new. But what? Well that’s the important part, isn’t it? Should I just up and tell you? Yawn. The climax of a movie is at the climax, if they tell you the crux at the beginning, it’s a lot less fun (usually). The standard bearer for web1 and ads was Google (with countless followers), and essentially every website adopted that model for their first pass at content monetization. Google has been… let’s call it fairly successful… so it’s not a bad way to look at things. How many websites live and die by selling advertising? The standard bearers for web2 and subscriptions were Salesforce (for B2B SaaS) and Netflix (for B2C SaaS), with countless followers, to the extent that SaaS has been the dominant startup monetization thesis for the last 15+ years. It’s more old and tired by now than most American politicians, but how many websites live and die by people entering payment details for a monthly or annual subscription? Evidence proves those models for web1 and web2 worked well enough that countless businesses depend on them, and countless fortunes have been made and lost surfing the waves, or crashing against the shorelines, of ads and subs. But it’s also apparent (to me, at least) that now that AI is the dominant startup thesis, neither ads nor subs are going to prevail in an AI-centered world, and for one simple reason: Those monetization strategies are for humans, and AI bots are not humans. Changing Environments Require Changing Strategies Every so often, there’s a fundamental shift that demands everything in the ecosystem adapt to a new habitation strategy to survive. We’ve seen this repeatedly across Earth’s ecology (for instance, introducing free oxygen to the atmosphere, producing respiration while destroying all the life forms that existed before oxygen permeated the atmosphere), and across human society (for example, how nuclear bombs changed war, and how drones are changing it again, for less violent examples, consider the adoption of computers and the subsequent adoption of smartphones). Now the ecosystem of the internet has changed irrevocably, opening up countless new and interesting niches to occupy. Humans may see an ad and buy something stupid (or, occasionally, not-stupid), but an AI won’t unless its programmed to. And subscriptions are designed for humans to consume content at a human rate, not for an AI that can choke down an entire database of content (whatever it may be) at whatever speed the servers can manage. Changing conditions require changing strategies. It was clear to me that: The introduction of AI bots to the internet ecosystem was, is, and will be massively disruptive for a very long time The internet population of bots already exceeds humans and is growing faster than the human population The two dominant monetization strategies are not relevant to bots That disruption of expectations across the ecosystem demands a third strategy, a new strategy to handle a massive change in an existing system. And that strategy needs to accommodate, support, and monetize the new demands from the vast armies of new participants in the internet ecology. Therefore, a method that converts bots from an expense into a revenue source would become a dominant monetization strategy, and therefore whoever owns that strategy will be a dominant player in the internet ecosystem. Set the realization of semi-practical, semi-useful AI against a backdrop of technology cycles that have, in the distant past (in internet terms) produced ads and subs, and more recently produced enormous investment into fintech and crypto, I started to see a path that felt like it would grow over time to become a new monetization strategy that works in the AI ecosystem. Sun Tzu had a couple drinks, saw a couple things… There’s at least, and possibly only, two things I know about fighting: You cannot fight the tide, and it’s much harder to fight an uphill battle. If my whole thesis on this go-around was to go with the flow, and that trickle of insight was leading me from my overlook along a roaring flow of cash coursing through a valley filled with AI startups, where exactly would it lead me? Most rivers lead to the sea eventually, but they can take winding paths, and sometimes the quickest route from the mountain to the sea isn’t to follow the river, but to understand where the river leads and go there instead. Getting a view from on high can save you a lot of time on your journey. But before I get to where the path has led (or is leading) that will explain the objective I’ve identified, and the deliverables I have to produce to reach it, let’s talk about a few of the steps on the path I’ve been taking that highlight the process I followed. I figure if I explain the steps I’m taking, as I’m taking them, it may be easier for people who haven’t trod this route before to follow me and understand how to carve their own course towards their own objectives. And maybe the real treasure will be the friends we make along the way. (I will not promote)

160 of Y Combinators 229 Startup Cohort are AI Startups with and 75% of the Cohort has 0 revenue
reddit
LLM Vibe Score0
Human Vibe Score1
DemocratizingfinanceThis week

160 of Y Combinators 229 Startup Cohort are AI Startups with and 75% of the Cohort has 0 revenue

Y Combinator (YC), one of the most prestigious startup accelerators in the world, has just unveiled its latest batch of innovative startups, providing key insights into what the future might hold. Y Combinators Summer 2023 Batch In a recent post by Garry Tan, YC's president, Tan offers a nostalgic look back at his first YC Demo Day in 2008, where he, as a budding entrepreneur, pitched his startup. Now, fifteen years later, he's at the helm, proudly launching the 37th Demo Day, this time for the Summer 2023 batch. Tan proudly declares this batch as one of YC's most impressive yet, emphasizing the deep technical talent of the participants. From a staggering pool of over 24,000 applications, only 229 startups were chosen, making this one of the most competitive batches to date. This batch marks a number of firsts and solidifies several rising trends within the startups landscape. 75% of these companies began their YC journey with zero revenue, and 81% hadn't raised any funding before joining the accelerator. YC's decision to focus on early-stage startups this round signals their commitment to nurturing raw, untapped potential. A Return to Face-to-Face Interaction After three years, YC has brought back the in-person Demo Day format, allowing startups, investors, and mentors to connect directly. While the virtual format has its merits, there's an unmistakable magic in the YC Demo Day room, filled with anticipation, hope, and innovation. AI Takes Center Stage Artificial Intelligence is the standout sector in the Summer 2023 batch. With recent advancements making waves across various industries, there's arguably no better time to launch an AI-focused startup, and no better platform than YC to foster its growth. This signals a clear trend in the startup investing and venture capital space: AI is just getting started. Of the entire Summer 2023 batch, 160 out of the entire 229 Summer 2023 batch that are utilizing or implementing artificial intelligence in some capacity. This means over 2 out of every 3 startups accepted is focused on artificial intelligence in some capacity. Some of the startups include: Quill AI: Automating the job of a financial analyst Fiber AI: Automating prospecting and outbound marketing Reworkd AI: Open Source Zapier of AI Agents Watto AI: AI-powered McKinsey-quality reports in seconds Agentive: AI-powered auditing platform Humanlike: Replace your call center with voice bots that sound human Greenlite: AI compliance team for fintech and banking atla: AI assistants to help in-house lawyers answer legal questions Studdy: An AI Match tutor Glade: League of Legends with AI-generated maps and gameplay and literally over 100 others. As you can see, there's a startup covering nearly every sector of AI in the new batch. YC By The Numbers YC continues to grow as a community. The accelerator now boasts over 10,000 founders spanning more than 4,500 startups. The success stories are impressive: over 350 startups valued at over $150 million and 90 valued at more than $1 billion. The unicorn creation rate of 5% is truly unparalleled in the industry. To cater to the ever-growing community, YC has added more full-time Group Partners than ever. This includes industry veterans such as Tom Blomfield, co-founder of billion-dollar startups GoCardless and Monzo, and YC alumni like Wayne Crosby (Zenter) and Emmett Shear (Twitch). YC Core Values YC's commitment to diversity is evident in the demographics of the S23 batch. They've also spotlighted the industries these startups operate in, with 70% in B2B SaaS/Enterprise, followed by fintech, healthcare, consumer, and proptech/industrials. Garry Tan emphasizes three core tenets for YC investors: to act ethically, to make decisions swiftly, and to commit long-term. He underlines the importance of the YC community, urging investors to provide valuable introductions and guidance to founders. The Road Ahead With YC's track record and the promise shown by the Summer 2023 batch, the future of the startup ecosystem looks promising. As always, YC remains at the forefront, championing innovation and shaping the next generation of global startups. Original Post: https://www.democratizing.finance/post/take-a-peek-into-the-future-with-y-combinators-finalized-summer-2023-batch

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies) (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies) (I will not promote)

AI Palette is an AI-driven platform that helps food and beverage companies predict emerging product trends. I had the opportunity recently to sit down with the founder to get his advice on building an AI-first startup, which he'll be going through in this post. (I will not promote) About AI Palette: Co-founders: >!2 (Somsubhra GanChoudhuri, Himanshu Upreti)!!100+!!$12.7M USD!!AI-powered predictive analytics for the CPG (Consumer Packaged Goods) industry!!Signed first paying customer in the first year!!65+ global brands, including Cargill, Diageo, Ajinomoto, Symrise, Mondelez, and L’Oréal, use AI Palette!!Every new product launched has secured a paying client within months!!Expanded into Beauty & Personal Care (BPC), onboarding one of India’s largest BPC companies within weeks!!Launched multiple new product lines in the last two years, creating a unified suite for brand innovation!Identify the pain points in your industry for ideas* When I was working in the flavour and fragrance industry, I noticed a major issue CPG companies faced: launching a product took at least one to two years. For instance, if a company decided today to launch a new juice, it wouldn’t hit the market until 2027. This long timeline made it difficult to stay relevant and on top of trends. Another big problem I noticed was that companies relied heavily on market research to determine what products to launch. While this might work for current consumer preferences, it was highly inefficient since the product wouldn’t actually reach the market for several years. By the time the product launched, the consumer trends had already shifted, making that research outdated. That’s where AI can play a crucial role. Instead of looking at what consumers like today, we realised that companies should use AI to predict what they will want next. This allows businesses to create products that are ahead of the curve. Right now, the failure rate for new product launches is alarmingly high, with 8 out of 10 products failing. By leveraging AI, companies can avoid wasting resources on products that won’t succeed, leading to better, more successful launches. Start by talking to as many industry experts as possible to identify the real problems When we first had the idea for AI Palette, it was just a hunch, a gut feeling—we had no idea whether people would actually pay for it. To validate the idea, we reached out to as many people as we could within the industry. Since our focus area was all about consumer insights, we spoke to professionals in the CPG sector, particularly those in the insights departments of CPG companies. Through these early conversations, we began to see a common pattern emerge and identified the exact problem we wanted to solve. Don’t tell people what you’re building—listen to their frustrations and challenges first. Going into these early customer conversations, our goal was to listen and understand their challenges without telling them what we were trying to build. This is crucial as it ensures that you can gather as much data about the problem to truly understand it and that you aren't biasing their answers by showing your solution. This process helped us in two key ways: First, it validated that there was a real problem in the industry through the number of people who spoke about experiencing the same problem. Second, it allowed us to understand the exact scale and depth of the problem—e.g., how much money companies were spending on consumer research, what kind of tools they were currently using, etc. Narrow down your focus to a small, actionable area to solve initially. Once we were certain that there was a clear problem worth solving, we didn’t try to tackle everything at once. As a small team of two people, we started by focusing on a specific area of the problem—something big enough to matter but small enough for us to handle. Then, we approached customers with a potential solution and asked them for feedback. We learnt that our solution seemed promising, but we wanted to validate it further. If customers are willing to pay you for the solution, it’s a strong validation signal for market demand. One of our early customer interviewees even asked us to deliver the solution, which we did manually at first. We used machine learning models to analyse the data and presented the results in a slide deck. They paid us for the work, which was a critical moment. It meant we had something with real potential, and we had customers willing to pay us before we had even built the full product. This was the key validation that we needed. By the time we were ready to build the product, we had already gathered crucial insights from our early customers. We understood the specific information they wanted and how they wanted the results to be presented. This input was invaluable in shaping the development of our final product. Building & Product Development Start with a simple concept/design to validate with customers before building When we realised the problem and solution, we began by designing the product, but not by jumping straight into coding. Instead, we created wireframes and user interfaces using tools like InVision and Figma. This allowed us to visually represent the product without the need for backend or frontend development at first. The goal was to showcase how the product would look and feel, helping potential customers understand its value before we even started building. We showed these designs to potential customers and asked for feedback. Would they want to buy this product? Would they pay for it? We didn’t dive into actual development until we found a customer willing to pay a significant amount for the solution. This approach helped us ensure we were on the right track and didn’t waste time or resources building something customers didn’t actually want. Deliver your solution using a manual consulting approach before developing an automated product Initially, we solved problems for customers in a more "consulting" manner, delivering insights manually. Recall how I mentioned that when one of our early customer interviewees asked us to deliver the solution, we initially did it manually by using machine learning models to analyse the data and presenting the results to them in a slide deck. This works for the initial stages of validating your solution, as you don't want to invest too much time into building a full-blown MVP before understanding the exact features and functionalities that your users want. However, after confirming that customers were willing to pay for what we provided, we moved forward with actual product development. This shift from a manual service to product development was key to scaling in a sustainable manner, as our building was guided by real-world feedback and insights rather than intuition. Let ongoing customer feedback drive iteration and the product roadmap Once we built the first version of the product, it was basic, solving only one problem. But as we worked closely with customers, they requested additional features and functionalities to make it more useful. As a result, we continued to evolve the product to handle more complex use cases, gradually developing new modules based on customer feedback. Product development is a continuous process. Our early customers pushed us to expand features and modules, from solving just 20% of their problems to tackling 50–60% of their needs. These demands shaped our product roadmap and guided the development of new features, ultimately resulting in a more complete solution. Revenue and user numbers are key metrics for assessing product-market fit. However, critical mass varies across industries Product-market fit (PMF) can often be gauged by looking at the size of your revenue and the number of customers you're serving. Once you've reached a certain critical mass of customers, you can usually tell that you're starting to hit product-market fit. However, this critical mass varies by industry and the type of customers you're targeting. For example, if you're building an app for a broad consumer market, you may need thousands of users. But for enterprise software, product-market fit may be reached with just a few dozen key customers. Compare customer engagement and retention with other available solutions on the market for product-market fit Revenue and the number of customers alone isn't always enough to determine if you're reaching product-market fit. The type of customer and the use case for your product also matter. The level of engagement with your product—how much time users are spending on the platform—is also an important metric to track. The more time they spend, the more likely it is that your product is meeting a crucial need. Another way to evaluate product-market fit is by assessing retention, i.e whether users are returning to your platform and relying on it consistently, as compared to other solutions available. That's another key indication that your solution is gaining traction in the market. Business Model & Monetisation Prioritise scalability Initially, we started with a consulting-type model where we tailor-made specific solutions for each customer use-case we encountered and delivered the CPG insights manually, but we soon realized that this wasn't scalable. The problem with consulting is that you need to do the same work repeatedly for every new project, which requires a large team to handle the workload. That is not how you sustain a high-growth startup. To solve this, we focused on building a product that would address the most common problems faced by our customers. Once built, this product could be sold to thousands of customers without significant overheads, making the business scalable. With this in mind, we decided on a SaaS (Software as a Service) business model. The benefit of SaaS is that once you create the software, you can sell it to many customers without adding extra overhead. This results in a business with higher margins, where the same product can serve many customers simultaneously, making it much more efficient than the consulting model. Adopt a predictable, simplistic business model for efficiency. Look to industry practices for guidance When it came to monetisation, we considered the needs of our CPG customers, who I knew from experience were already accustomed to paying annual subscriptions for sales databases and other software services. We decided to adopt the same model and charge our customers an annual upfront fee. This model worked well for our target market, aligning with industry standards and ensuring stable, recurring revenue. Moreover, our target CPG customers were already used to this business model and didn't have to choose from a huge variety of payment options, making closing sales a straightforward and efficient process. Marketing & Sales Educate the market to position yourself as a thought leader When we started, AI was not widely understood, especially in the CPG industry. We had to create awareness around both AI and its potential value. Our strategy focused on educating potential users and customers about AI, its relevance, and why they should invest in it. This education was crucial to the success of our marketing efforts. To establish credibility, we adopted a thought leadership approach. We wrote blogs on the importance of AI and how it could solve problems for CPG companies. We also participated in events and conferences to demonstrate our expertise in applying AI to the industry. This helped us build our brand and reputation as leaders in the AI space for CPG, and word-of-mouth spread as customers recognized us as the go-to company for AI solutions. It’s tempting for startups to offer products for free in the hopes of gaining early traction with customers, but this approach doesn't work in the long run. Free offerings don’t establish the value of your product, and customers may not take them seriously. You should always charge for pilots, even if the fee is minimal, to ensure that the customer is serious about potentially working with you, and that they are committed and engaged with the product. Pilots/POCs/Demos should aim to give a "flavour" of what you can deliver A paid pilot/POC trial also gives you the opportunity to provide a “flavour” of what your product can deliver, helping to build confidence and trust with the client. It allows customers to experience a detailed preview of what your product can do, which builds anticipation and desire for the full functionality. During this phase, ensure your product is built to give them a taste of the value you can provide, which sets the stage for a broader, more impactful adoption down the line. Fundraising & Financial Management Leverage PR to generate inbound interest from VCs When it comes to fundraising, our approach was fairly traditional—we reached out to VCs and used connections from existing investors to make introductions. However, looking back, one thing that really helped us build momentum during our fundraising process was getting featured in Tech in Asia. This wasn’t planned; it just so happened that Tech in Asia was doing a series on AI startups in Southeast Asia and they reached out to us for an article. During the interview, they asked if we were fundraising, and we mentioned that we were. As a result, several VCs we hadn’t yet contacted reached out to us. This inbound interest was incredibly valuable, and we found it far more effective than our outbound efforts. So, if you can, try to generate some PR attention—it can help create inbound interest from VCs, and that interest is typically much stronger and more promising than any outbound strategies because they've gone out of their way to reach out to you. Be well-prepared and deliberate about fundraising. Keep trying and don't lose heart When pitching to VCs, it’s crucial to be thoroughly prepared, as you typically only get one shot at making an impression. If you mess up, it’s unlikely they’ll give you a second chance. You need to have key metrics at your fingertips, especially if you're running a SaaS company. Be ready to answer questions like: What’s your retention rate? What are your projections for the year? How much will you close? What’s your average contract value? These numbers should be at the top of your mind. Additionally, fundraising should be treated as a structured process, not something you do on the side while juggling other tasks. When you start, create a clear plan: identify 20 VCs to reach out to each week. By planning ahead, you’ll maintain momentum and speed up the process. Fundraising can be exhausting and disheartening, especially when you face multiple rejections. Remember, you just need one investor to say yes to make it all worthwhile. When using funds, prioritise profitability and grow only when necessary. Don't rely on funding to survive. In the past, the common advice for startups was to raise money, burn through it quickly, and use it to boost revenue numbers, even if that meant operating at a loss. The idea was that profitability wasn’t the main focus, and the goal was to show rapid growth for the next funding round. However, times have changed, especially with the shift from “funding summer” to “funding winter.” My advice now is to aim for profitability as soon as possible and grow only when it's truly needed. For example, it’s tempting to hire a large team when you have substantial funds in the bank, but ask yourself: Do you really need 10 new hires, or could you get by with just four? Growing too quickly can lead to unnecessary expenses, so focus on reaching profitability as soon as possible, rather than just inflating your team or burn rate. The key takeaway is to spend your funds wisely and only when absolutely necessary to reach profitability. You want to avoid becoming dependent on future VC investments to keep your company afloat. Instead, prioritize reaching break-even as quickly as you can, so you're not reliant on external funding to survive in the long run. Team-Building & Leadership Look for complementary skill sets in co-founders When choosing a co-founder, it’s important to find someone with a complementary skill set, not just someone you’re close to. For example, I come from a business and commercial background, so I needed someone with technical expertise. That’s when I found my co-founder, Himanshu, who had experience in machine learning and AI. He was a great match because his technical knowledge complemented my business skills, and together we formed a strong team. It might seem natural to choose your best friend as your co-founder, but this can often lead to conflict. Chances are, you and your best friend share similar interests, skills, and backgrounds, which doesn’t bring diversity to the table. If both of you come from the same industry or have the same strengths, you may end up butting heads on how things should be done. Having diverse skill sets helps avoid this and fosters a more collaborative working relationship. Himanshu (left) and Somsubhra (right) co-founded AI Palette in 2018 Define roles clearly to prevent co-founder conflict To avoid conflict, it’s essential that your roles as co-founders are clearly defined from the beginning. If your co-founder and you have distinct responsibilities, there is no room for overlap or disagreement. This ensures that both of you can work without stepping on each other's toes, and there’s mutual respect for each other’s expertise. This is another reason as to why it helps to have a co-founder with a complementary skillset to yours. Not only is having similar industry backgrounds and skillsets not particularly useful when building out your startup, it's also more likely to lead to conflicts since you both have similar subject expertise. On the other hand, if your co-founder is an expert in something that you're not, you're less likely to argue with them about their decisions regarding that aspect of the business and vice versa when it comes to your decisions. Look for employees who are driven by your mission, not salary For early-stage startups, the first hires are crucial. These employees need to be highly motivated and excited about the mission. Since the salary will likely be low and the work demanding, they must be driven by something beyond just the paycheck. The right employees are the swash-buckling pirates and romantics, i.e those who are genuinely passionate about the startup’s vision and want to be part of something impactful beyond material gains. When employees are motivated by the mission, they are more likely to stick around and help take the startup to greater heights. A litmus test for hiring: Would you be excited to work with them on a Sunday? One of the most important rounds in the hiring process is the culture fit round. This is where you assess whether a candidate shares the same values as you and your team. A key question to ask yourself is: "Would I be excited to work with this person on a Sunday?" If there’s any doubt about your answer, it’s likely not a good fit. The idea is that you want employees who align with the company's culture and values and who you would enjoy collaborating with even outside of regular work hours. How we structure the team at AI Palette We have three broad functions in our organization. The first two are the big ones: Technical Team – This is the core of our product and technology. This team is responsible for product development and incorporating customer feedback into improving the technology Commercial Team – This includes sales, marketing, customer service, account managers, and so on, handling everything related to business growth and customer relations. General and Administrative Team – This smaller team supports functions like finance, HR, and administration. As with almost all businesses, we have teams that address the two core tasks of building (technical team) and selling (commercial team), but given the size we're at now, having the administrative team helps smoothen operations. Set broad goals but let your teams decide on execution What I've done is recruit highly skilled people who don't need me to micromanage them on a day-to-day basis. They're experts in their roles, and as Steve Jobs said, when you hire the right person, you don't have to tell them what to do—they understand the purpose and tell you what to do. So, my job as the CEO is to set the broader goals for them, review the plans they have to achieve those goals, and periodically check in on progress. For example, if our broad goal is to meet a certain revenue target, I break it down across teams: For the sales team, I’ll look at how they plan to hit that target—how many customers they need to sell to, how many salespeople they need, and what tactics and strategies they plan to use. For the technical team, I’ll evaluate our product offerings—whether they think we need to build new products to attract more customers, and whether they think it's scalable for the number of customers we plan to serve. This way, the entire organization's tasks are cascaded in alignment with our overarching goals, with me setting the direction and leaving the details of execution to the skilled team members that I hire.

I studied how 7 Founders found their first 100 customers for their businesses. Summarizing it here!
reddit
LLM Vibe Score0
Human Vibe Score1
adriannelestrangeThis week

I studied how 7 Founders found their first 100 customers for their businesses. Summarizing it here!

I am learning marketing, and so I combed through the internet to find specific advice that helped founders reach 100 users and not random Google answers. Here’s what I found: Llama Life by Marie Marie founder of Llama Life, a productivity app ($51.4K+ revenue) got her first 100 users using Snowballing effect. She shared great advice that I want to add here verbatim, “Need to think about what you have that you can leverage based on your current situation. eg..When you have no customers, think about where you can post to get the 1st customer eg Product Hunt. If you do well on PH, say you get #3 product of the day, then you post somewhere else saying ‘I got #3 product of the day’.. to get your next few customers. Maybe that post is on reddit with some learnings that you found. If the reddit post does well, then you might post it on Twitter, saying reddit did well and what learnings you got from that etc. or even if it doesn’t do well you can still post about it.” Another tip she shared is to build related products that get more viral than the product itself. These are small stand-alone sites that would appeal to the same target audience, but by nature, are more shareable. On these sites, you can mention your startup like: ‘brought to you by Llama Life’ and then provide a link to the main website if someone is interested. If one of those gets viral or ranks on Google, you’ll have a passive traffic source. Scraping bee by Pierre Pierre, founder of Scraping Bee, a web scraping tool has now reached $1.5M ARR. Pierre and his cofounder Kevin started with 10 Free Beta Users in 2019, and after 6 months asked them to take a paid subscription if they wanted to continue using the product. That’s how they got their first user within 50 minutes of that email. Then they listed it on dozens of startup directories but their core strategy was writing the best possible content for their target audience — Developers. 3 very successful pieces of content that worked were : A small tutorial on how to scrape single-page application An extensive general guide about web scraping without getting blocked A complete introduction to web scraping with Python They didn’t do content marketing for the sake of content marketing but deep-dived into the value they were providing their customer. One of these got 70K visits, and all this together got them to over 100 users. WePay by Bill Clerico Bill Clerico left his cushy corporate job to build WePay which was then acquired for $400M got his first users by using his app. He got his first users by using his app! The app was for group payments. So he hosted a Poker tournament at his house and collected payments only with his app. Then they hosted a barbecue for fraternity treasurers at San Jose State & helped them do their annual dues collection. Good old word-of-mouth marketing, that however, started with an event where they used what they made! RealWorld by Genevieve Genevieve — Founder and CEO of Realworld stands by the old-school advice of value giving. RealWorld is an app that helps GenZ navigate adulthood. So, before launching their direct-to-consumer platform, they had an educational course that they sold to college career centers and students. They already had a pipeline of adults who turned to Realworld for their adulting challenges. From there, she gained her first 100 followers. Saner dot ai by Austin Austin got 100 users from Reddit for his startup Saner.ai. Reddit hates advertising, and so his tips to market your startup on Reddit is to Write value-driven posts on your niche. Instead of writing posts, find posts where people are looking for solutions DM people facing problems that your SaaS solves. But instead of selling, ask about their problem to see if your product is a good fit Heartfelt posts about why you built it, aren’t gonna cut it To find posts and people, search Reddit with relevant keywords and join all the subreddits A Stock Portfolio Newsletter A financial investor got his first 100 paid newsletter subscribers for his stock portfolio newsletter. His tips : Don’t reinvent the wheel. Work what’s already working. He saw a company making $500M+ from stock picking newsletter, so decided to try that. Find the gaps in “already working” and leverage them. That newsletter did not have portfolios of advisors writing them. That was his USP. He added his own portfolio to his newsletter. Now to 100 users, he partnered with a guy running an investing website and getting good traffic. That guy got a cut of his revenue, in exchange. That one simple step got him to 100 users. Hypefury by Yannick and Samy Yannick and Samy from Hypefury, Twitter and Social Media Automation tool got their first beta testers and users from a paid community. They launched Hypefury there and asked if someone wanted to try it. A couple of people tried it and gave feedback. Samy conducted user interviews and product demos for them, And shared the reviews on Twitter. That alone, along with word-of-mouth marketing on Twitter got them their first 100 users. To conclude: Don’t reinvent the wheel, try what’s working. Find the gaps in what’s working, and leverage that. Instead of thinking about millions of customers, think about the first 10. Then first 100. Leverage what you have. Get the first 10 customers, then talk about this to get the next 100. Use your app. Find ways, events, and opportunities to use your app in front of people. And get them to use it. Write content not only for SEO but also to help people. It won’t work tomorrow, but it will work for years after it picks up. Leverage other sources of traffic by partnering up! Do things that don’t scale. I’m also doing SaaS marketing deep dives over 30 pieces of content. I'm posting here for the first time, so I'm not sure if it will stay or not, sorry if it doesn't. I've helped a SaaS grow from $19K to $100K MRR as a marketer in last 2 years, and now I wanna dive deep. Cheers! (1/30)

Joined an AI Startup with Ex-ShipStation Team - Need Tips on Finding Early Users
reddit
LLM Vibe Score0
Human Vibe Score1
welcomereadThis week

Joined an AI Startup with Ex-ShipStation Team - Need Tips on Finding Early Users

Hey Reddit, My name’s Welcome (Yes, that’s really my name), and I’ve been in tech for most of my career, mostly at bigger companies with established brands and resources. But recently, I decided to join a small startup called BotDojo. It’s my first time being part of a small team, and it’s been a pretty eye-opening experience so far. But, like with anything new, I’ve hit a few bumps along the way, and I’m hoping you all might have some advice. A little backstory: BotDojo was started by some of the engineers who used to work together at ShipStation. After ShipStation sold, they spent some time experimenting with AI but kept running into the same problems—having to patch together tools, getting inconsistent results, handling data ingestion, and struggling to track performance. So, they decided to build a platform to help developers build, test, and deploy AI solutions. Since I came on board, my focus has been on finding early users, and it’s been a mixed bag of wins and frustrations. We’ve got a solid group of people using the free version (which is great), but only a few have upgraded to the paid plan so far (ranging from startups to large enterprises). The cool thing is that those who have become paying customers absolutely love the product. It’s just been hard getting more people to that point. We’ve tried a bunch of things: Attending industry events, doing cold email outreach, running social ads (the usual stuff). And while we’ve seen some interest, we’re running into a few challenges:   Learning curve: The software is really powerful, but it takes a week or two for users to really see what it can do. Without a dedicated sales team to walk them through it, it’s been tough getting people to stick around long enough to see the value. Standing out is hard: The AI space is super crowded right now. I think a lot of people see “AI tool” and assume it’s just like everything else out there (even though BotDojo has some awesome features that really set it apart).  Sign-ups, but limited engagement: We’re on a freemium model to make it easy for people to try it out, but that also means we get a lot of bots and people who sign up but don’t really dive in. So, I thought I’d reach out here and see if anyone has been through this early stage before. How did you manage to break through and find those first paying users who really saw the value in what you were building?  Are there any strategies, communities, or tactics that worked particularly well for you? And if you had to do it all over again, what would you focus on? I figure I’m not the only one trying to navigate these waters, so I’m hoping this can be a helpful thread for others too. Thanks so much for reading, and I’d be super grateful for any advice or insights you can share! 🙏

I studied how 7 Founders found their first 100 customers for their businesses. Summarizing it here!
reddit
LLM Vibe Score0
Human Vibe Score1
adriannelestrangeThis week

I studied how 7 Founders found their first 100 customers for their businesses. Summarizing it here!

I am learning marketing, and so I combed through the internet to find specific advice that helped founders reach 100 users and not random Google answers. Here’s what I found: Llama Life by Marie Marie founder of Llama Life, a productivity app ($51.4K+ revenue) got her first 100 users using Snowballing effect. She shared great advice that I want to add here verbatim, “Need to think about what you have that you can leverage based on your current situation. eg..When you have no customers, think about where you can post to get the 1st customer eg Product Hunt. If you do well on PH, say you get #3 product of the day, then you post somewhere else saying ‘I got #3 product of the day’.. to get your next few customers. Maybe that post is on reddit with some learnings that you found. If the reddit post does well, then you might post it on Twitter, saying reddit did well and what learnings you got from that etc. or even if it doesn’t do well you can still post about it.” Another tip she shared is to build related products that get more viral than the product itself. These are small stand-alone sites that would appeal to the same target audience, but by nature, are more shareable. On these sites, you can mention your startup like: ‘brought to you by Llama Life’ and then provide a link to the main website if someone is interested. If one of those gets viral or ranks on Google, you’ll have a passive traffic source. Scraping bee by Pierre Pierre, founder of Scraping Bee, a web scraping tool has now reached $1.5M ARR. Pierre and his cofounder Kevin started with 10 Free Beta Users in 2019, and after 6 months asked them to take a paid subscription if they wanted to continue using the product. That’s how they got their first user within 50 minutes of that email. Then they listed it on dozens of startup directories but their core strategy was writing the best possible content for their target audience — Developers. 3 very successful pieces of content that worked were : A small tutorial on how to scrape single-page application An extensive general guide about web scraping without getting blocked A complete introduction to web scraping with Python They didn’t do content marketing for the sake of content marketing but deep-dived into the value they were providing their customer. One of these got 70K visits, and all this together got them to over 100 users. WePay by Bill Clerico Bill Clerico left his cushy corporate job to build WePay which was then acquired for $400M got his first users by using his app. He got his first users by using his app! The app was for group payments. So he hosted a Poker tournament at his house and collected payments only with his app. Then they hosted a barbecue for fraternity treasurers at San Jose State & helped them do their annual dues collection. Good old word-of-mouth marketing, that however, started with an event where they used what they made! RealWorld by Genevieve Genevieve — Founder and CEO of Realworld stands by the old-school advice of value giving. RealWorld is an app that helps GenZ navigate adulthood. So, before launching their direct-to-consumer platform, they had an educational course that they sold to college career centers and students. They already had a pipeline of adults who turned to Realworld for their adulting challenges. From there, she gained her first 100 followers. Saner dot ai by Austin Austin got 100 users from Reddit for his startup Saner.ai. Reddit hates advertising, and so his tips to market your startup on Reddit is to Write value-driven posts on your niche. Instead of writing posts, find posts where people are looking for solutions DM people facing problems that your SaaS solves. But instead of selling, ask about their problem to see if your product is a good fit Heartfelt posts about why you built it, aren’t gonna cut it To find posts and people, search Reddit with relevant keywords and join all the subreddits A Stock Portfolio Newsletter A financial investor got his first 100 paid newsletter subscribers for his stock portfolio newsletter. His tips : Don’t reinvent the wheel. Work what’s already working. He saw a company making $500M+ from stock picking newsletter, so decided to try that. Find the gaps in “already working” and leverage them. That newsletter did not have portfolios of advisors writing them. That was his USP. He added his own portfolio to his newsletter. Now to 100 users, he partnered with a guy running an investing website and getting good traffic. That guy got a cut of his revenue, in exchange. That one simple step got him to 100 users. Hypefury by Yannick and Samy Yannick and Samy from Hypefury, Twitter and Social Media Automation tool got their first beta testers and users from a paid community. They launched Hypefury there and asked if someone wanted to try it. A couple of people tried it and gave feedback. Samy conducted user interviews and product demos for them, And shared the reviews on Twitter. That alone, along with word-of-mouth marketing on Twitter got them their first 100 users. To conclude: Don’t reinvent the wheel, try what’s working. Find the gaps in what’s working, and leverage that. Instead of thinking about millions of customers, think about the first 10. Then first 100. Leverage what you have. Get the first 10 customers, then talk about this to get the next 100. Use your app. Find ways, events, and opportunities to use your app in front of people. And get them to use it. Write content not only for SEO but also to help people. It won’t work tomorrow, but it will work for years after it picks up. Leverage other sources of traffic by partnering up! Do things that don’t scale. I’m also doing SaaS marketing deep dives over 30 pieces of content. I'm posting here for the first time, so I'm not sure if it will stay or not, sorry if it doesn't. I've helped a SaaS grow from $19K to $100K MRR as a marketer in last 2 years, and now I wanna dive deep. Cheers! (1/30)

How to get funding for startup ? I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
wlynncorkThis week

How to get funding for startup ? I will not promote

I will not promote. Software startup based out of Minnesota us. I've built and launched a product that is gaining traction, solving a problem that has frustrated software developers and product teams for years. The problem: Software development is slow, expensive, and full of inefficiencies. Developers spend hours on repetitive coding tasks, project managers struggle with bottlenecks, and businesses waste time translating product requirements into actual code. The solution: My product automates a large portion of software development. It acts as an AI-powered assistant for developers, taking high-level requirements and turning them into functional code while integrating with existing codebases. It can read, understand, and modify software projects in a structured way—cutting development time drastically. The potential: Businesses are always looking for ways to cut costs and speed up development. With the rise of AI, companies are increasingly adopting automation, and this tool fits perfectly into that wave. Imagine a world where software teams are 10x more efficient because AI handles the grunt work, and developers focus on the bigger picture. It’s not about replacing developers—it’s about supercharging them. The current status: The product is live and in use. The user base is growing, and I’ve proven demand. Now, I need to figure out the best funding model to scale—whether that’s bootstrapping, VC, grants, or some hybrid approach. If you have experience in startup funding or have scaled a tech product, I'd love to hear your insights. DM me if you're open to discussing strategies!

No revenue for 6 months, then signed $10k MRR in 2 weeks with a new strategy. Here’s what I changed.
reddit
LLM Vibe Score0
Human Vibe Score0.6
xoyourwifeThis week

No revenue for 6 months, then signed $10k MRR in 2 weeks with a new strategy. Here’s what I changed.

This is my first company so I made A LOT of mistakes when starting out. I'll explain everything I did that worked so you don't have to waste your time either. For context, I built a SaaS tool that helps companies scale their new client outreach 10x (at human quality with AI) so they can secure more sales meetings. Pricing I started out pricing it way too low (1/10 as much as competitors) so that it'd be easier to get customers in the beginning. This is a HUGE mistake and wasted me a bunch of time. First, this low pricing meant that I was unable to pay for the tools I needed to make sure my product could be great. I was forced to use low-quality databases, AI models, sending infrastructure -- you name it. Second, my customers were less invested in the product, and I received less input from them to make the product better. None ended up converting from my free trial because my product sucked, and I couldn't even get good feedback from them. I decided to price my product much higher, which allowed me to use best-in class tools to make my product actually work well. Outreach Approach The only issue is that it's a lot harder to get people to pay $500/month than $50/month. I watched every single video on the internet about cold email for getting B2B clients and built up an outbound MACHINE for sending thousands of emails a day. I tried all the top recommended sales email formats and tricks (intro, painpoint, testimonial, CTA, etc). Nothing. I could send 1k emails and get a few out of office responses and a handful of 'F off' responses. I felt bad and decided I couldn't just spam the entire world and expect to make any progress. I decided I needed to take a step back and learn from people who'd succeeded before in sales. I started manually emailing CEOs/founders that fit my customer profile with personal messages asking for feedback on my product -- not even trying to sell them anything. Suddenly I was getting 4-6 meetings a day and just trying to learn from them (turns out people love helping others). And without even prompting, many of them said 'hey, I actually could use this for my own sales' and asked how they could start trying it out. That week I signed 5 clients between $500-$4k/month (depending how many contacts they want to reach). I then taught my product to do outreach the same way I did that worked (include company signals, make sure the person is a great match with web research, and DONT TALK SALESY). Now, 6 of my first 10 clients (still figuring out who it works for, lol) have converted from the free trial and successfully used it to book sales meetings. I'm definitely still learning, but this one change in my sales approach changed everything for me, so I wanted to share. If anyone has any other tips/advice that changed their business's sales, would love to hear!

Building in the open with Founder University - I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
Tim-SylvesterThis week

Building in the open with Founder University - I will not promote

Published Oct 30, 2024 I am on my fifth startup. I ran the last one for a decade, that’s a whole story. A hell of a story. But a different story. I’ll tell it to you when I can, but not right now. The one before that was an e-commerce site that did pretty well but I didn’t love it. Before that were two service businesses. The first one I did for the love of the game, the second one was an attempt to make people stop asking me to fix their computer by charging them outrageous prices, which backfired horribly when they were eager to pay. None are relevant except to say I’ve been around the block and have the scars to prove it. When it was time to get back out there, I wanted to use all I’ve learned to do better. Before I talk about what those lessons produced, I’m going to talk about what those lessons were. Cause before effect, after all. One thing I wanted to do better this time was pattern matching - making the startup look the way that the industry and investors “expect” a startup to look. My last startup was an awesome idea with awesome tech (still is, but like I said, another story), but that one didn’t match patterns. It didn’t match investor patterns, industry buying patterns, patterns of existing, immediate, recognized and admitted needs. Because it didn’t “look” right to anyone, everything about it was way harder than necessary. The “make it look right” approach runs the risk of building a cargo cult, imitating the trappings of something but without understanding the essence of that something, but then again, a thing that looks like a knife is going to make a better knife that a thing that looks like a bowling ball, so sometimes just sharing apparent similarities can get you pretty far, even if it doesn’t get you all the way there. Like how mimicking someone’s accent makes it easier for them to understand you. For this one, I wanted to adopt every tool, method, and pattern that I knew “the industry” wanted to see to minimize the friction from development, go-to-market, scaling, adoption, and that would make investment optional (and, therefore, available if desired) instead of necessary (and, therefore, largely unavailable). That required establishing some expectations for successful patterns I could match against. What patterns am I matching to? Here’s a general sketch of my pattern matching thought process: Software first and software only. It’s the easiest industry to start a business in, lowest startup costs, and easiest customer acquisition. I wanted to build software for an element of the industry that’s actively emerging (and therefore has room to grow) and part of an optimistic investor thesis (and therefore has a cohort of people who are intent on injecting capital into the market to help it grow). It needs to fills a niche that is underexplored (low competition) and highly potent (lots of opportunity), while being aligned to recognized and emerging needs within the industry (readily adopted). I wanted it to have evidence supporting the business thesis that proves the demand exists, but demonstrates that the demand is unanswered (as of yet) by sufficient or adequate supply.* I wanted the lowest number of dominoes to line up and tip for everything to work correctly - the more dominoes in the line, the less likely the last one will fall. I wanted to implement modern toolsets for everything, wherever possible. I wanted to obey the maxim, “When there’s a gold rush, don’t mine the gold, sell the picks and shovels.” Whatever I chose would need to produce cash flow almost immediately with minimal development time or go-to-market delays, because the end of ZIRP killed the “trust me bro” investment thesis predominant over the last 15 years. I wanted to match to YC best practices, not because YC can predict what will definitely work, but because they’ve churned through so many startups in the last 15 years that they have a good sense of what will definitely not work. And I wanted to build client-centric, because if my intent is to to produce cash flow immediately, we need to get clients immediately, and if we need to get clients immediately, we need to focus on what clients need right now. Extra credit: What’s the difference between a customer and a client? Note: Competition is awesome! Competition is validating and not scary, because competition proves a market exists. But competition, especially mature competition against an immature startup, makes it harder to break into a space. A first mover advantage isn’t everything, but seeing demand before it’s sufficiently supplied is a great advantage if you’re capital constrained or otherwise unproven. Think about how much money the first guy to sell fidget spinners or Silly Bandz made versus how much money the last guy to order a pallet of each made. Finding demand that exists already but is as of yet insufficiently satisfied is a great place to start. What opportunity spaces are most relevant? The industries and markets I chose to observe were: AI, because if I’m following a theme & pattern for today, it’s AI. Fintech, because cash is king, and fintech puts your hands on cash flow. Crypto/blockchain, because that’s the “new” fintech (or maybe the “old-new” fintech?), and crypto creates powerful incentives and capital formation strategies, along with a lot of flexibility for transaction systems. Tools, particularly unmet demand in tools, that enable these industries. If you wanted to do some brief and simple homework, you could map each of those bullets to several of the numbered list items preceding them. The reasoning was pretty simplistic - AI is what people want to build and invest in now, while fintech and crypto/blockchain are what people were building and investing in for the last major investment thesis. That means that there’s demand in the market for AI and AI-adjacent startups, while there’s a glut of underutilized and highly developed tools within fintech and crypto/blockchain, with a lot of motivated capital behind the adoption. When someone is thinking “I built this thing and not enough people are using it”, and you then build something that uses it creates a great way to find allies. This rationale harnesses technology that is being built and financed now (which means it needs tools and support methods, and a lot of other “picks and shovels”), while leveraging technology that was recently built and financed and is eager for more widespread adoption of the existing toolkits, which makes it suitable for using to build the AI-adjacent tools that are in demand now. It’s like two harmonics producing constructive interference - it makes two waves into one larger wave, which gives me more momentum to surf against. This was a learning process, and I iterated against my general paradigm repeatedly as I learned more. Neither of us have the patience to go through that in excruciating detail, so I’ll cover the highlights in my next post. Extra credit answer: A customer gets a product, a client gets a service. Challenge: Is software a product or a service?

I just had my best month after 18 months as a solopreneur
reddit
LLM Vibe Score0
Human Vibe Score0.778
stepitup9600This week

I just had my best month after 18 months as a solopreneur

Last month I reached important milestones both financially (60+ sales) and in terms of my personal brand (2.500+ new followers) But the most important part is that it has reinforced a belief in myself: it is possible, as long as I keep going, improving, learning and iterating. For the last year and a half, I've been grinding and launching project after project. But there was always something wrong: Product didn't solve a real problem Bad marketing (very often lol) Target market had low purchasing power Super-competitive niche (usually b2c) It's difficult to have failure after failure and keep going on. At times it would feel like everyone was making money, except for me. I was hacking on my projects every single day before and after my 9-5 and had mostly given up all my free time for this. But results were far from being what I wanted. So I would doubt myself all the time. One thing I had going for me is that I really enjoy building things - so that helped me a lot in staying consistent. I always knew this was a long-term thing and that I'd probably have to fail again and again before seeing some success. But even so, it was really hard to keep up the spirits at all time, especially after working so hard for so long. I wasn't going to give up but I also knew that continuing like this would lead nowhere. So I decided that for my next project I would do 2 things: 1) prioritize marketing and 2) build something strategic 1) Prioritize marketing I decided I was going to put in the same amount of effort into marketing as I put into building. Usually my time would be split 90% coding - 10% marketing. Now, for the first time ever it's probably 65% coding - 35% marketing. I organized myself and made an entire gameplan for it. This forced me to learn a lot about: Video editing Cold emails Copywriting Running ads Short-form content There are a lot of items I still need to execute on - but at least I have a good idea of how to approach most things. 2) Build something strategic I had to build something that I would be able to use even if nobody else did. For the last year and a half I had been building AI apps and my plan was to continue doing that. So I decided to leverage that and thought about how I could build something that would give me an unfair advantage + have a compounding effect over the long term: a) Unfair advantage Having AI demo apps that cover all type of AI functionalities would make my life easier & would allow me to ship new apps quickly, regardless of the required model/functionality So even if nobody bought this - I'd have built something really useful for myself & would have a slight edge over other people b) Compound over the long term Building "AnotherWrapper' (my new project) would have a good synergy with my future projects: It would allow me to build new projects faster While building new projects, I'd learn new things, which I would then be able to implement into AnotherWrapper and improve the product that way A win-win. Closing thoughts I did not expect things to go this well - it's been an amazing month and I'm truly grateful to everyone that has been supporting me. But at the end of the day, there is still a lot of work to be done. The initial 'hype' & effects from some viral tweets are starting to wear off. I still don't have a reliable distribution channel that guarantees me traffic. So I need to figure that out. I think the product has a lot of potential - it has been well received and has been a success so far, but my distribution is still lacking. The good thing is that I now have some extra cash to spend on things like ads, influencers, freelancers etc. So it opens some new doors that were previously closed! I also have some other projects down the pipeline which are coming soon. Will keep you guys updated!

Hot Take: Not all your startups need AI forced into them
reddit
LLM Vibe Score0
Human Vibe Score1
bitorsicThis week

Hot Take: Not all your startups need AI forced into them

I'm a final year Computer Engineering student, hence applying for jobs all around. There's this particular trend I've noticed with startups that are coming up these days. That is, even for the absolute basic stuff they'll use 'AI', and they'll think they built something 'revolutionary'. No. You're breaking your product in ways you don't realise. An example, that even some well established companies are guilty of: AI Chatbots You absolutely don't need them and it's an entire gimmick. If you really wanna implement a chatbot, connect the user to an actual person on your end, which I think is not possible if you're at a 'startup' stage. You'll need employees who can handle user queries in real time. If the user really is stuck let them use the 'Contact Us' page. A really close relative of mine is very vocal about the frustration he faces whenever he tries to use the AI Chatbot on any well known e-com website. The only case for AI Chatbot that makes sense is when it's directing the customer to an actual customer support rep if none of the AI's solutions is working for the customer. Even then, implementing a search page for FAQ is extremely easy and user friendly. Another example: AI Interviewer I recently interviewed for a startup, and their whole interviewing process was AI'zed?!?! No real person at the other end, I was answering to their questions which were in video format. They even had a 'mascot' / 'AI interviewer' avatar designed by an AI (AI-ception???). This mascot just text-to-speech'ed all the questions for me to rewind and hear what I missed again. And I had to record video and audio to answer these questions on their platform itself. The entire interview process just could've been a questionnaire, or if you're really concerned on the integrity of the interviewee, just take a few minutes out of your oh-so-busy schedule as a startup owner. Atleast for hiring employees who would make the most impact on your product going ahead. I say the most impact, because (atleast as a developer) the work done by these employees would define how robust your product is, and/or how easily other features can be integrated into the codebase. Trust me, refactoring code later on would only cost you time and money. These resources would rather be more useful in other departments of your startup. The only use case for an AI Interviewer I see is for preparing for an actual interview, provided that feedback is given to the user at the earliest, which you don't need to worry about as a startup owner. So yeah, you're probably better off without integrating AI in your product. Thank you for reading. TLDR; The title; I know AI is the new thing and gets everyone drooling and all, but for the love of God, just focus on what your startup does best and put real people behind it; Integrating AI without human intervention is as good as a broken product; Do your hiring yourself, or through real people, emphasizing on the fact that the people you hire at an early stage will define your growth ahead;

Content aggregation that acts as a middleman for content discovery via third-party marketplace & revenue sharing (i will not promote but I'm looking for fellow researchers)
reddit
LLM Vibe Score0
Human Vibe Score1
colbyn-wadmanThis week

Content aggregation that acts as a middleman for content discovery via third-party marketplace & revenue sharing (i will not promote but I'm looking for fellow researchers)

High level I’m considering a content aggregation business model, but one that acts as an open marketplace where third party devs and where world class data scientists compete to build the best recommenders for different use cases. (E.g. the incentives can be ad revenue sharing or subscription based for niche professional markets.) The idea is to facilitate more bottom up innovation from third party data scientists. The platform itself just acts as the middleman. (Also something that strips out original ads and makes it easy to skip paid sponsorship sections would be great.)  I’ve seen startups building web crawlers and content aggregation systems for other AI startups. My proposal is better in the sense that third party devs are instead responsible for implementing whatever questionable hacks are necessarily to scrape platforms that don’t necessarily want to be scraped.  Personally, I’m more concerned about getting the right information than ever before, to this end I can’t rely on platform specific recommenders. The solution is more bottom up innovation in content promotion. More generally, if you’re also concerned about consuming game changing information that’s too easily missed: we need a platform that incentivizes bottom up innovation of content promotion. What we need is a platform that functions like a marketplace where third party devs and where world class data scientists compete to build the best recommenders for different use cases. Here’s some elevator pitches I’m considering:  Did you know that the magic behind YouTube is its recommendation engine? Now, imagine an open platform where independent engines compete to deliver the most personalized content feed—from news to local events—directly to you. Interested in rethinking how we find content? “In today’s fragmented digital landscape, a single platform no longer holds sway over content discovery. The Network Effect is dead: audiences are more mobile than ever; and big tech killed it. In such a fragmented landscape we’re building a bottom-up, decentralized marketplace for recommendation engines—a solution that taps into diverse revenue streams through subscriptions, ad revenue, and affiliate partnerships. Invest in the future of personalized content aggregation.” “Are you a developer passionate about algorithms and content discovery? Our open marketplace lets you build and monetize your own recommendation engine, competing to deliver the most engaging, personalized feeds. Join a revolution where your innovation can directly shape how the world finds content.” “Are you tired of being told what to watch or read by one mysterious algorithm? Imagine taking control—choosing from a marketplace of smart recommendation engines that curate content just for you. It’s a revolution in content discovery where you hold the power.” (As a Utahn this one is interesting because even mormons are talking about the dangers of “doom scrolling” though it’s seldom discussed in society at large.) As far as simple hooks I’m considering:  One platform to rule them all and in the darkness bind them.  Choose how you discover—content recommenders that work for you.  The area where recommender engines battle to win your feed. Request I would love to start prototyping this idea and see what else I can uncover from such preliminary research. But I want to get a couple other likeminded individuals onboard.  I'm the best when it comes to iOS/macOS development, but there's tons of backend work that needs to be done which I wouldn’t have the time for if i'm focused on the native clients. Who am I 'ideally' looking for?  I’ve heard of weird stats to the effect that if you scale up a population to billions of people, the number of life overlaps starts skyrocketing. Not just physical lookalikes, but people with eerily similar life paths, personalities, habits, and even thoughts — without ever knowing each other. Where are my clones? Such is whom I’m looking for in an ideal world.  Take a hunch  People nowadays have no concept of going out on a limb, taking a ‘hunch’, and backing their instincts. Everything has to be calculated, proven, and guaranteed before they make a move. In contrast consider the success of the Chinese DeepSeek project: According to Asianometry’s YouTube video on DeepSeek, their “memory-saving multi-head latent architecture” (whatever that means, just quoting the name) came about from a researchers ‘hunch’, which the company bet big on and the result was drastically improved performance on low end hardware…  Here in the west the idea of betting on a hunch is inconceivable. We have no balls to chase long term insights. My own instincts when it comes to software is such because I’ve wasted too much of my life on small scale projects. All I’m trying to do is attempt a more scaled up experiment based on some hunches with me and a few other likeminded individuals.  Just as the early oil prospectors didn’t have precise maps—just intuition and test drills. They had to drill, analyze the pressure, and adjust. The best oil fields weren’t found by foresight alone, but by adaptive exploration. The startup space itself is liken to the first prospectors who got the gold nuggets lying in the riverbed. In such an environment moving first has its advantages but nowadays I wish I could have all those shitty ‘engineers’ sent to their maker.  Today the reality is such that you’ve got to dig deep—where vast stores of wealth can be found—or go home, and those who dig into the depths cannot use mere forethought, for what lies beneath cannot be seen by the mind’s eye.  I will not promote but I'm looking for fellow research oriented minds.

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

Lessons from 139 YC AI startups (S23)
reddit
LLM Vibe Score0
Human Vibe Score0.333
minophenThis week

Lessons from 139 YC AI startups (S23)

YC's Demo Day was last week, and with it comes another deluge of AI companies. A record-breaking 139 startups were in some way related to AI or ML - up from 112 in the last batch. Here are 5 of my biggest takeaways: AI is (still) eating the world. It's remarkable how diverse the industries are - over two dozen verticals were represented, from materials science to social media to security. However, the top four categories were: AI Ops: Tooling and platforms to help companies deploy working AI models. We'll discuss more below, but AI Ops has become a huge category, primarily focused on LLMs and taming them for production use cases. Developer Tools: Apps, plugins, and SDKs making it easier to write code. There were plenty of examples of integrating third-party data, auto-generating code/tests, and working with agents/chatbots to build and debug code. Healthcare + Biotech: It seems like healthcare has a lot of room for automation, with companies working on note-taking, billing, training, and prescribing. And on the biotech side, there are some seriously cool companies building autonomous surgery robots and at-home cancer detection. Finance + Payments: Startups targeting banks, fintechs, and compliance departments. This was a wide range of companies, from automated collections to AI due diligence to "Copilot for bankers." Those four areas covered over half of the startups. The first two make sense: YC has always filtered for technical founders, and many are using AI to do what they know - improve the software developer workflow. But it's interesting to see healthcare and finance not far behind. Previously, I wrote: Large enterprises, healthcare, and government are not going to send sensitive data to OpenAI. This leaves a gap for startups to build on-premise, compliant \[LLMs\] for these verticals. And we're now seeing exactly that - LLMs focused on healthcare and finance and AI Ops companies targeting on-prem use cases. It also helps that one of the major selling points of generative AI right now is cost-cutting - an enticing use case for healthcare and finance. Copilots are king. In the last batch, a lot of startups positioned themselves as "ChatGPT for X," with a consumer focus. It seems the current trend, though, is "Copilot for X" - B2B AI assistants to help you do everything from KYC checks to corporate event planning to chip design to negotiate contracts. Nearly two dozen companies were working on some sort of artificial companion for businesses - and a couple for consumers. It's more evidence for the argument that AI will not outright replace workers - instead, existing workers will collaborate with AI to be more productive. And as AI becomes more mainstream, this trend of making specialized tools for specific industries or tasks will only grow. That being said - a Bing-style AI that lives in a sidebar and is only accessible via chat probably isn't the most useful form factor for AI. But until OpenAI, Microsoft, and Google change their approach (or until another company steps up), we'll probably see many more Copilots. AI Ops is becoming a key sector. "AI Ops" has been a term for only a few years. "LLM Ops" has existed for barely a year. And yet, so many companies are focused on training, fine-tuning, deploying, hosting, and post-processing LLMs it's quickly becoming a critical piece of the AI space. It's a vast industry that's sprung up seemingly overnight, and it was pretty interesting to see some of the problems being solved at the bleeding edge. For example: Adding context to language models with as few as ten samples. Pausing and moving training runs in real-time. Managing training data ownership and permissions. Faster vector databases. Fine-tuning models with synthetic data. But as much ~~hype~~ enthusiasm and opportunity as there might be, the size of the AI Ops space also shows how much work is needed to really productionalize LLMs and other models. There are still many open questions about reliability, privacy, observability, usability, and safety when it comes to using LLMs in the wild. Who owns the model? Does it matter? Nine months ago, anyone building an LLM company was doing one of three things: Training their own model from scratch. Fine-tuning a version of GPT-3. Building a wrapper around ChatGPT. Thanks to Meta, the open-source community, and the legions of competitors trying to catch up to OpenAI, there are now dozens of ways to integrate LLMs. However, I found it interesting how few B2B companies mentioned whether or not they trained their own model. If I had to guess, I'd say many are using ChatGPT or a fine-tuned version of Llama 2. But it raises an interesting question - if the AI provides value, does it matter if it's "just" ChatGPT behind the scenes? And once ChatGPT becomes fine-tuneable, when (if ever) will startups decide to ditch OpenAI and use their own model instead? "AI" isn't a silver bullet. At the end of the day, perhaps the biggest lesson is that "AI" isn't a magical cure-all - you still need to build a defensible company. At the beginning of the post-ChatGPT hype wave, it seemed like you just had to say "we're adding AI" to raise your next round or boost your stock price. But competition is extremely fierce. Even within this batch, there were multiple companies with nearly identical pitches, including: Solving customer support tickets. Negotiating sales contracts. Writing drafts of legal documents. Building no-code LLM workflows. On-prem LLM deployment. Automating trust and safety moderation. As it turns out, AI can be a competitive advantage, but it can't make up for a bad business. The most interesting (and likely valuable) companies are the ones that take boring industries and find non-obvious use cases for AI. In those cases, the key is having a team that can effectively distribute a product to users, with or without AI. Where we’re headed I'll be honest - 139 companies is a lot. In reviewing them all, there were points where it just felt completely overwhelming. But after taking a step back, seeing them all together paints an incredibly vivid picture of the current AI landscape: one that is diverse, rapidly evolving, and increasingly integrated into professional and personal tasks. These startups aren't just building AI for the sake of technology or academic research, but are trying to address real-world problems. Technology is always a double-edged sword - and some of the startups felt a little too dystopian for my taste - but I'm still hopeful about AI's ability to improve productivity and the human experience.

Competing with much bigger companies that have lame products? How do I market and carve out a niche? (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
YoKevinTrueThis week

Competing with much bigger companies that have lame products? How do I market and carve out a niche? (I will not promote)

I've been working on a product for the last few months that competes with CapCut, Adobe Premier, Veed, Descript, DaVinci Resolve, etc. Basically, it's a fancy video editor. (no link and I will not promote but just some background context) I'm very technical and started creating videos for TikTok but really wanted to take my game to the next level. My channel sort of blew up on me in the first month and I was able to get 2M views and 10k followers. My initial thinking was that I was going to use AI to make video editing fancy/faster and sort of have this as a "script" that I used personally. Basically, give myself a serious competitive advantage. However, it sort of spiraled out of control! What started off as a weekend project, turned into 2 weekends, which turned into about 2 months of continuous hacking. If I'm going to spend a significant amount of time on this, I might as well try to productize it and try to at least make enough money that I break even on my time. The thing I'm worried about, in the back of my mind, is that if I shop this, that my competitors, with their signifiant resources, could clone what I'm doing quickly. However, at the same time, why haven't they done so already? I mean maybe I have a better understanding of the market than they do because they don't actually use their products. I know that sounds like a bit of a cop out in a way but there are plenty of entrepreneurs who have started companies and crushed it just because they were heads down and focused. Another problem I face, is that I think VCs may not be super excited about this because it's B2C-ish and it's not in a super exciting space. Maybe you could say it's in the AI video space, and they're excited about AI video, but it's just an AI video editor, not fully creating AI videos from scratch like SORA. I think since I blew up my TikTok feed before, that I could do it again, and if I get 2M views, and I have a outro on my video, that I could start to convert some of these as customers. Especially, if I started to create videos for creators which is more focused on the target market. So without funding, can I really tackle these existing competitors? PS. "I will not promote" but I have to talk about this somewhat abstractly but I won't link to anything.

Experienced Software Developer looking for startup to help. I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
DB010112This week

Experienced Software Developer looking for startup to help. I will not promote

My passion for programming started at the age of 9 when I began playing video games. It was during this time that I first dived into programming, creating scripts for SA:MP (San Andreas Multiplayer) using the Pawn language. SA:MP is a modification for the popular game Grand Theft Auto: San Andreas, allowing players to experience multiplayer gameplay. My early experiences in programming were all about problem-solving—finding ways to enhance the game and improve the player experience. This was when I realized how satisfying it is to solve a problem through code, and that feeling has stayed with me throughout my career. I am a self-taught programmer, and everything I know today comes from my own initiative to learn and improve. After five years of working with local clients, I decided to expand my knowledge and started learning more widely applicable programming languages like Java and Python. I’ve always been the type of person who thrives on challenges. Whenever I encounter a problem, I don’t just look for a quick fix—I dive deep into researching and understanding the problem, and I find a solution that works in the long run. This is what drives me. The ability to solve problems, no matter how complex, and the satisfaction that comes with it is what fuels my passion for programming. My big break came when I had the opportunity to work at \\\\. There, I replaced two senior and two junior developers, which led to significant cost savings for the company. I completed all tasks ahead of schedule, focusing on Java-based applications that were multithreaded and communicated with embedded systems. This experience taught me how to work under pressure and how to manage and solve complex technical problems efficiently. Following my time at \\\\, I transitioned into freelance work as a FullStack Developer, working with technologies such as HTML, CSS, Bootstrap, JavaScript, Django, Spring, MySQL, and PostgreSQL. As a freelancer, I was responsible for finding solutions to a wide range of problems, often working independently and making decisions on the fly. I learned that self-reliance is key in this industry, and being resourceful is one of the most important qualities a developer can have. Later, I joined \\\\ elecom, where I worked on system integration with foreign teams, BPM process solutions, and the merging of complex systems in Oracle databases. I continued to solve challenges, often working with teams across borders and tackling technical obstacles that required creative and well-thought-out solutions. Eventually, I founded my own company, \\\\, where I focus on developing software solutions, Artificial Intelligence (AI), Cybersecurity, and Ethical Hacking. As an entrepreneur, I take pride in finding innovative solutions to problems, whether they come from clients or from technical obstacles I encounter along the way. I’ve also had the privilege of working with the Serbian Ministry of Defense and the police, handling sensitive projects that demand both technical expertise and trustworthiness. Being a self-taught programmer means that I have had to learn and adapt on my own, and I’ve learned to embrace challenges as opportunities for growth. I am constantly driven by the process of solving problems, and it is what keeps me engaged and fulfilled in my work. I am always open to new collaborations and am eager to take on new challenges that push my boundaries in technology, cybersecurity, and software development.

Why raise in 2025? - I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
Able_Swimming_4909This week

Why raise in 2025? - I will not promote

I will not promote Lately, I've been thinking about how AI tools are completely reshaping what it means to bootstrap a startup. It honestly feels like we're living through a golden age for entrepreneurs where you don't necessarily need venture capital to build something big or meaningful. At my company, we're a small team of just four people, bootstrapping our AI-focused startup. Thanks to AI-powered tools, we're able to keep our burn rate ridiculously low, quickly test new ideas, and scale our operations way faster than we ever expected. It’s honestly pretty incredible how accessible advanced technology has become, even compared to just a few years ago. Of course, bootstrapping definitely comes with its own share of headaches. For example, we've noticed that funded startups get significantly better access to cloud credits, advertising budgets, and enterprise-level tools. We do have access to some discounts and free resources, but it rarely compares to what funded startups enjoy. This can feel frustrating, especially when you know you're competing directly with businesses that have those extra advantages. Visibility is another major challenge we've noticed. Without big funding announcements or a well-connected investor backing us, getting attention from media or even early adopters can be tough. It's just harder to make a splash without someone else's endorsement. We've had to accept and work around creatively. That said, there's something genuinely empowering about staying bootstrapped, prioritizing profitability, and maintaining control over our vision. After speaking with several investors, we've become aware of how investors can significantly influence or even redirect the trajectory of a business. We've heard stories where investors gained enough leverage to replace the original founders or have killed perfectly profitable businesses that were not growing "fast enough", which certainly gave us pause. They can definitely be helpful but giving the control over the future of my business to someone else would definitely make me feel anxious. At this time, we simply don't feel raising external capital aligns with our current goals, but we're also aware that this could change in the future. For now, maintaining autonomy and staying close to our original vision remains a priority. I'm curious to hear from others here who've been through this. Have you successfully bootstrapped an AI a tech business? What obstacles did you encounter, and how did you overcome them? EDIT: To give you a bit of perspective, my company is a B2B SaaS in the finance industry based in Europe. We have received VC funding in the past but it was an exceptionally good deal and we don't plan to raise in the near future even-thought it may change if we see the need to help us scale. We have also raised a significant amount in soft funding. Right now, we are growing on our revenues, and we plan to continue this trajectory. Recently, one of our developers left, and although we are a small team, we noticed that it had little to no impact on our productivity.

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

How I made a high tech salary in my first selling month
reddit
LLM Vibe Score0
Human Vibe Score1
Ok_Negotiation_2587This week

How I made a high tech salary in my first selling month

For over 7 years I worked as a full-stack developer, helping other companies bring their ideas to life. But one day, I thought “Why not try making my own dream come true?”. That’s when I decided to quit my job and start my own journey to becoming an entrepreneur. At first, it wasn’t easy. I didn’t make any money for months and had no idea where to start. I felt lost. Then, I decided to focus on something popular and trending. AI was everywhere, and ChatGPT was the most used AI platform. So I looked into it and I found the OpenAI community forum where people had been asking for features that weren’t being added. That gave me an idea. Why not build those features myself? I created a Chrome extension and I worked on some of the most requested features, like: Downloading the advanced voice mode and messages as MP3 Adding folders to organize chats Saving and reusing prompts Pinning important chats Exporting chats to TXT/JSON files Deleting or archiving multiple chats at once Making chat history searches faster and better It took me about a week to build the first version, and when I published it, the response was incredible. People loved it! Some even said things like, “You’re a lifesaver!” That’s when I realized I had something that could not only help people but also turn into a real business. I kept the first version free to see how people would respond. Many users have been downloading my extension, which prompted Chrome to review it to determine if it qualified for the featured badge. I received the badge, and it has significantly boosted traffic to my extension ever since. After all the positive feedback, I launched a paid version one month ago. A few minutes after publishing it, I made my first sale! That moment was so exciting, and it motivated me to keep going. I already have over 4,000 users and have made more than $4,500 in my first selling month. I’ve decided to release 1-2 new features every month to keep improving the extension based on what users ask for. I also created the same extension for Firefox and Edge users because many people have been asking for it! I also started a Reddit community, where I share updates, sales, discount codes, and ideas for new features. It’s been awesome to connect with users directly and get their feedback. Additionally, I’ve started working on another extension for Claude, which I’m hoping will be as successful as this one. My message to you is this: never give up on your dreams. It might feel impossible at first, but with patience, hard work, and some creativity, you can make it happen. I hope this inspires you to go after what you want. Good luck to all of us!

🚀 Revolutionizing IT and Network Operations: A Vision for the Future, Smarter, Faster, Proactive  🚀
reddit
LLM Vibe Score0
Human Vibe Score1
Psychological_Cod_50This week

🚀 Revolutionizing IT and Network Operations: A Vision for the Future, Smarter, Faster, Proactive 🚀

Solution that I am building: IT teams today are bogged down by fragmented tools, reactive troubleshooting, and escalating downtime costs. This hampers innovation, inflates operational expenses, and delays business growth. We’re building something game-changing: an AI/ML-powered platform that transforms IT operations with: ✔️ Proactive issue prevention via real-time anomaly detection. ✔️ Automated remediation, reducing resolution time by up to 90%. ✔️ Unified monitoring, integrating infrastructure, apps, and services into a single-pane-of-glass dashboard. ✔️ Advanced network automation, with features like configuration drift detection, root cause analysis, and dynamic topology mapping. The goal? Less firefighting, more innovation. With faster ROI, 30–40% cost savings, and seamless scalability across hybrid and multi-cloud environments, we aim to redefine IT operations. 💡 We’d love your thoughts: 👉 Does this resonate with the challenges you’ve faced? 👉 What features would make this an essential tool for your organization? If you’d like to share insights, contribute to the vision, or even explore investment opportunities, let’s connect! Together, we can shape the future of proactive IT operations. Drop your feedback in the comments or DM me directly. Let’s innovate together! 🙌 \#ITInnovation #NetworkAutomation #AIOps #DigitalTransformation #FutureOfIT

10y of product development, 2 bankruptcies, and 1 Exit — what next? [Extended Story]
reddit
LLM Vibe Score0
Human Vibe Score1
Slight-Explanation29This week

10y of product development, 2 bankruptcies, and 1 Exit — what next? [Extended Story]

10 years of obsessive pursuit from the bottom to impressive product-market fit and exit. Bootstrapping tech products as Software Developer and 3x Startup Founder (2 bankruptcies and 1 exit). Hi everyone, your motivation has inspired me to delve deeper into my story. So, as promised to some of you, I've expanded on it a bit more, along with my brief reflections. There are many founders, product creators, and proactive individuals, I’ve read many of your crazy stories and lessons so I decided to share mine and the lessons I learned from the bottom to impressive product-market fit and exit. I've spent almost the past 10 years building tech products as a Corporate Team Leader, Senior Software Developer, Online Course Creator, Programming Tutor, Head of Development/CTO, and 3x Startup Founder (2 bankruptcies, and 1 exit). And what next? good question... A brief summary of my journey: Chapter 1: Software Developer / Team Leader / Senior Software Developer I’ve always wanted to create products that win over users’ hearts, carry value, and influence users. Ever since my school days, I’ve loved the tech part of building digital products. At the beginning of school, I started hosting servers for games, blogs and internet forums, and other things that did not require much programming knowledge. My classmates and later even over 100 people played on servers that I hosted on my home PC. Later, as the only person in school, I passed the final exam in computer science. During my computer science studies, I started my first job as a software developer. It was crazy, I was spending 200–300 hours a month in the office attending also to daily classes. Yes, I didn’t have a life, but it truly was the fulfillment of my dreams. I was able to earn good money doing what I love, and I devoted fully myself to it. My key to effectively studying IT and growing my knowledge at rocket speed was learning day by day reading guides, building products to the portfolio, watching youtube channels and attending conferences, and even watching them online, even if I didn’t understand everything at the beginning. In one year we’ve been to every possible event within 400km. We were building healthcare products that were actually used in hospitals and medical facilities. It was a beautiful adventure and tons of knowledge I took from this place. That time I built my first product teams, hired many great people, and over the years became a senior developer and team leader. Even I convinced my study mates to apply to this company and we studied together and worked as well. Finally, there were 4 of us, when I left a friend of mine took over my position and still works there. If you’re reading this, I’m sending you a flood of love and appreciation. I joined as the 8th person, and after around 4 years, when I left hungry for change, there were already over 30 of us, now around 100. It was a good time, greetings to everyone. I finished my Master’s and Engineering degrees in Computer Science, and it was time for changes. Chapter 2: 1st time as a Co-founder — Marketplace In the meantime, there was also my first startup (a marketplace) with four of my friends. We all worked on the product, each of us spent thousands of hours, after hours, entire weekends… and I think finally over a year of work. As you might guess, we lacked the most important things: sales, marketing, and product-market fit. We thought users think like us. We all also worked commercially, so the work went very smoothly, but we didn’t know what we should do next with it… Finally, we didn’t have any customers, but you know what, I don’t regret it, a lot of learning things which I used many times later. The first attempts at validating the idea with the market and business activities. In the end, the product was Airbnb-sized. Landing pages, listings, user panels, customer panels, admin site, notifications, caches, queues, load balancing, and much more. We wanted to publish the fully ready product to the market. It was a marketplace, so if you can guess, we had to attract both sides to be valuable. “Marketplace” — You can imagine something like Uber, if you don’t have passengers it was difficult to convince taxi drivers, if you don’t have a large number of taxi drivers you cannot attract passengers. After a year of development, we were overloaded, and without business, marketing, sales knowledge, and budget. Chapter 3: Corp Team Lead / Programming Tutor / Programming Architecture Workshop Leader Working in a corporation, a totally different environment, an international fintech, another learning experience, large products, and workmates who were waiting for 5 pm to finish — it wasn’t for me. Very slow product development, huge hierarchy, being an ant at the bottom, and low impact on the final product. At that time I understood that being a software developer is not anything special and I compared my work to factory worker. Sorry for that. High rates have been pumped only by high demand. Friends of mine from another industry do more difficult things and have a bigger responsibility for lower rates. That’s how the market works. This lower responsibility time allowed for building the first online course after hours, my own course platform, individual teaching newbies programming, and my first huge success — my first B2C customers, and B2B clients for workshops. I pivoted to full focus on sales, marketing, funnels, advertisements, demand, understanding the market, etc. It was 10x easier than startups but allowed me to learn and validate my conceptions and ideas on an easier market and showed me that it’s much easier to locate their problem/need/want and create a service/product that responds to it than to convince people of your innovative ideas. It’s just supply and demand, such a simple and basic statement, in reality, is very deep and difficult to understand without personal experience. If you’re inexperienced and you think you understand, you don’t. To this day, I love to analyze this catchword in relation to various industries / services / products and rediscover it again and again... While writing this sentence, I’m wondering if I’m not obsessed. Chapter 4: Next try — 2nd time as a founder — Edtech Drawing upon my experiences in selling services, offering trainings, and teaching programming, I wanted to broaden my horizons, delve into various fields of knowledge, involve more teachers, and so on. We started with simple services in different fields of knowledge, mainly relying on teaching in the local area (without online lessons). As I had already gathered some knowledge and experience in marketing and sales, things were going well and were moving in the right direction. The number of teachers in various fields was growing, as was the number of students. I don’t remember the exact statistics anymore, but it was another significant achievement that brought me a lot of satisfaction and new experiences. As you know, I’m a technology lover and couldn’t bear to look at manual processes — I wanted to automate everything: lessons, payments, invoices, customer service, etc. That’s when I hired our first developers (if you’re reading this, I’m sending you a flood of love — we spent a lot of time together and I remember it as a very fruitful and great year) and we began the process of tool and automation development. After a year we had really extended tools for students, teachers, franchise owners, etc. We had really big goals, we wanted to climb higher and higher. Maybe I wouldn’t even fully call it Startup, as the client was paying for the lessons, not for the software. But it gave us positive income, bootstrap financing, and tool development for services provided. Scaling this model was not as costless as SaaS because customer satisfaction was mainly on the side of the teacher, not the quality of the product (software). Finally, we grew to nearly 10 people and dozens of teachers, with zero external funding, and almost $50k monthly revenue. We worked very hard, day and night, and by November 2019, we were packed with clients to the brim. And as you know, that’s when the pandemic hit. It turned everything upside down by 180 degrees. Probably no one was ready for it. With a drastic drop in revenues, society started to save. Tired from the previous months, we had to work even harder. We had to reduce the team, change the model, and save what we had built. We stopped the tool’s development and sales, and with the developers, we started supporting other product teams to not fire them in difficult times. The tool worked passively for the next two years, reducing incomes month by month. With a smaller team providing programming services, we had full stability and earned more than relying only on educational services. At the peak of the pandemic, I promised myself that it was the last digital product I built… Never say never… Chapter 5: Time for fintech — Senior Software Developer / Team Lead / Head of Development I worked for small startups and companies. Building products from scratch, having a significant impact on the product, and complete fulfillment. Thousands of hours and sacrifices. This article mainly talks about startups that I built, so I don’t want to list all the companies, products, and applications that I supported as a technology consultant. These were mainly start-ups with a couple of people up to around 100 people on board. Some of the products were just a rescue mission, others were building an entire tech team. I was fully involved in all of them with the hope that we would work together for a long time, but I wasn’t the only one who made mistakes when looking for a product-market fit. One thing I fully understood: You can’t spend 8–15 hours a day writing code, managing a tech team, and still be able to help build an audience. In marketing and sales, you need to be rested and very creative to bring results and achieve further results and goals. If you have too many responsibilities related to technology, it becomes ineffective. I noticed that when I have more free time, more time to think, and more time to bounce the ball against the wall, I come up with really working marketing/sales strategies and solutions. It’s impossible when you are focused on code all day. You must know that this chapter of my life was long and has continued until now. Chapter 6: 3rd time as a founder — sold Never say never… right?\\ It was a time when the crypto market was really high and it was really trending topic. You know that I love technology right? So I cannot miss the blockchain world. I had experience in blockchain topics by learning on my own and from startups where I worked before. I was involved in crypto communities and I noticed a “starving crowd”. People who did things manually and earned money(crypto) on it.I found potential for building a small product that solves a technological problem. I said a few years before that I don’t want to start from scratch. I decided to share my observations and possibilities with my good friend. He said, “If you gonna built it, I’m in”. I couldn’t stop thinking about it. I had thought and planned every aspect of marketing and sales. And you know what. On this huge mindmap “product” was only one block. 90% of the mindmap was focused on marketing and sales. Now, writing this article, I understood what path I went from my first startup to this one. In the first (described earlier) 90% was the product, but in the last one 90% was sales and marketing. Many years later, I did this approach automatically. What has changed in my head over the years and so many mistakes? At that time, the company for which I provided services was acquired. The next day I got a thank you for my hard work and all my accounts were blocked. Life… I was shocked. We were simply replaced by their trusted technology managers. They wanted to get full control. They acted a bit unkindly, but I knew that they had all my knowledge about the product in the documentation, because I’m used to drawing everything so that in the moment of my weakness (illness, whatever) the team could handle it. That’s what solid leaders do, right? After a time, I know that these are normal procedures in financial companies, the point is that under the influence of emotions, do not do anything inappropriate. I quickly forgot about it, that I was brutally fired. All that mattered was to bring my plan to life. And it has been started, 15–20 hours a day every day. You have to believe me, getting back into the game was incredibly satisfying for me. I didn’t even know that I would be so excited. Then we also noticed that someone was starting to think about the same product as me. So the race began a game against time and the market. I assume that if you have reached this point, you are interested in product-market fit, marketing, and sales, so let me explain my assumptions to you: Product: A very very small tool that allowed you to automate proper tracking and creation of on-chain transactions. Literally, the whole app for the user was located on only three subpages. Starving Crowd: We tapped into an underserved market. The crypto market primarily operates via communities on platforms like Discord, Reddit, Twitter, Telegram, and so on. Therefore, our main strategy was directly communicating with users and demonstrating our tool. This was essentially “free marketing” (excluding the time we invested), as we did not need to invest in ads, promotional materials, or convince people about the efficacy of our tool. The community could directly observe on-chain transactions executed by our algorithms, which were processed at an exceptionally fast rate. This was something they couldn’t accomplish manually, so whenever someone conducted transactions using our algorithm, it was immediately noticeable and stirred a curiosity within the community (how did they do that!). Tests: I conducted the initial tests of the application on myself — we had already invested significantly in developing the product, but I preferred risking my own resources over that of the users. I provided the tool access to my wallet, containing 0.3ETH, and went to sleep. Upon waking up, I discovered that the transactions were successful and my wallet had grown to 0.99ETH. My excitement knew no bounds, it felt like a windfall. But, of course, there was a fair chance I could have lost it too. It worked. As we progressed, some users achieved higher results, but it largely hinged on the parameters set by them. As you can surmise, the strategy was simple — buy low, sell high. There was considerable risk involved. Churn: For those versed in marketing, the significance of repeat visitors cannot be overstated. Access to our tool was granted only after email verification and a special technique that I’d prefer to keep confidential. And this was all provided for free. While we had zero followers on social media, we saw an explosion in our email subscriber base and amassed a substantial number of users and advocates. Revenue Generation: Our product quickly gained popularity as we were effectively helping users earn — an undeniable value proposition. Now, it was time to capitalize on our efforts. We introduced a subscription model charging $300 per week or $1,000 per month — seemingly high rates, but the demand was so intense that it wasn’t an issue. Being a subscriber meant you were prioritized in the queue, ensuring you were among the first to reap benefits — thus adding more “value”. Marketing: The quality of our product and its ability to continually engage users contributed to it achieving what can best be described as viral. It was both a source of pride and astonishment to witness users sharing charts and analyses derived from our tool in forum discussions. They weren’t actively promoting our product but rather using screenshots from our application to illustrate certain aspects of the crypto world. By that stage, we had already assembled a team to assist with marketing, and programming, and to provide round-the-clock helpdesk support. Unforgettable Time: Despite the hype, my focus remained steadfast on monitoring our servers, their capacity, and speed. Considering we had only been on the market for a few weeks, we were yet to implement alerts, server scaling, etc. Our active user base spanned from Japan to the West Coast of the United States. Primarily, our application was used daily during the evenings, but considering the variety of time zones, the only time I could afford to sleep was during the evening hours in Far Eastern Europe, where we had the least users. However, someone always needed to be on guard, and as such, my phone was constantly by my side. After all, we couldn’t afford to let our users down. We found ourselves working 20 hours a day, catering to thousands of users, enduring physical fatigue, engaging in talks with VCs, and participating in conferences. Sudden Downturn: Our pinnacle was abruptly interrupted by the war in Ukraine (next macroeconomic shot straight in the face, lucky guy), a precipitous drop in cryptocurrency value, and swiftly emerging competition. By this time, there were 5–8 comparable tools had infiltrated the market. It was a challenging period as we continually stumbled upon new rivals. They immediately embarked on swift fundraising endeavors — a strategy we overlooked, which in retrospect was a mistake. Although our product was superior, the competitors’ rapid advancement and our insufficient funds for expeditious scaling posed significant challenges. Nonetheless, we made a good decision. We sold the product (exit) to competitors. The revenue from “exit” compensated for all the losses, leaving us with enough rest. We were a small team without substantial budgets for rapid development, and the risk of forming new teams without money to survive for more than 1–2 months was irresponsible. You have to believe me that this decision consumed us sleepless nights. Finally, we sold it. They turned off our app but took algorithms and users. Whether you believe it or not, after several months of toiling day and night, experiencing burnout, growing weary of the topic, and gaining an extra 15 kg in weight, we finally found our freedom… The exit wasn’t incredibly profitable, but we knew they had outdone us. The exit covered all our expenses and granted us a well-deserved rest for the subsequent quarter. It was an insane ride. Despite the uncertainty, stress, struggles, and sleepless nights, the story and experience will remain etched in my memory for the rest of my life. Swift Takeaways: Comprehending User Needs: Do you fully understand the product-market fit? Is your offering just an accessory or does it truly satisfy the user’s needs? The Power of Viral Marketing: Take inspiration from giants like Snapchat, ChatGPT, and Clubhouse. While your product might not attain the same scale (but remember, never say never…), the closer your concept is to theirs, the easier your journey will be. If your user is motivated to text a friend saying, “Hey, check out how cool this is” (like sharing ChatGPT), then you’re on the best track. Really. Even if it doesn’t seem immediately evident, there could be a way to incorporate this into your product. Keep looking until you find it. Niche targeting — the more specific and tailored your product is to a certain audience, the easier your journey will be People love buying from people — establishing a personal brand and associating yourself with the product can make things easier. Value: Seek to understand why users engage with your product and keep returning. The more specific and critical the issue you’re aiming to solve, the easier your path will be. Consider your offerings in terms of products and services and focus on sales and marketing, regardless of personal sentiments. These are just a few points, I plan to elaborate on all of them in a separate article. Many products undergo years of development in search of market fit, refining the user experience, and more. And guess what? There’s absolutely nothing wrong with that. Each product and market follows its own rules. Many startups have extensive histories before they finally make their mark (for instance, OpenAI). This entire journey spanned maybe 6–8 months. I grasped and capitalized on the opportunity, but we understood from the start that establishing a startup carried a significant risk, and our crypto product was 10 times riskier. Was it worth it? Given my passion for product development — absolutely. Was it profitable? — No, considering the hours spent — we lose. Did it provide a stable, problem-free life — nope. Did this entire adventure offer a wealth of happiness, joy, and unforgettable experiences — definitely yes. One thing is certain — we’ve amassed substantial experience and it’s not over yet :) So, what lies ahead? Chapter 7: Reverting to the contractor, developing a product for a crypto StartupReturning to the past, we continue our journey… I had invested substantial time and passion into the tech rescue mission product. I came on board as the technical Team Leader of a startup that had garnered over $20M in seed round funding, affiliated with the realm of cryptocurrencies. The investors were individuals with extensive backgrounds in the crypto world. My role was primarily technical, and there was an abundance of work to tackle. I was fully immersed, and genuinely devoted to the role. I was striving for excellence, knowing that if we secured another round of financing, the startup would accelerate rapidly. As for the product and marketing, I was more of an observer. After all, there were marketing professionals with decades of experience on board. These were individuals recruited from large crypto-related firms. I had faith in them, kept an eye on their actions, and focused on my own responsibilities. However, the reality was far from satisfactory. On the last day, the principal investor for the Series A round withdrew. The board made the tough decision to shut down. It was a period of intense observation and gaining experience in product management. This was a very brief summary of the last 10 years. And what next? (Last) Chapter 8: To be announced — Product Owner / Product Consultant / Strategist / CTO After spending countless hours and days deliberating my next steps, one thing is clear: My aspiration is to continue traversing the path of software product development, with the hopeful anticipation that one day, I might ride the crest of the next big wave and ascend to the prestigious status of a unicorn company. I find myself drawn to the process of building products, exploring product-market fit, strategizing, engaging in software development, seeking out new opportunities, networking, attending conferences, and continuously challenging myself by understanding the market and its competitive landscape. Product Owner / Product Consultant / CTO / COO: I’m not entirely sure how to categorize this role, as I anticipate that it will largely depend on the product to which I will commit myself fully. My idea is to find one startup/company that wants to build a product / or already has a product, want to speed up, or simply doesn’t know what’s next. Alternatively, I could be a part of an established company with a rich business history, which intends to invest in digitization and technological advancements. The goal would be to enrich their customer experience by offering complementary digital products Rather than initiating a new venture from ground zero with the same team, I am receptive to new challenges. I am confident that my past experiences will prove highly beneficial for the founders of promising, burgeoning startups that already possess a product, or are in the initial phases of development. ‘Consultant’ — I reckon we interpret this term differently. My aim is to be completely absorbed in a single product, crafting funnels, niches, strategies, and all that is necessary to repeatedly achieve the ‘product-market fit’ and significant revenue. To me, ‘consultant’ resonates more akin to freelancing than being an employee. My current goal is to kickstart as a consultant and aide, dealing with facilitating startups in their journey from point A to B. Here are two theoretical scenarios to illustrate my approach: Scenario 1: (Starting from point A) You have a product but struggle with marketing, adoption, software, strategy, sales, fundraising, or something else. I conduct an analysis and develop a strategy to reach point B. I take on the “dirty work” and implement necessary changes, including potential pivots or shifts (going all-in) to guide the product to point B. The goal is to reach point B, which could involve achieving a higher valuation, expanding the user base, increasing sales, or generating monthly revenue, among other metrics. Scenario 2: (Starting from point A) You have a plan or idea but face challenges with marketing, adoption, strategy, software, sales, fundraising, or something else. I analyze the situation and devise a strategy to reach point B. I tackle the necessary tasks, build the team, and overcome obstacles to propel the product to point B. I have come across the view that finding the elusive product-market fit is the job of the founder, and it’s hard for me to disagree. However, I believe that my support and experiences can help save money, many failures, and most importantly, time. I have spent a great deal of time learning from my mistakes, enduring failure after failure, and even had no one to ask for support or opinion, which is why I offer my help. Saving even a couple of years, realistically speaking, seems like a value I’m eager to provide… I invite you to share your thoughts and insights on these scenarios :) Closing Remarks: I appreciate your time and effort in reaching this point. This has been my journey, and I wouldn’t change it for the world. I had an extraordinary adventure, and now I’m ready for the next exciting battle with the market and new software products. While my entire narrative is centered around startups, especially the ones I personally built, I’m planning to share more insights drawn from all of my experiences, not just those as a co-founder. If you’re currently developing your product or even just considering the idea, I urge you to reach out to me. Perhaps together, we can create something monumental :) Thank you for your time and insights. I eagerly look forward to engaging in discussions and hearing your viewpoints. Please remember to like and subscribe. Nothing motivates to write more than positive feedback :) Matt.

36 startup ideas found by analyzing podcasts (problem, solution & source episode)
reddit
LLM Vibe Score0
Human Vibe Score1
joepigeonThis week

36 startup ideas found by analyzing podcasts (problem, solution & source episode)

Hey, I've been a bit of a podcast nerd for a long time. Around a year ago I began experimenting with transcription of podcasts for a SaaS I was running. I realized pretty quickly that there's a lot of knowledge and value in podcast discussions that is for all intents and purposes entirely unsearchable or discoverable to most people. I ended up stopping work on that SaaS product (party for lack of product/market fit, and partly because podcasting was far more interesting), and focusing on the podcast technology full-time instead. I'm a long-time lurker and poster of r/startups and thought this would make for some interesting content and inspiration for folks. Given I'm in this space, have millions of transcripts, and transcribe thousands daily... I've been exploring fun ways to expose some of the interesting knowledge and conversations taking place that utilize our own data/API. I'm a big fan of the usual startup podcasts (My First Million, Greg Isenberg, etc. etc.) and so I built an automation that turns all of the startup ideas discussed into a weekly email digest. I always struggle to listen to as many episodes as I'd actually like to, so I thought I'd summarise the stuff I care about instead (startup opportunities being discussed). I thought it would be interesting to post some of the ideas extracted so far. They range from being completely whacky and blue sky, to pretty boring but realistic. A word of warning before anyone complains – this is a big mixture of tech, ai, non-tech, local services, etc. ideas: Some of the ideas are completely mundane, but realistic (e.g. local window cleaning service) Some of the ideas are completely insane, blue sky, but sound super interesting Here's the latest 36 ideas: |Idea Name|Problem|Solution|Source| |:-|:-|:-|:-| |SalesForce-as-a-Service - White Label Enterprise Sales Teams|White-label enterprise sales teams for B2B SaaS. Companies need sales but can't hire/train. Recruit retail sellers, train for tech, charge 30% of deals closed.|Create a white-label enterprise sales team by recruiting natural salespeople from retail and direct sales backgrounds (e.g. mall kiosks, cutco knives). Train them specifically in B2B SaaS sales techniques and processes. Offer this trained sales force to tech companies on a contract basis.|My First Million - "Life Hacks From The King of Introverts + 7 Business Ideas| |TechButler - Mobile Device Maintenance Service|Mobile tech maintenance service. Clean/optimize devices, improve WiFi, basic support. $100/visit to homes. Target affluent neighborhoods.|Mobile tech support service providing in-home device cleaning, optimization, and setup. Focus on common issues like WiFi improvement, device maintenance, and basic tech support.|My First Million - "Life Hacks From The King of Introverts + 7 Business Ideas| |MemoryBox - At-Home Video Digitization Service|Door-to-door VHS conversion service. Parents have boxes of old tapes. Pick up, digitize, deliver. $30/tape with minimum order. Going extinct.|Door-to-door VHS to digital conversion service that handles everything from pickup to digital delivery. Make it extremely convenient for customers to preserve their memories.|My First Million - "Life Hacks From The King of Introverts + 7 Business Ideas| |Elite Match Ventures - Success-Based Luxury Matchmaking|High-end matchmaking for 50M+ net worth individuals. Only charge $1M+ when they get married. No upfront fees. Extensive vetting process.|Premium matchmaking service exclusively for ultra-high net worth individuals with a pure contingency fee model - only get paid ($1M+) upon successful marriage. Focus on quality over quantity with extensive vetting and personalized matching.|My First Million - "Life Hacks From The King of Introverts + 7 Business Ideas| |LocalHost - Simple Small Business Websites|Simple WordPress sites for local businesses. $50/month includes hosting, updates, security. Target restaurants and shops. Recurring revenue play.|Simplified web hosting and WordPress management service targeting local small businesses. Focus on basic sites with standard templates, ongoing maintenance, and reliable support for a fixed monthly fee.|My First Million - "Life Hacks From The King of Introverts + 7 Business Ideas| |VoiceJournal AI - Voice-First Smart Journaling|Voice-to-text journaling app with AI insights. 8,100 monthly searches. $15/month subscription. Partners with journaling YouTubers.|AI-powered journaling app that combines voice recording, transcription, and intelligent insights. Users can speak their thoughts, which are automatically transcribed and analyzed for patterns, emotions, and actionable insights.|Where It Happens - "7 $1M+ AI startup ideas you can launch tomorrow with $0"| |AIGenAds - AI-Generated UGC Content Platform|AI platform turning product briefs into UGC-style video ads. Brands spending $500/video for human creators. Generate 100 variations for $99/month.|AI platform that generates UGC-style video ads using AI avatars and scripting. System would allow rapid generation of multiple ad variations at a fraction of the cost. Platform would use existing AI avatar technology combined with script generation to create authentic-looking testimonial-style content.|Where It Happens - "7 $1M+ AI startup ideas you can launch tomorrow with $0"| |InfographAI - Automated Infographic Generation Platform|AI turning blog posts into branded infographics. Marketers spending hours on design. $99/month unlimited generation.|AI-powered platform that automatically converts blog posts and articles into visually appealing infographics. System would analyze content, extract key points, and generate professional designs using predefined templates and brand colors.|Where It Happens - "7 $1M+ AI startup ideas you can launch tomorrow with $0"| |KidFinance - Children's Financial Education Entertainment|Children's media franchise teaching financial literacy. Former preschool teacher creating 'Dora for money'. Books, videos, merchandise potential.|Character-driven financial education content for kids, including books, videos, and potentially TV show. Focus on making money concepts fun and memorable.|The Side Hustle Show - "How a Free Challenge Turned Into a $500,000 a Year Business (Greatest Hits)"| |FinanceTasker - Daily Financial Task Challenge|Free 30-day financial challenge with daily action items. People overwhelmed by money management. Makes $500k/year through books, speaking, and premium membership.|A free 30-day financial challenge delivering one simple, actionable task per day via email. Each task includes detailed scripts and instructions. Participants join a Facebook community for support and accountability. The program focuses on quick wins to build momentum. Automated delivery allows scaling.|The Side Hustle Show - "How a Free Challenge Turned Into a $500,000 a Year Business (Greatest Hits)"| |FinanceAcademy - Expert Financial Training Platform|Premium financial education platform. $13/month for expert-led courses and live Q&As. 4000+ members generating $40k+/month.|Premium membership site with expert-led courses, live Q&As, and community support. Focus on specific topics like real estate investing, business creation, and advanced money management.|The Side Hustle Show - "How a Free Challenge Turned Into a $500,000 a Year Business (Greatest Hits)"| |SecurityFirst Compliance - Real Security + Compliance Platform|Security-first compliance platform built by hackers. Companies spending $50k+ on fake security. Making $7M/year showing why current solutions don't work.|A compliance platform built by security experts that combines mandatory compliance requirements with real security measures. The solution includes hands-on security testing, expert guidance, and a focus on actual threat prevention rather than just documentation. It merges traditional compliance workflows with practical security implementations.|In the Pit with Cody Schneider| |LinkedInbound - Automated Professional Visibility Engine|LinkedIn automation for inbound job offers. Professionals spending hours on manual outreach. $99/month per job seeker.|Automated system for creating visibility and generating inbound interest on LinkedIn through coordinated profile viewing and engagement. Uses multiple accounts to create visibility patterns that trigger curiosity and inbound messages.|In the Pit with Cody Schneider| |ConvoTracker - Community Discussion Monitoring Platform|Community discussion monitoring across Reddit, Twitter, HN. Companies missing sales opportunities. $499/month per brand tracked.|Comprehensive monitoring system that tracks competitor mentions and industry discussions across multiple platforms (Reddit, Twitter, Hacker News, etc.) with automated alerts and engagement suggestions.|In the Pit with Cody Schneider| |ContentAds Pro - Smart Display Ad Implementation|Display ad implementation service for content creators. Bloggers losing thousands in ad revenue monthly. Makes $3-5k per site setup plus ongoing optimization fees.|Implementation of professional display advertising through networks like Mediavine that specialize in optimizing ad placement and revenue while maintaining user experience. Include features like turning off ads for email subscribers and careful placement to minimize impact on core metrics.|The Side Hustle Show - "636: Is Business Coaching Worth It? A Look Inside the last 12 months of Side Hustle Nation"| |MoneyAppReviews - Professional Side Hustle App Testing|Professional testing service for money-making apps. People wasting time on low-paying apps. Makes $20k/month from affiliate commissions and ads.|Professional app testing service that systematically reviews money-making apps and creates detailed, honest reviews including actual earnings data, time investment, and practical tips.|The Side Hustle Show - "636: Is Business Coaching Worth It? A Look Inside the last 12 months of Side Hustle Nation"| |LightPro - Holiday Light Installation Service|Professional Christmas light installation service. Homeowners afraid of ladders. $500-2000 per house plus storage.|Professional Christmas light installation service targeting residential and commercial properties. Full-service offering including design, installation, maintenance, removal and storage. Focus on safety and premium aesthetic results.|The Side Hustle Show - "639: 30 Ways to Make Extra Money for the Holidays"| |FocusMatch - Research Participant Marketplace|Marketplace connecting companies to paid research participants. Companies spending weeks finding people. $50-150/hour per study.|Online platform connecting companies directly with paid research participants. Participants create detailed profiles and get matched to relevant studies. Companies get faster access to their target demographic while participants earn money sharing opinions.|The Side Hustle Show - "639: 30 Ways to Make Extra Money for the Holidays"| |SolarShine Pro - Specialized Solar Panel Cleaning Service|Solar panel cleaning service using specialized equipment. Panels lose 50% efficiency when dirty. $650 per job, automated scheduling generates $18k/month from repeat customers.|Professional solar panel cleaning service using specialized deionized water system and European cleaning equipment. Includes automated 6-month scheduling, professional liability coverage, and warranty-safe cleaning processes. Service is bundled with inspection and performance monitoring.|The UpFlip Podcast - "156. $18K/Month with This ONE Service — Niche Business Idea"| |ExteriorCare Complete - One-Stop Exterior Maintenance Service|One-stop exterior home cleaning service (solar, windows, gutters, bird proofing). Automated scheduling. $650 average ticket. 60% repeat customers on 6-month contracts.|All-in-one exterior cleaning service offering comprehensive maintenance packages including solar, windows, gutters, roof cleaning and bird proofing. Single point of contact, consistent quality, and automated scheduling for all services.|The UpFlip Podcast - "156. $18K/Month with This ONE Service — Niche Business Idea"| |ContentMorph - Automated Cross-Platform Content Adaptation|AI platform converting blog posts into platform-optimized social content. Marketing teams spending 5hrs/post on manual adaptation. $199/mo per brand with 50% margins.|An AI-powered platform that automatically transforms long-form content (blog posts, podcasts, videos) into platform-specific formats (Instagram reels, TikToks, tweets). The system would preserve brand voice while optimizing for each platform's unique requirements and best practices.|Entrepreneurs on Fire - "Digital Threads: The Entrepreneur Playbook for Digital-First Marketing with Neal Schaffer"| |MarketerMatch - Verified Digital Marketing Talent Marketplace|Marketplace for pre-vetted digital marketing specialists. Entrepreneurs spending 15hrs/week on marketing tasks. Platform takes 15% commission averaging $900/month per active client.|A specialized marketplace exclusively for digital marketing professionals, pre-vetted for specific skills (video editing, social media, SEO, etc.). Platform includes skill verification, portfolio review, and specialization matching.|Entrepreneurs on Fire - "Digital Threads: The Entrepreneur Playbook for Digital-First Marketing with Neal Schaffer"| |Tiger Window Cleaning - Premium Local Window Service|Local window cleaning service targeting homeowners. Traditional companies charging 2x market rate. Making $10k/month from $200 initial investment.|Local window cleaning service combining competitive pricing ($5/pane), excellent customer service, and quality guarantees. Uses modern tools like water-fed poles for efficiency. Implements systematic approach to customer communication and follow-up.|The Side Hustle Show - "630: How this College Student’s Side Hustle Brings in $10k a Month"| |RealViz3D - Real Estate Visualization Platform|3D visualization service turning architectural plans into photorealistic renderings for real estate agents. Agents struggling with unbuilt property sales. Making $30-40k/year per operator.|Professional 3D modeling and rendering service that creates photorealistic visualizations of properties before they're built or renovated. The service transforms architectural plans into immersive 3D representations that show lighting, textures, and realistic details. This helps potential buyers fully understand and connect with the space before it physically exists.|Side Hustle School - "#2861 - TBT: An Architect’s Side Hustle in 3D Real Estate Modeling"| |Somewhere - Global Talent Marketplace|Platform connecting US companies with vetted overseas talent. Tech roles costing $150k locally filled for 50% less. Grew from $15M to $52M valuation in 9 months.|Platform connecting US companies with pre-vetted overseas talent at significantly lower rates while maintaining high quality. Handles payments, contracts, and quality assurance to remove friction from global hiring.|My First Million - "I Lost Everything Twice… Then Made $26M In 18 Months| |GymLaunch - Rapid Gym Turnaround Service|Consultants flying to struggling gyms to implement proven member acquisition systems. Gym owners lacking sales expertise. Made $100k in first 21 days.|Expert consultants fly in to implement proven member acquisition systems, train staff, and rapidly fill gyms with new members. The service combines sales training, marketing automation, and proven conversion tactics to transform struggling gyms into profitable businesses within weeks.|My First Million - "I Lost Everything Twice… Then Made $26M In 18 Months| |PublishPlus - Publishing Backend Monetization|Backend monetization system for publishing companies. One-time customers becoming recurring revenue. Grew business from $2M to $110M revenue.|Add complementary backend products and services to increase customer lifetime value. Develop software tools and additional services that natural extend from initial publishing product. Focus on high-margin recurring revenue streams.|My First Million - "I Lost Everything Twice… Then Made $26M In 18 Months| |WelcomeBot - Automated Employee Onboarding Platform|Automated employee welcome platform. HR teams struggling with consistent onboarding. $99/month per 100 employees.|An automated onboarding platform that creates personalized welcome experiences through pre-recorded video messages, scheduled check-ins, and automated swag delivery. The platform would ensure consistent high-quality onboarding regardless of timing or location.|Entrepreneurs on Fire - "Free Training on Building Systems and Processes to Scale Your Business with Chris Ronzio: An EOFire Classic from 2021"| |ProcessBrain - Business Knowledge Documentation Platform|SaaS platform turning tribal knowledge into documented processes. Business owners spending hours training new hires. $199/month per company.|A software platform that makes it easy to document and delegate business processes and procedures. The platform would include templates, guided documentation flows, and tools to easily share and update procedures. It would help businesses create a comprehensive playbook of their operations.|Entrepreneurs on Fire - "Free Training on Building Systems and Processes to Scale Your Business with Chris Ronzio: An EOFire Classic from 2021"| |TradeMatch - Modern Manufacturing Job Marketplace|Modern job board making manufacturing sexy again. Factory jobs paying $40/hr but can't recruit. $500 per successful referral.|A specialized job marketplace and recruitment platform focused exclusively on modern manufacturing and trade jobs. The platform would combine TikTok-style content marketing, referral programs, and modern UX to make manufacturing jobs appealing to Gen Z and young workers. Would leverage existing $500 referral fees and industry demand.|My First Million - "He Sold His Company For $15M, Then Got A Job At McDonald’s"| |GroundLevel - Executive Immersion Program|Structured program putting CEOs in front-line jobs. Executives disconnected from workers. $25k per placement.|A structured program that places executives and founders in front-line jobs (retail, warehouse, service) for 2-4 weeks with documentation and learning framework. Similar to Scott Heiferman's McDonald's experience but productized.|My First Million - "He Sold His Company For $15M, Then Got A Job At McDonald’s"| |OneStepAhead - Micro-Mentorship Marketplace|Marketplace for 30-min mentorship calls with people one step ahead. Professionals seeking specific guidance. Takes 15% of session fees.|MicroMentor Marketplace - Platform connecting people with mentors who are just one step ahead in their journey for focused, affordable micro-mentorship sessions.|Entrepreneurs on Fire - "How to Create an Unbroken Business with Michael Unbroken: An EOFire Classic from 2021"| |VulnerableLeader - Leadership Authenticity Training Platform|Leadership vulnerability training platform. Leaders struggling with authentic communication. $2k/month per company subscription.|Leadership Vulnerability Platform - A digital training platform combining assessment tools, guided exercises, and peer support to help leaders develop authentic communication skills. The platform would include real-world scenarios, video coaching, and measurable metrics for tracking leadership growth through vulnerability.|Entrepreneurs on Fire - "How to Create an Unbroken Business with Michael Unbroken: An EOFire Classic from 2021"| |NetworkAI - Smart Network Intelligence Platform|AI analyzing your network to find hidden valuable connections. Professionals missing opportunities in existing contacts. $49/month per user.|AI Network Navigator - Smart tool that analyzes your professional network across platforms, identifies valuable hidden connections, and suggests specific actionable ways to leverage relationships for mutual benefit.|Entrepreneurs on Fire - "How to Create an Unbroken Business with Michael Unbroken: An EOFire Classic from 2021"| |Porch Pumpkins - Seasonal Decoration Service|Full-service porch pumpkin decoration. Homeowners spend $300-1350 per season. One operator making $1M in 8 weeks seasonal revenue.|Full-service seasonal porch decoration service focused on autumn/Halloween, including design, installation, maintenance, and removal. Offering premium curated pumpkin arrangements with various package tiers.|My First Million - "The guy who gets paid $80K/yr to do nothing"| |Silent Companion - Professional Presence Service|Professional silent companions for lonely people. Huge problem in Japan/globally. $68/session, $80k/year per companion. Non-sexual, just presence.|A professional companion service where individuals can rent a non-judgmental, quiet presence for various activities. The companion provides silent company without the pressure of conversation or social performance. They accompany clients to events, meals, or just sit quietly together.|My First Million - "The guy who gets paid $80K/yr to do nothing"| Hope this is useful. If anyone would like to ensure I include any particular podcasts or episodes etc. in future posts, very happy to do so. I'll generally send \~5 ideas per week in a short weekly digest format (you can see the format I'd usually use in here: podcastmarketwatch.beehiiv.com). I find it mindblowing that the latest models with large context windows make it even possible to analyze full transcripts at such scale. It's a very exciting time we're living through! Would love some feedback on this stuff, happy to iterate and improve the analysis/ideas... or create a new newsletter on a different topic if anyone would like. Cheers!

From "There's an App for that" to "There's YOUR App for that" - AI workflows will transform generic apps into deeply personalized experiences
reddit
LLM Vibe Score0
Human Vibe Score1
Important-Ostrich69This week

From "There's an App for that" to "There's YOUR App for that" - AI workflows will transform generic apps into deeply personalized experiences

I will not promote. For the past decade mobile apps were a core element of daily life for entertainment, productivity and connectivity. However, as the ecosystem saturated the general desire to download "just one more app" became apprehensive. There were clear monopolistic winners in different categories, such as Instagram and TikTok, which completely captured the majority of people's screentime. The golden age of creating indie apps and becoming a millionaire from them was dead. Conceptual models of these popular apps became ingrained in the general consciousness, and downloading new apps where re-learning new UI layouts was required, became a major friction point. There is high reluctance to download a new app rather than just utilizing the tooling of the growing market share of the existing winners. Content marketing and white labeled apps saw a resurgence of new app downloads, as users with parasympathetic relationships with influencers could be more easily persuaded to download them. However, this has led to a series of genericized tooling that lacks the soul of the early indie developer apps from the 2010s (Flappy bird comes to mind). A seemingly grim spot to be in, until everything changed on November 30th 2022. Sam Altman, Ilya Sutskever and team announced chatGPT, a Large Language Model that was the first publicly available generative AI tool. The first non-deterministic tool that could reason probablisitically in a similar (if flawed) way, to the human mind. At first, it was a clear paradigm shift in the world of computing, this was obvious from the fact that it climbed to 1 Million users within the first 5 days of its launch. However, despite the insane hype around the AI, its utility was constrained to chatbot interfaces for another year or more. As the models reasoning abilities got better and better, engineers began to look for other ways of utilizing this new paradigm shift, beyond chatbots. It became clear that, despite the powerful abilities to generate responses to prompts, the LLMs suffered from false hallucinations with extreme confidence, significantly impacting the reliability of their use, in search, coding and general utility. Retrieval Augmented Generation (RAG) was coined to provide a solution to this. Now, the LLM would apply a traditional search for data, via a database, a browser or other source of truth, and then feed that information into the prompt as it generates, allowing for more accurate results. Furthermore, it became clear that you could enhance an LLM by providing them metadata to interact with tools such as APIs for other services, allowing LLMs to perform actions typically reserved for humans, like fetching data, manipulating it and acting as an independent Agent. This prompted engineers to start treating LLMs, not as a database and a search engine, but rather a reasoning system, that could be part of a larger system of inputs and feedback to handle workflows independently. These "AI Agents" are poised to become the core technology in the next few years for hyper-personalizing and automating processes for specific users. Rather than having a generic B2B SaaS product that is somewhat useful for a team, one could standup a modular system of Agents that can handle the exactly specified workflow for that team. Frameworks such as LlangChain and LLamaIndex will help enable this for companies worldwide. The power is back in the hands of the people. However, it's not just big tech that is going to benefit from this revolution. AI Agentic workflows will allow for a resurgence in personalized applications that work like personal digital employee's. One could have a Personal Finance agent keeping track of their budgets, a Personal Trainer accountability coaching you making sure you meet your goals, or even a silly companion that roasts you when you're procrastinating. The options are endless ! At the core of this technology is the fact that these agents will be able to recall all of your previous data and actions, so they will get better at understanding you and your needs as a function of time. We are at the beginning of an exciting period in history, and I'm looking forward to this new period of deeply personalized experiences. What are your thoughts ? Let me know in the comments !

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies) (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies) (I will not promote)

AI Palette is an AI-driven platform that helps food and beverage companies predict emerging product trends. I had the opportunity recently to sit down with the founder to get his advice on building an AI-first startup, which he'll be going through in this post. (I will not promote) About AI Palette: Co-founders: >!2 (Somsubhra GanChoudhuri, Himanshu Upreti)!!100+!!$12.7M USD!!AI-powered predictive analytics for the CPG (Consumer Packaged Goods) industry!!Signed first paying customer in the first year!!65+ global brands, including Cargill, Diageo, Ajinomoto, Symrise, Mondelez, and L’Oréal, use AI Palette!!Every new product launched has secured a paying client within months!!Expanded into Beauty & Personal Care (BPC), onboarding one of India’s largest BPC companies within weeks!!Launched multiple new product lines in the last two years, creating a unified suite for brand innovation!Identify the pain points in your industry for ideas* When I was working in the flavour and fragrance industry, I noticed a major issue CPG companies faced: launching a product took at least one to two years. For instance, if a company decided today to launch a new juice, it wouldn’t hit the market until 2027. This long timeline made it difficult to stay relevant and on top of trends. Another big problem I noticed was that companies relied heavily on market research to determine what products to launch. While this might work for current consumer preferences, it was highly inefficient since the product wouldn’t actually reach the market for several years. By the time the product launched, the consumer trends had already shifted, making that research outdated. That’s where AI can play a crucial role. Instead of looking at what consumers like today, we realised that companies should use AI to predict what they will want next. This allows businesses to create products that are ahead of the curve. Right now, the failure rate for new product launches is alarmingly high, with 8 out of 10 products failing. By leveraging AI, companies can avoid wasting resources on products that won’t succeed, leading to better, more successful launches. Start by talking to as many industry experts as possible to identify the real problems When we first had the idea for AI Palette, it was just a hunch, a gut feeling—we had no idea whether people would actually pay for it. To validate the idea, we reached out to as many people as we could within the industry. Since our focus area was all about consumer insights, we spoke to professionals in the CPG sector, particularly those in the insights departments of CPG companies. Through these early conversations, we began to see a common pattern emerge and identified the exact problem we wanted to solve. Don’t tell people what you’re building—listen to their frustrations and challenges first. Going into these early customer conversations, our goal was to listen and understand their challenges without telling them what we were trying to build. This is crucial as it ensures that you can gather as much data about the problem to truly understand it and that you aren't biasing their answers by showing your solution. This process helped us in two key ways: First, it validated that there was a real problem in the industry through the number of people who spoke about experiencing the same problem. Second, it allowed us to understand the exact scale and depth of the problem—e.g., how much money companies were spending on consumer research, what kind of tools they were currently using, etc. Narrow down your focus to a small, actionable area to solve initially. Once we were certain that there was a clear problem worth solving, we didn’t try to tackle everything at once. As a small team of two people, we started by focusing on a specific area of the problem—something big enough to matter but small enough for us to handle. Then, we approached customers with a potential solution and asked them for feedback. We learnt that our solution seemed promising, but we wanted to validate it further. If customers are willing to pay you for the solution, it’s a strong validation signal for market demand. One of our early customer interviewees even asked us to deliver the solution, which we did manually at first. We used machine learning models to analyse the data and presented the results in a slide deck. They paid us for the work, which was a critical moment. It meant we had something with real potential, and we had customers willing to pay us before we had even built the full product. This was the key validation that we needed. By the time we were ready to build the product, we had already gathered crucial insights from our early customers. We understood the specific information they wanted and how they wanted the results to be presented. This input was invaluable in shaping the development of our final product. Building & Product Development Start with a simple concept/design to validate with customers before building When we realised the problem and solution, we began by designing the product, but not by jumping straight into coding. Instead, we created wireframes and user interfaces using tools like InVision and Figma. This allowed us to visually represent the product without the need for backend or frontend development at first. The goal was to showcase how the product would look and feel, helping potential customers understand its value before we even started building. We showed these designs to potential customers and asked for feedback. Would they want to buy this product? Would they pay for it? We didn’t dive into actual development until we found a customer willing to pay a significant amount for the solution. This approach helped us ensure we were on the right track and didn’t waste time or resources building something customers didn’t actually want. Deliver your solution using a manual consulting approach before developing an automated product Initially, we solved problems for customers in a more "consulting" manner, delivering insights manually. Recall how I mentioned that when one of our early customer interviewees asked us to deliver the solution, we initially did it manually by using machine learning models to analyse the data and presenting the results to them in a slide deck. This works for the initial stages of validating your solution, as you don't want to invest too much time into building a full-blown MVP before understanding the exact features and functionalities that your users want. However, after confirming that customers were willing to pay for what we provided, we moved forward with actual product development. This shift from a manual service to product development was key to scaling in a sustainable manner, as our building was guided by real-world feedback and insights rather than intuition. Let ongoing customer feedback drive iteration and the product roadmap Once we built the first version of the product, it was basic, solving only one problem. But as we worked closely with customers, they requested additional features and functionalities to make it more useful. As a result, we continued to evolve the product to handle more complex use cases, gradually developing new modules based on customer feedback. Product development is a continuous process. Our early customers pushed us to expand features and modules, from solving just 20% of their problems to tackling 50–60% of their needs. These demands shaped our product roadmap and guided the development of new features, ultimately resulting in a more complete solution. Revenue and user numbers are key metrics for assessing product-market fit. However, critical mass varies across industries Product-market fit (PMF) can often be gauged by looking at the size of your revenue and the number of customers you're serving. Once you've reached a certain critical mass of customers, you can usually tell that you're starting to hit product-market fit. However, this critical mass varies by industry and the type of customers you're targeting. For example, if you're building an app for a broad consumer market, you may need thousands of users. But for enterprise software, product-market fit may be reached with just a few dozen key customers. Compare customer engagement and retention with other available solutions on the market for product-market fit Revenue and the number of customers alone isn't always enough to determine if you're reaching product-market fit. The type of customer and the use case for your product also matter. The level of engagement with your product—how much time users are spending on the platform—is also an important metric to track. The more time they spend, the more likely it is that your product is meeting a crucial need. Another way to evaluate product-market fit is by assessing retention, i.e whether users are returning to your platform and relying on it consistently, as compared to other solutions available. That's another key indication that your solution is gaining traction in the market. Business Model & Monetisation Prioritise scalability Initially, we started with a consulting-type model where we tailor-made specific solutions for each customer use-case we encountered and delivered the CPG insights manually, but we soon realized that this wasn't scalable. The problem with consulting is that you need to do the same work repeatedly for every new project, which requires a large team to handle the workload. That is not how you sustain a high-growth startup. To solve this, we focused on building a product that would address the most common problems faced by our customers. Once built, this product could be sold to thousands of customers without significant overheads, making the business scalable. With this in mind, we decided on a SaaS (Software as a Service) business model. The benefit of SaaS is that once you create the software, you can sell it to many customers without adding extra overhead. This results in a business with higher margins, where the same product can serve many customers simultaneously, making it much more efficient than the consulting model. Adopt a predictable, simplistic business model for efficiency. Look to industry practices for guidance When it came to monetisation, we considered the needs of our CPG customers, who I knew from experience were already accustomed to paying annual subscriptions for sales databases and other software services. We decided to adopt the same model and charge our customers an annual upfront fee. This model worked well for our target market, aligning with industry standards and ensuring stable, recurring revenue. Moreover, our target CPG customers were already used to this business model and didn't have to choose from a huge variety of payment options, making closing sales a straightforward and efficient process. Marketing & Sales Educate the market to position yourself as a thought leader When we started, AI was not widely understood, especially in the CPG industry. We had to create awareness around both AI and its potential value. Our strategy focused on educating potential users and customers about AI, its relevance, and why they should invest in it. This education was crucial to the success of our marketing efforts. To establish credibility, we adopted a thought leadership approach. We wrote blogs on the importance of AI and how it could solve problems for CPG companies. We also participated in events and conferences to demonstrate our expertise in applying AI to the industry. This helped us build our brand and reputation as leaders in the AI space for CPG, and word-of-mouth spread as customers recognized us as the go-to company for AI solutions. It’s tempting for startups to offer products for free in the hopes of gaining early traction with customers, but this approach doesn't work in the long run. Free offerings don’t establish the value of your product, and customers may not take them seriously. You should always charge for pilots, even if the fee is minimal, to ensure that the customer is serious about potentially working with you, and that they are committed and engaged with the product. Pilots/POCs/Demos should aim to give a "flavour" of what you can deliver A paid pilot/POC trial also gives you the opportunity to provide a “flavour” of what your product can deliver, helping to build confidence and trust with the client. It allows customers to experience a detailed preview of what your product can do, which builds anticipation and desire for the full functionality. During this phase, ensure your product is built to give them a taste of the value you can provide, which sets the stage for a broader, more impactful adoption down the line. Fundraising & Financial Management Leverage PR to generate inbound interest from VCs When it comes to fundraising, our approach was fairly traditional—we reached out to VCs and used connections from existing investors to make introductions. However, looking back, one thing that really helped us build momentum during our fundraising process was getting featured in Tech in Asia. This wasn’t planned; it just so happened that Tech in Asia was doing a series on AI startups in Southeast Asia and they reached out to us for an article. During the interview, they asked if we were fundraising, and we mentioned that we were. As a result, several VCs we hadn’t yet contacted reached out to us. This inbound interest was incredibly valuable, and we found it far more effective than our outbound efforts. So, if you can, try to generate some PR attention—it can help create inbound interest from VCs, and that interest is typically much stronger and more promising than any outbound strategies because they've gone out of their way to reach out to you. Be well-prepared and deliberate about fundraising. Keep trying and don't lose heart When pitching to VCs, it’s crucial to be thoroughly prepared, as you typically only get one shot at making an impression. If you mess up, it’s unlikely they’ll give you a second chance. You need to have key metrics at your fingertips, especially if you're running a SaaS company. Be ready to answer questions like: What’s your retention rate? What are your projections for the year? How much will you close? What’s your average contract value? These numbers should be at the top of your mind. Additionally, fundraising should be treated as a structured process, not something you do on the side while juggling other tasks. When you start, create a clear plan: identify 20 VCs to reach out to each week. By planning ahead, you’ll maintain momentum and speed up the process. Fundraising can be exhausting and disheartening, especially when you face multiple rejections. Remember, you just need one investor to say yes to make it all worthwhile. When using funds, prioritise profitability and grow only when necessary. Don't rely on funding to survive. In the past, the common advice for startups was to raise money, burn through it quickly, and use it to boost revenue numbers, even if that meant operating at a loss. The idea was that profitability wasn’t the main focus, and the goal was to show rapid growth for the next funding round. However, times have changed, especially with the shift from “funding summer” to “funding winter.” My advice now is to aim for profitability as soon as possible and grow only when it's truly needed. For example, it’s tempting to hire a large team when you have substantial funds in the bank, but ask yourself: Do you really need 10 new hires, or could you get by with just four? Growing too quickly can lead to unnecessary expenses, so focus on reaching profitability as soon as possible, rather than just inflating your team or burn rate. The key takeaway is to spend your funds wisely and only when absolutely necessary to reach profitability. You want to avoid becoming dependent on future VC investments to keep your company afloat. Instead, prioritize reaching break-even as quickly as you can, so you're not reliant on external funding to survive in the long run. Team-Building & Leadership Look for complementary skill sets in co-founders When choosing a co-founder, it’s important to find someone with a complementary skill set, not just someone you’re close to. For example, I come from a business and commercial background, so I needed someone with technical expertise. That’s when I found my co-founder, Himanshu, who had experience in machine learning and AI. He was a great match because his technical knowledge complemented my business skills, and together we formed a strong team. It might seem natural to choose your best friend as your co-founder, but this can often lead to conflict. Chances are, you and your best friend share similar interests, skills, and backgrounds, which doesn’t bring diversity to the table. If both of you come from the same industry or have the same strengths, you may end up butting heads on how things should be done. Having diverse skill sets helps avoid this and fosters a more collaborative working relationship. Himanshu (left) and Somsubhra (right) co-founded AI Palette in 2018 Define roles clearly to prevent co-founder conflict To avoid conflict, it’s essential that your roles as co-founders are clearly defined from the beginning. If your co-founder and you have distinct responsibilities, there is no room for overlap or disagreement. This ensures that both of you can work without stepping on each other's toes, and there’s mutual respect for each other’s expertise. This is another reason as to why it helps to have a co-founder with a complementary skillset to yours. Not only is having similar industry backgrounds and skillsets not particularly useful when building out your startup, it's also more likely to lead to conflicts since you both have similar subject expertise. On the other hand, if your co-founder is an expert in something that you're not, you're less likely to argue with them about their decisions regarding that aspect of the business and vice versa when it comes to your decisions. Look for employees who are driven by your mission, not salary For early-stage startups, the first hires are crucial. These employees need to be highly motivated and excited about the mission. Since the salary will likely be low and the work demanding, they must be driven by something beyond just the paycheck. The right employees are the swash-buckling pirates and romantics, i.e those who are genuinely passionate about the startup’s vision and want to be part of something impactful beyond material gains. When employees are motivated by the mission, they are more likely to stick around and help take the startup to greater heights. A litmus test for hiring: Would you be excited to work with them on a Sunday? One of the most important rounds in the hiring process is the culture fit round. This is where you assess whether a candidate shares the same values as you and your team. A key question to ask yourself is: "Would I be excited to work with this person on a Sunday?" If there’s any doubt about your answer, it’s likely not a good fit. The idea is that you want employees who align with the company's culture and values and who you would enjoy collaborating with even outside of regular work hours. How we structure the team at AI Palette We have three broad functions in our organization. The first two are the big ones: Technical Team – This is the core of our product and technology. This team is responsible for product development and incorporating customer feedback into improving the technology Commercial Team – This includes sales, marketing, customer service, account managers, and so on, handling everything related to business growth and customer relations. General and Administrative Team – This smaller team supports functions like finance, HR, and administration. As with almost all businesses, we have teams that address the two core tasks of building (technical team) and selling (commercial team), but given the size we're at now, having the administrative team helps smoothen operations. Set broad goals but let your teams decide on execution What I've done is recruit highly skilled people who don't need me to micromanage them on a day-to-day basis. They're experts in their roles, and as Steve Jobs said, when you hire the right person, you don't have to tell them what to do—they understand the purpose and tell you what to do. So, my job as the CEO is to set the broader goals for them, review the plans they have to achieve those goals, and periodically check in on progress. For example, if our broad goal is to meet a certain revenue target, I break it down across teams: For the sales team, I’ll look at how they plan to hit that target—how many customers they need to sell to, how many salespeople they need, and what tactics and strategies they plan to use. For the technical team, I’ll evaluate our product offerings—whether they think we need to build new products to attract more customers, and whether they think it's scalable for the number of customers we plan to serve. This way, the entire organization's tasks are cascaded in alignment with our overarching goals, with me setting the direction and leaving the details of execution to the skilled team members that I hire.

How to get funding for startup ? I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
wlynncorkThis week

How to get funding for startup ? I will not promote

I will not promote. Software startup based out of Minnesota us. I've built and launched a product that is gaining traction, solving a problem that has frustrated software developers and product teams for years. The problem: Software development is slow, expensive, and full of inefficiencies. Developers spend hours on repetitive coding tasks, project managers struggle with bottlenecks, and businesses waste time translating product requirements into actual code. The solution: My product automates a large portion of software development. It acts as an AI-powered assistant for developers, taking high-level requirements and turning them into functional code while integrating with existing codebases. It can read, understand, and modify software projects in a structured way—cutting development time drastically. The potential: Businesses are always looking for ways to cut costs and speed up development. With the rise of AI, companies are increasingly adopting automation, and this tool fits perfectly into that wave. Imagine a world where software teams are 10x more efficient because AI handles the grunt work, and developers focus on the bigger picture. It’s not about replacing developers—it’s about supercharging them. The current status: The product is live and in use. The user base is growing, and I’ve proven demand. Now, I need to figure out the best funding model to scale—whether that’s bootstrapping, VC, grants, or some hybrid approach. If you have experience in startup funding or have scaled a tech product, I'd love to hear your insights. DM me if you're open to discussing strategies!

Feeling stuck—built a startup, got rejected from YC & IVI, met smarter people, and now I don’t know what to do. ( i will not promote )
I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
vishwa1238This week

Feeling stuck—built a startup, got rejected from YC & IVI, met smarter people, and now I don’t know what to do. ( i will not promote ) I will not promote

I will not promote I don’t even know where to start, but I just feel completely stuck right now. I’m 20 years oldI don’t even know where to start, but I just feel completely stuck right now. I’m 20 years old, have been grinding non-stop for months, and it feels like I have nothing to show for it. I built an AI agent that automates workflows for businesses. I can build tech, but I can’t sell. That’s been my biggest realization recently—I thought building would be enough, but it’s not. I need customers, I need a co-founder, I need to figure out the business side… and I have no idea how. I applied to YC, IVI at ISB, and EF, met a lot of insanely smart people—some were impressed with me and my work, but they were wiser, more experienced, and honestly, just better at all of this than I am. It made me realize how much I don’t know. I got rejected from YC & IVI. 💔 YC didn’t even give much feedback—just a standard rejection. 💔 IVI told me: “You're too young, you need more experience, and you should work with a team before trying to start something.” That hit me hard. I had already been struggling to find a co-founder, and this just made me wonder if I even belong in this space yet. The Frustrating Part? I KNOW my tool Has a Unique Edge. I’m not just another AI automation tool—I know my tool has a strong USP that competitors lack. It has the potential to be an AI employee for businesses, not just another workflow tool. But I still haven’t built the “perfect product” I originally envisioned. And that’s what’s eating at me. I see what it COULD be, but I haven’t made it happen yet. At the same time, the competition in the AI agent space is exploding. YC-backed companies are working on AI agent startups. OpenAI is making huge progress with Operator. Competitors are moving fast, while I feel stuck. I’ve delayed development because I’m unsure whether to double down, pivot, or just move on entirely. Where I’m Stuck Right Now 🔹 Do I keep pushing and try to crack sales somehow? 🔹 Do I join a startup as a founding engineer to get experience, make connections, and learn sales before trying again? 🔹 Do I move to Bangalore, meet founders, and figure out what’s next? 🔹 Do I pivot to something nicher instead of competing in the AI agent race? If so, how do I even find a niche worth pursuing? 🔹 Do I even belong in startups? Or am I just forcing something that’s not working? I feel stuck in a weird middle zone where I’m not a beginner, but I’m also not successful. I’ve done enough to see what’s possible, but not enough to make it real. Every rejection makes me question if I’m even on the right path. I don’t know if I’m posting this for advice or just to get it out of my system. Maybe both. Has anyone else felt like this before? If you’ve been in this situation—how did you figure out whether to keep going or move on? TL;DR: I’m 20, built an AI agent for automating workflows, got rejected from YC & IVI, met insanely smart and experienced people, realized I can build tech but can’t sell, struggling to find a co-founder, AI agent competition is growing, delaying development, confused about the future—don’t know whether to double down, pivot, or move on. The frustrating part? I\ know I have a unique edge that others lack, but I still haven’t built the perfect product I originally envisioned.* edit: removed the tool's name

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
GummyBear8659This week

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)

Hey everyone! Long-time lurker, first-time founder here. I’m reaching out to get feedback on a recent startup experience—what went wrong, what I could have done better, and how I should approach future opportunities. The Background There were three founders in this venture: • Founder A (CEO, 50%) – The product/growth guy who identified the problem space. • Founder B (Me, CTO, 37.5%) – A software engineer with a software dev shop and multiple clients. I wanted to diversify into building my own products but am not inherently a “product person.” • Founder C (COO, 12.5%) – Brought into the mix by Founder A, with the goal of leveraging his network for traction once the product was built. The idea was to create Product X, a solution targeting the SMB space while competitors were moving upmarket. It wasn’t revolutionary—more of a strategic market play. The Initial Plan & My Role • Founder A would define and prioritize product specs, guiding what needed to be built. • I (Founder B) didn’t have time to code myself, so I allocated engineers from my dev shop (which I personally paid for). My stake was adjusted from 32.5% to 37.5% to reflect this contribution. • Founder C was more of an observer early on, planning to help with traction once we had a product ready. We agreed on a 1-year cliff and a 4-year vesting schedule for equity. Where Things Started to Go Wrong • Lack of a Clear Product Roadmap – Founder A was very focused on getting something built fast, but we never signed off on a structured roadmap or milestones. I underestimated the complexity of what was actually needed for customer conversations. • Engineering Expectations vs. Reality – The team (one part-time lead + two full-time juniors from my dev shop) faced early feedback that development was too slow. In response, I ramped up the lead to full-time and added a part-time PM. But Founder A continued pushing for speed, despite real hurdles (OAuth integrations, etc.). • Shifting MVP Goalposts – Midway, Founder A concluded that an MVP wouldn’t cut it—we needed a more complete product to be competitive. This meant more engineering, more delays, and more of my own money spent on development. The Breaking Point Near the 1-year vesting mark, we had an opportunity: a paying client willing to fund an app. I didn’t have devs on the bench, so I asked Founder A to hold off our project briefly while I hired more engineers to avoid stalling either effort. This was the final straw. Founder A (with Founder C somewhat aligned) decided the arrangement wasn’t working—citing past disagreements and the “slowness” issue. The decision was made to end the partnership. Now, Founder A, as majority holder, is requesting a full handover of the code, Founder C is indifferent, and all engineering costs I covered are essentially lost. Key Takeaways (So Far) Crystal-Clear Agreements Upfront – A formalized product roadmap and timeline should’ve been locked in from day one. Business Needs > Engineering Standards – I wanted to build something solid and scalable, but in an early-stage startup, speed to market is king. This was before AI tools became mainstream, so our approach wasn’t as optimized. Don’t Overextend Without Protection – I personally financed all engineering, but without clear safeguards, that investment became a sunk cost. Expenses Must Be Distributed – I was solely covering engineering salaries, which created an imbalance in financial risk. Future partnerships should ensure costs are shared proportionally, rather than one person shouldering the burden. Where I Need Advice Looking back, I want to improve as an engineer, CEO, and co-founder. • What should I have done differently in structuring this partnership? • How do you balance engineering quality with the startup need for speed? • As a dev shop owner, how can I better navigate equity deals where I’m also bringing in engineering resources? I really appreciate everyone who went through this long post and provide any insights from founders, engineers, or anyone who has been in a similar situation. Thanks for reading! ===================================================================== For readers who might be thinking what set this type of expectation? Because I had a dev shop and I thought my co-founders will be understanding of my business circumstance and I was a bit trigger to build a product with a C-exec team, I gave the impression of "unlimited" engineering which I later realized down the line that it was not feasible for me. Something I learned that I have to be more careful with and set expectations accordingly from the very beginning. And from the feedback of the commenters here, I am much more aware what I should offer and how to set expectations, esp. in the early stages of execution. So thank you all! 🙏🏾 EDIT: I would like to thank everyone who contributed to this thread. You not only helped me but future founders who are considering to get into the startup scene!

I started a Tech Startup, and I feel totally STUCK.
reddit
LLM Vibe Score0
Human Vibe Score1
BetAltruistic6556This week

I started a Tech Startup, and I feel totally STUCK.

I made "Visual Love," a Computer Vision/AI-driven matchmaking platform. The idea is that although appearance is one of the biggest factors for starting a relationship, current matchmaking services and dating apps do not have the capability to search for people based on appearance. On Visual Love, you can find your ideal match simply by uploading a picture of your "ideal type." Also, you can connect with someone who thinks of you as their ideal type, simply by uploading your own picture. Or, there might be a perfect (mutually ideal) match. I made this CV/AI algorithm to scan faces, retrieve facial features, and make it possible to find the closest match among millions of others in a second. On average, regular dating app users swipe 8000 times over 8 months until they find their love. On Visual Love, users can find one in a million just in a second. You can try the tech demo on the website if you want to (find the link through my LinkedIn at the bottom of the post; I have to follow the "I will not promote" rule.) I thought this app would have the best chance in Asia, as people care a lot more about appearance in Asia (especially Korea and Japan). Also, my nationality is Korean, and I speak both Korean and Japanese as fluently as I speak English. So I came to Korea, and pitched to a number of VC/AC firms in Korea and Japan, and two of them were typically intersted in making investment. However, they both required me to provide market validation: how much it would cost per user acquisition, how much each user would pay on average, and etc, even after I provided them with a 3-years financial projection including market research based on other dating apps. &#x200B; Everything might be going just as expected, or even better than anticipated, but I'm feeling very stuck now. I am not a business expert, and I don't have much idea on how to proceed from here. The problem is, it wouldn't quite work as expected when there are not many users. If I start with a small group of users, it's not any better than any other dating app. Matching users within a small group doesn't quite reflect the values of Visual Love. So I figured a way around: making a game version of Visual Love targeting 100k to 500k users to work as an initial distribution channel. This version will include finding look-alike celebrities, and solving look-alike face puzzles, and etc. But now, the problem is, I cannot continue this project by myself. I have no social/financial support, and I'm running low on cash. Also, although I'm from Korea, I lived in many different countries. I did my undergraduate in New York (Columbia University) and all my friends are in the US. I don't feel very included here. I can't stop feeling frustrated and distressed :( I'm sure Visual Love can reshape the future of the matchmaking market. But, only if I can continue this project by getting the fund I require. I'm open to any advice, and if you're interested in providing any help or working with me, please contact me through LinkedIn. https://www.linkedin.com/in/don-lee-3853b1264/

Hot Take: Not all your startups need AI forced into them
reddit
LLM Vibe Score0
Human Vibe Score1
bitorsicThis week

Hot Take: Not all your startups need AI forced into them

I'm a final year Computer Engineering student, hence applying for jobs all around. There's this particular trend I've noticed with startups that are coming up these days. That is, even for the absolute basic stuff they'll use 'AI', and they'll think they built something 'revolutionary'. No. You're breaking your product in ways you don't realise. An example, that even some well established companies are guilty of: AI Chatbots You absolutely don't need them and it's an entire gimmick. If you really wanna implement a chatbot, connect the user to an actual person on your end, which I think is not possible if you're at a 'startup' stage. You'll need employees who can handle user queries in real time. If the user really is stuck let them use the 'Contact Us' page. A really close relative of mine is very vocal about the frustration he faces whenever he tries to use the AI Chatbot on any well known e-com website. The only case for AI Chatbot that makes sense is when it's directing the customer to an actual customer support rep if none of the AI's solutions is working for the customer. Even then, implementing a search page for FAQ is extremely easy and user friendly. Another example: AI Interviewer I recently interviewed for a startup, and their whole interviewing process was AI'zed?!?! No real person at the other end, I was answering to their questions which were in video format. They even had a 'mascot' / 'AI interviewer' avatar designed by an AI (AI-ception???). This mascot just text-to-speech'ed all the questions for me to rewind and hear what I missed again. And I had to record video and audio to answer these questions on their platform itself. The entire interview process just could've been a questionnaire, or if you're really concerned on the integrity of the interviewee, just take a few minutes out of your oh-so-busy schedule as a startup owner. Atleast for hiring employees who would make the most impact on your product going ahead. I say the most impact, because (atleast as a developer) the work done by these employees would define how robust your product is, and/or how easily other features can be integrated into the codebase. Trust me, refactoring code later on would only cost you time and money. These resources would rather be more useful in other departments of your startup. The only use case for an AI Interviewer I see is for preparing for an actual interview, provided that feedback is given to the user at the earliest, which you don't need to worry about as a startup owner. So yeah, you're probably better off without integrating AI in your product. Thank you for reading. TLDR; The title; I know AI is the new thing and gets everyone drooling and all, but for the love of God, just focus on what your startup does best and put real people behind it; Integrating AI without human intervention is as good as a broken product; Do your hiring yourself, or through real people, emphasizing on the fact that the people you hire at an early stage will define your growth ahead;

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

We received 25k investment offer, need advice [I will not promote]
reddit
LLM Vibe Score0
Human Vibe Score1
Agreeable_Ad6424This week

We received 25k investment offer, need advice [I will not promote]

We received a $25k for 2.5% on a convertible note offer from a US based investor. The note matures in 18 months with an interest rate of 5%, but the investor said they can extend it further. It’s an AI SaaS in graphic design. We have been bootstrapping till now, and we feel that this money could help us hire better engineers and marketeers, we want to grow it to a good revenue, but don't see it becoming a billion dollar startup as such. Our initial plans were to build it like an indie-hacker, grow it a decent revenue and sell it to someone who can take better care of it. We built it as a side project with full time jobs. We already have decent traction with 10k+ signups and $600+ in revenue per month with <100 dollars spent on marketing. But our AI model costs are high, 0.2 USD per user that we onboard and provide free credits. But we as founders are more interested in another idea that we have been thinking about and see a bigger potential + founder market fit in. The current product is good, and we can foresee that with better hiring and marketing, we can grow our revenue to about 10-20k a month, like a regular online business. What should we do? We don't want to simply let go of the product because it's not that it doesn't work, it's just that we as founders are better fit for something else. We can't sell it yet as the revenue isn't too high and we haven't even incorporated. Is it okay if we think of growing it to 10-20k+ a month and then intend to sell it to someone who can take better care of it? Should we take the investment in such a case, given this investment is definitely gonna help us grow? Process of incorporation will also help us in selling this business later I think?

How I made a high tech salary in my first selling month
reddit
LLM Vibe Score0
Human Vibe Score1
Ok_Negotiation_2587This week

How I made a high tech salary in my first selling month

For over 7 years I worked as a full-stack developer, helping other companies bring their ideas to life. But one day, I thought “Why not try making my own dream come true?”. That’s when I decided to quit my job and start my own journey to becoming an entrepreneur. At first, it wasn’t easy. I didn’t make any money for months and had no idea where to start. I felt lost. Then, I decided to focus on something popular and trending. AI was everywhere, and ChatGPT was the most used AI platform. So I looked into it and I found the OpenAI community forum where people had been asking for features that weren’t being added. That gave me an idea. Why not build those features myself? I created a Chrome extension and I worked on some of the most requested features, like: Downloading the advanced voice mode and messages as MP3 Adding folders to organize chats Saving and reusing prompts Pinning important chats Exporting chats to TXT/JSON files Deleting or archiving multiple chats at once Making chat history searches faster and better It took me about a week to build the first version, and when I published it, the response was incredible. People loved it! Some even said things like, “You’re a lifesaver!” That’s when I realized I had something that could not only help people but also turn into a real business. I kept the first version free to see how people would respond. Many users have been downloading my extension, which prompted Chrome to review it to determine if it qualified for the featured badge. I received the badge, and it has significantly boosted traffic to my extension ever since. After all the positive feedback, I launched a paid version one month ago. A few minutes after publishing it, I made my first sale! That moment was so exciting, and it motivated me to keep going. I already have over 4,000 users and have made more than $4,500 in my first selling month. I’ve decided to release 1-2 new features every month to keep improving the extension based on what users ask for. I also created the same extension for Firefox and Edge users because many people have been asking for it! I also started a Reddit community, where I share updates, sales, discount codes, and ideas for new features. It’s been awesome to connect with users directly and get their feedback. Additionally, I’ve started working on another extension for Claude, which I’m hoping will be as successful as this one. My message to you is this: never give up on your dreams. It might feel impossible at first, but with patience, hard work, and some creativity, you can make it happen. I hope this inspires you to go after what you want. Good luck to all of us!

Just raised and here are the stats (July 2024)
reddit
LLM Vibe Score0
Human Vibe Score0.6
tylersellarsThis week

Just raised and here are the stats (July 2024)

CEO of a startup - bootstrapped for 10 months with a team of 7 - Built a waitlist of $15B AUM (fintech) and here's what it took (with no intended story structure) I didn't want to spam, so I decided to go the old school route and manually write every single email (some copy and paste) In order to feel prepared, I would do my research prior to reaching out, albeit sometimes limited due to time x reward Sent over 350 emails to around 300 funds and we received three yes' (to be honest we received more than three, but they wanted too much equity, or they weren’t a good fit culturally) Pre Seed is different for everyone. Some accept pre revenue others expect 100-300k ARR - this was one of the more frustrating pieces for me, due to the fact that it's extremely subjective to what "Earliest Stage" means to some funds. We're pre-product, we had to remove our CTO in the process due to performance issues, we let go of our front-end, hired a new CTO, hired an AI Engineer, and replaced our front end. It's a numbers game. We received 94% no reply - 5% nos (with meetings) - 1% yes' I used OpenVC as my main resource, highly recommend even their free product. I recommend using discount with a SAFE. Some Angels like it better, some VC’s hate it. You have to be willing to play ball with whoever leads. Mercury for banking, perks (like Carta), and SAFE agreement. All great at Mercury. Raising for a B2B business in an enterprise market is much easier, unless deep tech or science backed. But consumer products right now are not raising pre seed from what I can tell. If it means anything - it's a numbers game. Go get what you deserve, but put in the work because no one will just hand it to you. Love this community, always here to help anyone I can.

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
GummyBear8659This week

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)

Hey everyone! Long-time lurker, first-time founder here. I’m reaching out to get feedback on a recent startup experience—what went wrong, what I could have done better, and how I should approach future opportunities. The Background There were three founders in this venture: • Founder A (CEO, 50%) – The product/growth guy who identified the problem space. • Founder B (Me, CTO, 37.5%) – A software engineer with a software dev shop and multiple clients. I wanted to diversify into building my own products but am not inherently a “product person.” • Founder C (COO, 12.5%) – Brought into the mix by Founder A, with the goal of leveraging his network for traction once the product was built. The idea was to create Product X, a solution targeting the SMB space while competitors were moving upmarket. It wasn’t revolutionary—more of a strategic market play. The Initial Plan & My Role • Founder A would define and prioritize product specs, guiding what needed to be built. • I (Founder B) didn’t have time to code myself, so I allocated engineers from my dev shop (which I personally paid for). My stake was adjusted from 32.5% to 37.5% to reflect this contribution. • Founder C was more of an observer early on, planning to help with traction once we had a product ready. We agreed on a 1-year cliff and a 4-year vesting schedule for equity. Where Things Started to Go Wrong • Lack of a Clear Product Roadmap – Founder A was very focused on getting something built fast, but we never signed off on a structured roadmap or milestones. I underestimated the complexity of what was actually needed for customer conversations. • Engineering Expectations vs. Reality – The team (one part-time lead + two full-time juniors from my dev shop) faced early feedback that development was too slow. In response, I ramped up the lead to full-time and added a part-time PM. But Founder A continued pushing for speed, despite real hurdles (OAuth integrations, etc.). • Shifting MVP Goalposts – Midway, Founder A concluded that an MVP wouldn’t cut it—we needed a more complete product to be competitive. This meant more engineering, more delays, and more of my own money spent on development. The Breaking Point Near the 1-year vesting mark, we had an opportunity: a paying client willing to fund an app. I didn’t have devs on the bench, so I asked Founder A to hold off our project briefly while I hired more engineers to avoid stalling either effort. This was the final straw. Founder A (with Founder C somewhat aligned) decided the arrangement wasn’t working—citing past disagreements and the “slowness” issue. The decision was made to end the partnership. Now, Founder A, as majority holder, is requesting a full handover of the code, Founder C is indifferent, and all engineering costs I covered are essentially lost. Key Takeaways (So Far) Crystal-Clear Agreements Upfront – A formalized product roadmap and timeline should’ve been locked in from day one. Business Needs > Engineering Standards – I wanted to build something solid and scalable, but in an early-stage startup, speed to market is king. This was before AI tools became mainstream, so our approach wasn’t as optimized. Don’t Overextend Without Protection – I personally financed all engineering, but without clear safeguards, that investment became a sunk cost. Expenses Must Be Distributed – I was solely covering engineering salaries, which created an imbalance in financial risk. Future partnerships should ensure costs are shared proportionally, rather than one person shouldering the burden. Where I Need Advice Looking back, I want to improve as an engineer, CEO, and co-founder. • What should I have done differently in structuring this partnership? • How do you balance engineering quality with the startup need for speed? • As a dev shop owner, how can I better navigate equity deals where I’m also bringing in engineering resources? I really appreciate everyone who went through this long post and provide any insights from founders, engineers, or anyone who has been in a similar situation. Thanks for reading! ===================================================================== For readers who might be thinking what set this type of expectation? Because I had a dev shop and I thought my co-founders will be understanding of my business circumstance and I was a bit trigger to build a product with a C-exec team, I gave the impression of "unlimited" engineering which I later realized down the line that it was not feasible for me. Something I learned that I have to be more careful with and set expectations accordingly from the very beginning. And from the feedback of the commenters here, I am much more aware what I should offer and how to set expectations, esp. in the early stages of execution. So thank you all! 🙏🏾 EDIT: I would like to thank everyone who contributed to this thread. You not only helped me but future founders who are considering to get into the startup scene!

I studied how 7 Founders found their first 100 customers for their businesses. Summarizing it here!
reddit
LLM Vibe Score0
Human Vibe Score1
adriannelestrangeThis week

I studied how 7 Founders found their first 100 customers for their businesses. Summarizing it here!

I am learning marketing, and so I combed through the internet to find specific advice that helped founders reach 100 users and not random Google answers. Here’s what I found: Llama Life by Marie Marie founder of Llama Life, a productivity app ($51.4K+ revenue) got her first 100 users using Snowballing effect. She shared great advice that I want to add here verbatim, “Need to think about what you have that you can leverage based on your current situation. eg..When you have no customers, think about where you can post to get the 1st customer eg Product Hunt. If you do well on PH, say you get #3 product of the day, then you post somewhere else saying ‘I got #3 product of the day’.. to get your next few customers. Maybe that post is on reddit with some learnings that you found. If the reddit post does well, then you might post it on Twitter, saying reddit did well and what learnings you got from that etc. or even if it doesn’t do well you can still post about it.” Another tip she shared is to build related products that get more viral than the product itself. These are small stand-alone sites that would appeal to the same target audience, but by nature, are more shareable. On these sites, you can mention your startup like: ‘brought to you by Llama Life’ and then provide a link to the main website if someone is interested. If one of those gets viral or ranks on Google, you’ll have a passive traffic source. Scraping bee by Pierre Pierre, founder of Scraping Bee, a web scraping tool has now reached $1.5M ARR. Pierre and his cofounder Kevin started with 10 Free Beta Users in 2019, and after 6 months asked them to take a paid subscription if they wanted to continue using the product. That’s how they got their first user within 50 minutes of that email. Then they listed it on dozens of startup directories but their core strategy was writing the best possible content for their target audience — Developers. 3 very successful pieces of content that worked were : A small tutorial on how to scrape single-page application An extensive general guide about web scraping without getting blocked A complete introduction to web scraping with Python They didn’t do content marketing for the sake of content marketing but deep-dived into the value they were providing their customer. One of these got 70K visits, and all this together got them to over 100 users. WePay by Bill Clerico Bill Clerico left his cushy corporate job to build WePay which was then acquired for $400M got his first users by using his app. He got his first users by using his app! The app was for group payments. So he hosted a Poker tournament at his house and collected payments only with his app. Then they hosted a barbecue for fraternity treasurers at San Jose State & helped them do their annual dues collection. Good old word-of-mouth marketing, that however, started with an event where they used what they made! RealWorld by Genevieve Genevieve — Founder and CEO of Realworld stands by the old-school advice of value giving. RealWorld is an app that helps GenZ navigate adulthood. So, before launching their direct-to-consumer platform, they had an educational course that they sold to college career centers and students. They already had a pipeline of adults who turned to Realworld for their adulting challenges. From there, she gained her first 100 followers. Saner dot ai by Austin Austin got 100 users from Reddit for his startup Saner.ai. Reddit hates advertising, and so his tips to market your startup on Reddit is to Write value-driven posts on your niche. Instead of writing posts, find posts where people are looking for solutions DM people facing problems that your SaaS solves. But instead of selling, ask about their problem to see if your product is a good fit Heartfelt posts about why you built it, aren’t gonna cut it To find posts and people, search Reddit with relevant keywords and join all the subreddits A Stock Portfolio Newsletter A financial investor got his first 100 paid newsletter subscribers for his stock portfolio newsletter. His tips : Don’t reinvent the wheel. Work what’s already working. He saw a company making $500M+ from stock picking newsletter, so decided to try that. Find the gaps in “already working” and leverage them. That newsletter did not have portfolios of advisors writing them. That was his USP. He added his own portfolio to his newsletter. Now to 100 users, he partnered with a guy running an investing website and getting good traffic. That guy got a cut of his revenue, in exchange. That one simple step got him to 100 users. Hypefury by Yannick and Samy Yannick and Samy from Hypefury, Twitter and Social Media Automation tool got their first beta testers and users from a paid community. They launched Hypefury there and asked if someone wanted to try it. A couple of people tried it and gave feedback. Samy conducted user interviews and product demos for them, And shared the reviews on Twitter. That alone, along with word-of-mouth marketing on Twitter got them their first 100 users. To conclude: Don’t reinvent the wheel, try what’s working. Find the gaps in what’s working, and leverage that. Instead of thinking about millions of customers, think about the first 10. Then first 100. Leverage what you have. Get the first 10 customers, then talk about this to get the next 100. Use your app. Find ways, events, and opportunities to use your app in front of people. And get them to use it. Write content not only for SEO but also to help people. It won’t work tomorrow, but it will work for years after it picks up. Leverage other sources of traffic by partnering up! Do things that don’t scale. I’m also doing SaaS marketing deep dives over 30 pieces of content. I'm posting here for the first time, so I'm not sure if it will stay or not, sorry if it doesn't. I've helped a SaaS grow from $19K to $100K MRR as a marketer in last 2 years, and now I wanna dive deep. Cheers! (1/30)

Feeling stuck—built a startup, got rejected from YC & IVI, met smarter people, and now I don’t know what to do. ( i will not promote )
I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
vishwa1238This week

Feeling stuck—built a startup, got rejected from YC & IVI, met smarter people, and now I don’t know what to do. ( i will not promote ) I will not promote

I will not promote I don’t even know where to start, but I just feel completely stuck right now. I’m 20 years oldI don’t even know where to start, but I just feel completely stuck right now. I’m 20 years old, have been grinding non-stop for months, and it feels like I have nothing to show for it. I built an AI agent that automates workflows for businesses. I can build tech, but I can’t sell. That’s been my biggest realization recently—I thought building would be enough, but it’s not. I need customers, I need a co-founder, I need to figure out the business side… and I have no idea how. I applied to YC, IVI at ISB, and EF, met a lot of insanely smart people—some were impressed with me and my work, but they were wiser, more experienced, and honestly, just better at all of this than I am. It made me realize how much I don’t know. I got rejected from YC & IVI. 💔 YC didn’t even give much feedback—just a standard rejection. 💔 IVI told me: “You're too young, you need more experience, and you should work with a team before trying to start something.” That hit me hard. I had already been struggling to find a co-founder, and this just made me wonder if I even belong in this space yet. The Frustrating Part? I KNOW my tool Has a Unique Edge. I’m not just another AI automation tool—I know my tool has a strong USP that competitors lack. It has the potential to be an AI employee for businesses, not just another workflow tool. But I still haven’t built the “perfect product” I originally envisioned. And that’s what’s eating at me. I see what it COULD be, but I haven’t made it happen yet. At the same time, the competition in the AI agent space is exploding. YC-backed companies are working on AI agent startups. OpenAI is making huge progress with Operator. Competitors are moving fast, while I feel stuck. I’ve delayed development because I’m unsure whether to double down, pivot, or just move on entirely. Where I’m Stuck Right Now 🔹 Do I keep pushing and try to crack sales somehow? 🔹 Do I join a startup as a founding engineer to get experience, make connections, and learn sales before trying again? 🔹 Do I move to Bangalore, meet founders, and figure out what’s next? 🔹 Do I pivot to something nicher instead of competing in the AI agent race? If so, how do I even find a niche worth pursuing? 🔹 Do I even belong in startups? Or am I just forcing something that’s not working? I feel stuck in a weird middle zone where I’m not a beginner, but I’m also not successful. I’ve done enough to see what’s possible, but not enough to make it real. Every rejection makes me question if I’m even on the right path. I don’t know if I’m posting this for advice or just to get it out of my system. Maybe both. Has anyone else felt like this before? If you’ve been in this situation—how did you figure out whether to keep going or move on? TL;DR: I’m 20, built an AI agent for automating workflows, got rejected from YC & IVI, met insanely smart and experienced people, realized I can build tech but can’t sell, struggling to find a co-founder, AI agent competition is growing, delaying development, confused about the future—don’t know whether to double down, pivot, or move on. The frustrating part? I\ know I have a unique edge that others lack, but I still haven’t built the perfect product I originally envisioned.* edit: removed the tool's name

How to get funding for startup ? I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
wlynncorkThis week

How to get funding for startup ? I will not promote

I will not promote. Software startup based out of Minnesota us. I've built and launched a product that is gaining traction, solving a problem that has frustrated software developers and product teams for years. The problem: Software development is slow, expensive, and full of inefficiencies. Developers spend hours on repetitive coding tasks, project managers struggle with bottlenecks, and businesses waste time translating product requirements into actual code. The solution: My product automates a large portion of software development. It acts as an AI-powered assistant for developers, taking high-level requirements and turning them into functional code while integrating with existing codebases. It can read, understand, and modify software projects in a structured way—cutting development time drastically. The potential: Businesses are always looking for ways to cut costs and speed up development. With the rise of AI, companies are increasingly adopting automation, and this tool fits perfectly into that wave. Imagine a world where software teams are 10x more efficient because AI handles the grunt work, and developers focus on the bigger picture. It’s not about replacing developers—it’s about supercharging them. The current status: The product is live and in use. The user base is growing, and I’ve proven demand. Now, I need to figure out the best funding model to scale—whether that’s bootstrapping, VC, grants, or some hybrid approach. If you have experience in startup funding or have scaled a tech product, I'd love to hear your insights. DM me if you're open to discussing strategies!

I spent 6 months on building a web product, and got zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a web product, and got zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I have stuff to post on Reddit very rarely, but I share how my project is going on, random stuff, and memes on X. Just in case few might want to keep in touch 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age
reddit
LLM Vibe Score0
Human Vibe Score1
ExpenectThis week

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age Introduction In an era characterized by rapid technological advancements, the field of finance is undergoing a transformative journey. The emergence of financial technology, or fintech, is reshaping the way businesses manage their finances, and Chartered Accountants (CAs) are at the forefront of this evolution. In this blog post, we'll explore how CAs are embracing fintech and leveraging its potential to enhance financial management, analysis, and advisory services. Fintech's Impact on Financial Services Fintech encompasses a wide range of technologies that leverage data analytics, artificial intelligence, blockchain, and automation to improve financial services. For CAs, this means new tools to streamline processes, enhance decision-making, and offer innovative solutions to clients. Automation of Routine Tasks CAs are increasingly using automation tools to handle repetitive tasks such as data entry, reconciliations, and transaction processing. This not only reduces the risk of human error but also frees up CAs to focus on higher-value tasks like strategic planning and analysis. Advanced Data Analytics Data analytics tools enable CAs to extract meaningful insights from large volumes of financial data. These insights can help businesses identify trends, anticipate risks, and make informed decisions to drive growth. Real-Time Financial Reporting Fintech enables CAs to provide clients with real-time financial reporting, giving businesses immediate access to critical information. This enhances transparency and empowers business owners to respond quickly to changing market conditions. Enhancing Audit Efficiency Fintech tools are revolutionizing the audit process. CAs can use AI-powered algorithms to analyze vast amounts of data, detect anomalies, and identify potential instances of fraud more efficiently. Personalized Financial Planning CAs can leverage fintech to offer personalized financial planning services. With access to detailed financial data, CAs can create tailored strategies that align with a client's unique goals and circumstances. Strengthening Cybersecurity As businesses become more reliant on digital tools, cybersecurity becomes paramount. CAs are playing a critical role in advising clients on cybersecurity measures to protect sensitive financial information. Virtual CFO Services Fintech enables CAs to offer virtual CFO services to startups and small businesses. Through digital platforms, CAs can provide expert financial advice and guidance remotely, making their expertise accessible to a wider range of clients. Embracing Blockchain Technology Blockchain's potential for secure and transparent record-keeping is of interest to CAs. They can explore applications in supply chain finance, smart contracts, and even audit trail verification. Continuous Learning in Fintech CAs recognize the importance of staying updated with fintech trends. Many are investing in continuous learning to master the use of new tools and technologies that can optimize their services. Conclusion The integration of fintech into the realm of finance is reshaping the landscape in profound ways. CAs are embracing these technologies to elevate their roles from traditional number-crunchers to strategic advisors, equipped with tools that enhance efficiency, accuracy, and insight. As fintech continues to evolve, CAs will remain pivotal in guiding businesses through the ever-changing financial landscape, leveraging technology to drive growth, innovation, and success. Find the top verified CA in your City Feel free to let me know if you'd like more blogs on different topics or if you have specific requirements for the content.

The Future of AI in eCommerce Marketing: What to Expect 🚀
reddit
LLM Vibe Score0
Human Vibe Score0
McFlyAdsThis week

The Future of AI in eCommerce Marketing: What to Expect 🚀

Hey Reddit community! As we dive deeper into 2025, the integration of AI in eCommerce marketing is becoming more sophisticated and impactful. Here’s a look at where AI is headed and how it's revolutionizing the industry: Personalized Shopping Experiences: AI is enhancing personalization by analyzing consumer behavior and preferences, allowing retailers to offer tailored recommendations and promotions. This not only boosts customer satisfaction but also increases conversion rates. Chatbots and Virtual Assistants: AI-powered chatbots are becoming more intuitive and capable of handling complex queries, providing 24/7 customer support, and improving overall user experience. They’re a game-changer for eCommerce businesses looking to enhance customer engagement. Predictive Analytics: With AI, businesses can leverage predictive analytics to forecast trends, optimize inventory, and refine marketing strategies. This helps in making data-driven decisions that align with consumer demands and market dynamics. Automated Content Creation: AI tools are being used to generate product descriptions, social media posts, and even ad copy. This automation saves time and ensures consistency across marketing channels. Visual and Voice Search: AI is powering visual and voice search capabilities, making it easier for consumers to find products using images or voice commands. This technology is set to transform how users interact with eCommerce platforms. Fraud Detection: AI algorithms are improving fraud detection by analyzing transaction patterns and identifying anomalies. This is crucial for maintaining trust and security in online shopping. As AI continues to evolve, it will undoubtedly reshape the eCommerce landscape, offering new opportunities for innovation and growth. What are your thoughts on the future of AI in eCommerce marketing? Let's discuss!

Ai C-Level team
reddit
LLM Vibe Score0
Human Vibe Score1
thestoicdesignerThis week

Ai C-Level team

I've been exploring ways to run a company where I'm essentially the only internal team member, relying entirely on a suite of specialized AIs for executive roles, supported occasionally by external consultants for niche expertise. My goal is to stay lean, agile, and highly creative, especially in a fashion/tech brand context. Essentially, I'm building an AI-driven C-Level team, or what I like to call a "C-Level AI Wallet." Here's what I'm thinking for the key executive roles I'd need to cover with AI: CEO AI – Responsible for overall strategy, decision-making, trend analysis, and guiding the company's vision. I'd probably lean on something advanced like Gemini, GPT-4, or similar models, fine-tuned with market-specific data. COO AI (Operations): I'd need tools that streamline and automate logistics, supply chain management, and day-to-day operations (think something along the lines of Zapier AI integrations or Make). CMO AI (Marketing & Content): For branding, content creation, digital marketing, and consumer insights, I'd use Jasper or Copy.ai, combined with predictive analytics tools like Google Vertex AI to understand trends better. Additionally, for generating engaging visual and multimedia content, tools like Midjourney, DALL·E, Adobe Firefly, and Runway ML would be perfect. CFO AI (Financial Management): For financial management, cash flow control, and investment decisions, I'd probably leverage AI tools like Bloomberg GPT, combined with AI-powered forecasting platforms. CHRO AI (Human Resources & Culture): Although the internal team is minimal (just myself!), I'd still rely on AI for tasks like project management, freelancer hiring, and performance tracking—tools like HireVue AI, Motion, or even Notion's AI could be beneficial here. CSO AI (Sustainability & Compliance): Since sustainability and ethical sourcing are critical, I'd integrate ESG-focused AI tools to ensure transparency and responsible sourcing. My idea is that, with the right AI tools seamlessly integrated, I can manage the strategic vision and creative direction personally, leveraging external consultants only when necessary. This setup would ideally allow me to operate as a one-person internal team supported by a robust "wallet" of AI executives. Has anyone tried a similar approach? What AI tools would you recommend for a truly lean, innovative brand structure? I'm very curious about your experiences or suggestions—let me know your thoughts!

 Struggling with Cold Start for Our AI PowerPoint Tool - Seeking Platform and Strategy Suggestions!
reddit
LLM Vibe Score0
Human Vibe Score1
yamaggieThis week

Struggling with Cold Start for Our AI PowerPoint Tool - Seeking Platform and Strategy Suggestions!

Hello everyone, I'm one of the co-founders of a new AI-generated PowerPoint company, and I handle the marketing side of things. Our product is currently in the cold start phase, and we’re facing some challenges in gaining traction. We've already tried some influencer marketing, but the results have been underwhelming. We're looking for advice on the best platforms and strategies to effectively launch our product and reach our target audience. Here’s a bit more about our product: AI-Powered: Our tool leverages AI to help users quickly create professional PowerPoint presentations by simply entering their desired topic. User-Friendly: The process is streamlined to save users time and effort, making it ideal for professionals, educators, and students. Given our current situation, we would greatly appreciate any suggestions on: Platforms: Which platforms have you found most effective for cold starts, especially for tech or AI products? Strategies: What marketing strategies or tactics have worked for you in the early stages? Any tips on refining our influencer marketing approach or alternative methods to consider? Partnerships: Are there any specific types of partnerships or collaborations that you’ve found beneficial for similar products? Thank you in advance for your insights and advice. We're eager to learn from this community and hopefully turn things around for our launch. Best, Maggie

I spent 6 months on building a web product, and got zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a web product, and got zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I have stuff to post on Reddit very rarely, but I share how my project is going on, random stuff, and memes on X. Just in case few might want to keep in touch 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

I spent 6 months on building a web product, and got zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a web product, and got zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I have stuff to post on Reddit very rarely, but I share how my project is going on, random stuff, and memes on X. Just in case few might want to keep in touch 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

80+ Social Media Updates Related to Business Marketing That Occurred in last 5 months
reddit
LLM Vibe Score0
Human Vibe Score0.333
lazymentorsThis week

80+ Social Media Updates Related to Business Marketing That Occurred in last 5 months

Tiktok expanded its caption limits from 100 to 500 Characters. Reddit Updates Search tools, Now you can search User Comments. “Comment search is here”. Pinterest Announces New Partnership with WooCommerce to Expand Product Listings. Google’s launched ‘multisearch’ feature that lets you search using text and image at the same time. Etsy sellers went on strike after platform increases transaction fees. Reddit launched $1 million fund to support various projects going on platform. Instagram is updating its ranking algorithm to put more focus on Original Content LinkedIn Added New tools In creator mode: improved content analytics and Updates profile video Options. Tiktok launched its own gif library “Effect House”. Instagram Updates Reels editing tools adding reordering clips feature. Google Search got a new label to direct people to original news sources YouTube launches new Profile Rings for Stories and Live. Snapchat launched YouTube Link stickers to make video sharing easier! Messenger adds new shortcuts, including a slack like @everyone feature. Pinterest Expands it’s Creator funds program to help more Underrepresented creators. Reddit brings back r/place after 5 years. Google Adds New Seller Performance Badges, New Pricing Insights for eCommerce Brands. Meta and Google agrees to New Data Transfer agreement to keep Instagram and Facebook running in EU. Twitter tests New Interactive Ad types to boost its promotional Appeal. Instagram removed In-stream Ads from its Advertising Options. Tiktok launched new program “CAP” to help creative agencies reach its audience. Twitch shuts down its desktop app. Meta launched the ability to add “share to Reels” feature to third Party Apps. TikTok Adds New ‘Background Player’ Option for Live-Streams. Twitter rolls out ALT badge and improved image description. Fast, A Checkout Startup with $15 billion valuation shuts down after spending all the funds raised in 2021. Wordpress announced new pricing with more traffic and storage limits after receiving backlash from the community. Sales force upgrades marketing field services and sales tools with AI. Dropbox shop launches in open beta to allow creators to sell digital content. Tiktok is the most downloaded app in Quarter 1 of 2022. WhatsApp announced launch of ‘Communities’ - more structured group chats with admin controls. Tiktok expands testing a private dislike button for comments. Twitter acquired “Openback” A notification app to improve timeline and relevance of push notifications YouTube and Tiktok added New options for Automated Captions, Improving Accessibility. A new social media App “Be Real” is trending across the internet grabbing Gen-Zs attention to try the app. WhatsApp got permission to expand payment services to its Indian user base of 100 Million. YouTube Shorts now allows creators to splice in long-form videos. You can use long form video audios and clips for YT shorts. New Snapchat feature ‘Dynamic Stories’ uses a publisher’s RSS feed to automatically create Stories posts. Zoom launches AI-powered features aimed at sales teams. Tiktok started testing who viewed your profile feature. Ogilvy Announced they will no longer work with who edit their bodies and faces for ads. If you don’t know “Oglivy” is the most successful advertising agency of the decade. YouTube Launches New ‘Search Insights’ for all creators. Snapchat Added 13 million new users in Q1 2022 more than both Twitter and Facebook. Google is Introduced new options to reject tracking cookies in Europe after receiving fines from violating EU data laws. Sony & Microsoft are planning to integrate Ads into their gaming platforms Xbox and PlayStation. YouTube Adds new Shorts Shelf to Trending Tab to show Top Shorts in an alternative section. Instagram started testing a reels template feature which enables creators to copy formats from other reels. Google Tests “What People Are Saying” Search Results. Twitter Launches New Test of Promotions for Third Party Tools Within the App. Instagram is changing how hashtags work by experimenting removing Recents tab from hashtags section. Google Adds New Publisher Verification Badges to Extension Listings in the Google Web Store Amazon AWS launches $30M accelerator program aimed at minority founders. Meta launched more fundraising options for Instagram Reels in 30 countries. Brave Search and DuckDuckGo will no longer support Google AMP due to privacy issues. Instagram is working on a pinned post feature and will officially launch in next few months. Meta: You can now add Music to your Facebook comments Twitter tests new closed caption button to switch on captions in Video Clip Elon Musk Bought Twitter $44 Billion and Company is set to go private. Google now lets you request the removal of personal contact information from search results YouTube reveals that Ads between YT Shorts are being tested with selective brands. LinkedInis rolling out a new website link feature. Google Reduces Visibility Of Business Edits With Color Changes To Profile Updates. Instagram expands testing of 90 second Reels. Microsoft Advertising now offers incentive features like cash-back and adding stock images from your website. Facebook & Pinterest are growing again despite all the hype around slow growth of both platform in last quarter. Google Added 9 new Ad policies to prevent misleading ads taking place. Tiktok Introduces Third-party cookies to its Pixel. (like Facebook Pixel) Twitter reportedly overcounted number of daily active users for last 3 years. Google launched Media CDN to compete on content delivery. YouTube expands Thank You Monetisation tool to all eligible creators. Twitch is looking to expand their cut from streamers earnings from 30 to 50% and also thinks of boosting Ads. Snapchat launches a $230 flying drone camera and new e-commerce integrations in Snap Summit 2022. YouTube Expands its ‘Pre-Publish Checks’ Tool to the Mobile App Google Search Console’s URL parameter tool is officially removed for a time period. Twitter creators can now get paid through Cryptocurrency on Twitter with Stripe. Jellysmack- One of the Influencer marketing agency acquires YouTube analytics tool Google & Microsoft Ads brought more revenue in last quarter- 22% Gains! WhatsApp is working on a paid subscription for multi-phone and tablet chatting. Instagram users now spend 20% of their time in the reels section. Google tests new Color for clicked search results by you. Now Clicked results are in Purple. Twitter: Elon plans to remove employees and focus more on influencers for twitter’s growth + new monetisation ideas were shared. YouTube revenue falls as more users spend time on shorts tab than consuming long form content. Drop 👋 to receive June Updates!

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age
reddit
LLM Vibe Score0
Human Vibe Score1
ExpenectThis week

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age Introduction In an era characterized by rapid technological advancements, the field of finance is undergoing a transformative journey. The emergence of financial technology, or fintech, is reshaping the way businesses manage their finances, and Chartered Accountants (CAs) are at the forefront of this evolution. In this blog post, we'll explore how CAs are embracing fintech and leveraging its potential to enhance financial management, analysis, and advisory services. Fintech's Impact on Financial Services Fintech encompasses a wide range of technologies that leverage data analytics, artificial intelligence, blockchain, and automation to improve financial services. For CAs, this means new tools to streamline processes, enhance decision-making, and offer innovative solutions to clients. Automation of Routine Tasks CAs are increasingly using automation tools to handle repetitive tasks such as data entry, reconciliations, and transaction processing. This not only reduces the risk of human error but also frees up CAs to focus on higher-value tasks like strategic planning and analysis. Advanced Data Analytics Data analytics tools enable CAs to extract meaningful insights from large volumes of financial data. These insights can help businesses identify trends, anticipate risks, and make informed decisions to drive growth. Real-Time Financial Reporting Fintech enables CAs to provide clients with real-time financial reporting, giving businesses immediate access to critical information. This enhances transparency and empowers business owners to respond quickly to changing market conditions. Enhancing Audit Efficiency Fintech tools are revolutionizing the audit process. CAs can use AI-powered algorithms to analyze vast amounts of data, detect anomalies, and identify potential instances of fraud more efficiently. Personalized Financial Planning CAs can leverage fintech to offer personalized financial planning services. With access to detailed financial data, CAs can create tailored strategies that align with a client's unique goals and circumstances. Strengthening Cybersecurity As businesses become more reliant on digital tools, cybersecurity becomes paramount. CAs are playing a critical role in advising clients on cybersecurity measures to protect sensitive financial information. Virtual CFO Services Fintech enables CAs to offer virtual CFO services to startups and small businesses. Through digital platforms, CAs can provide expert financial advice and guidance remotely, making their expertise accessible to a wider range of clients. Embracing Blockchain Technology Blockchain's potential for secure and transparent record-keeping is of interest to CAs. They can explore applications in supply chain finance, smart contracts, and even audit trail verification. Continuous Learning in Fintech CAs recognize the importance of staying updated with fintech trends. Many are investing in continuous learning to master the use of new tools and technologies that can optimize their services. Conclusion The integration of fintech into the realm of finance is reshaping the landscape in profound ways. CAs are embracing these technologies to elevate their roles from traditional number-crunchers to strategic advisors, equipped with tools that enhance efficiency, accuracy, and insight. As fintech continues to evolve, CAs will remain pivotal in guiding businesses through the ever-changing financial landscape, leveraging technology to drive growth, innovation, and success. Find the top verified CA in your City Feel free to let me know if you'd like more blogs on different topics or if you have specific requirements for the content.

Seeking Feedback on My Business Idea – SaaS + Lead Generation for Small Businesses
reddit
LLM Vibe Score0
Human Vibe Score1
sarveshpandey89This week

Seeking Feedback on My Business Idea – SaaS + Lead Generation for Small Businesses

Edit: TL;DR I’m Sarvesh, a digital marketer with 10 years of experience in paid ads. After losing my job last year, I started freelancing and discovered how much small businesses struggle with getting reviews (Google, Yelp, TrustPilot, etc.). My Business Idea – SaaS + Paid Ads Free Plan: Businesses can track & reply to reviews across 40+ platforms in one dashboard. Paid Plan ($99/month): Automates review collection, AI-powered responses, social media posting, and spam detection. Custom Plan: Paid ads to generate leads, offered only to businesses on my paid plan for 3+ months. Goal: SaaS platform attracts users → Some upgrade to paid plan → Best clients get lead-generation help → More leads → More reviews → More organic customers → A profitable business cycle. Need Feedback: Does this idea have potential? How can I get my first beta users? Any features I should add/remove? Would love your thoughts—thanks for reading! 😊 TL: Hi everyone, I’m Sarvesh, and I’m in the process of starting my own business. Since my target audience is small businesses, I’d love to get some input, advice, or critiques from this community. A Little About Me I’ve spent the last 10 years working in paid advertising, helping medium and large businesses generate leads through Facebook and Google Ads. I also have experience running e-commerce campaigns. You can check out my background on LinkedIn: LinkedIn Profile Last year, my second daughter was born, and around the same time, my company shut down all its offices (India & UK), leaving me without a job. I decided to take a break and spend time with my wife and newborn, something I regretted not doing with my first child. By November, I started job hunting again, but in the meantime, I got some freelance work through Reddit, helping small businesses with ads for the first time. For context, in my previous jobs, I managed ad campaigns with daily budgets of £4K–£8K. Working with small businesses was a new challenge, but to my surprise, I was able to generate solid leads for beauty salons, hair salons, and nail salons, helping them grow. What stood out to me was how much impact my work had—unlike my corporate job, where I was just another person in the system, here I felt truly valued. That feeling led me to explore starting my own business. The Problem I Noticed While working with small businesses, I realized that online reviews (Google, Yelp, Trustpilot, etc.) are critical for them, yet many struggle to get them. Customers often don’t leave reviews, and employees are either too shy or don’t prioritize asking for them. This gave me an idea—to build a system that helps businesses get more genuine Google reviews from customers. I developed the system but struggled to find businesses willing to test it, even for free. My target audience is U.S. small businesses, but since I’m based in India, cold emails and Reddit outreach didn’t get much traction. My Business Idea – SaaS + Custom Plans I’m now thinking of pivoting my business model into a SaaS platform with optional paid upgrades. Here’s how it would work: Free Plan (Review Tracking & Management) Businesses can track their reviews across 40+ platforms (Google, Yelp, Facebook, Trustpilot, TripAdvisor, etc.) in one dashboard. They can reply to reviews manually from a single place instead of switching between platforms. This will be completely free forever. Paid Plan ($99/month, Plus SMS/Email Costs) For businesses that struggle to get reviews, they can upgrade to a paid plan that includes: Automated Review Requests – Automatically send review requests via SMS & email. Website Widget – Showcase 4- and 5-star reviews dynamically. Social Media Automation – Automatically post positive reviews on Facebook/Instagram. AI-Powered Responses – AI can reply to reviews automatically. Spam Detection – The system will notify businesses of suspicious reviews (but won’t take direct action). Custom Plan (Lead Generation via Paid Ads) I will personally manage paid ad campaigns to generate leads. Pricing depends on the niche, budget, and contract duration. Money-Back Guarantee – If I don’t deliver results, I refund the month’s fee. Small businesses can’t afford wasted ad spend, and I want to ensure I provide real value. Limited spots per month to maintain quality and avoid burnout. How Everything Ties Together The SaaS platform serves as a lead generation tool for my custom plans: Businesses use the free plan to track their reviews. Some upgrade to the paid plan to automate and improve reviews. A select few, after 3 months on the paid plan, can join my custom plan for paid ads to generate more leads. More leads → More reviews → Better Google Maps ranking → More organic customers → A more profitable business. Would Love Your Feedback! What do you think about this approach? Do you see potential for this business to take off? Any features I should add or remove? Any suggestions on how I can get my first beta users to test the SaaS platform? What about pricing? Do you think $99 is good pricing? I know this is a long post, but I really appreciate anyone taking the time to read and share their thoughts. Thanks in advance!

Ai C-Level team
reddit
LLM Vibe Score0
Human Vibe Score1
thestoicdesignerThis week

Ai C-Level team

I've been exploring ways to run a company where I'm essentially the only internal team member, relying entirely on a suite of specialized AIs for executive roles, supported occasionally by external consultants for niche expertise. My goal is to stay lean, agile, and highly creative, especially in a fashion/tech brand context. Essentially, I'm building an AI-driven C-Level team, or what I like to call a "C-Level AI Wallet." Here's what I'm thinking for the key executive roles I'd need to cover with AI: CEO AI – Responsible for overall strategy, decision-making, trend analysis, and guiding the company's vision. I'd probably lean on something advanced like Gemini, GPT-4, or similar models, fine-tuned with market-specific data. COO AI (Operations): I'd need tools that streamline and automate logistics, supply chain management, and day-to-day operations (think something along the lines of Zapier AI integrations or Make). CMO AI (Marketing & Content): For branding, content creation, digital marketing, and consumer insights, I'd use Jasper or Copy.ai, combined with predictive analytics tools like Google Vertex AI to understand trends better. Additionally, for generating engaging visual and multimedia content, tools like Midjourney, DALL·E, Adobe Firefly, and Runway ML would be perfect. CFO AI (Financial Management): For financial management, cash flow control, and investment decisions, I'd probably leverage AI tools like Bloomberg GPT, combined with AI-powered forecasting platforms. CHRO AI (Human Resources & Culture): Although the internal team is minimal (just myself!), I'd still rely on AI for tasks like project management, freelancer hiring, and performance tracking—tools like HireVue AI, Motion, or even Notion's AI could be beneficial here. CSO AI (Sustainability & Compliance): Since sustainability and ethical sourcing are critical, I'd integrate ESG-focused AI tools to ensure transparency and responsible sourcing. My idea is that, with the right AI tools seamlessly integrated, I can manage the strategic vision and creative direction personally, leveraging external consultants only when necessary. This setup would ideally allow me to operate as a one-person internal team supported by a robust "wallet" of AI executives. Has anyone tried a similar approach? What AI tools would you recommend for a truly lean, innovative brand structure? I'm very curious about your experiences or suggestions—let me know your thoughts!

What I learn from my $200 MRR App I built 4 months ago?
reddit
LLM Vibe Score0
Human Vibe Score0.857
ricky0603This week

What I learn from my $200 MRR App I built 4 months ago?

4 month ago, I am just a 10-years experienced product manager without any software development experience. I have an $3K/month job, but I am so tired, I don’t like my life, don’t like my boss, don’t like my daily work, that make me feeling I already died however I am still living. I yearn for freedom and want to live each day the way I want to. So I quit my job, and become a Indie developer to build my own business, my own app, even my own life. I am so grateful for this time and experience, now my app reach $200 MRR, still very little compared to my previous salary, but I never regret. I have learned lots of things from this time and experience, more than I had in last 10 years. Here is the time-line of my App: \- Sep 2023: Launch first version to iOS App store \- Oct 2023: Release in-app-purchase features and have first subscriber, the revenue in October is $154 \- Nov 2023: Change from subscription to pay per use, and I did lots of marketing jobs in November, however, the revenue reduced to only $40. \- Dec 2023: Change back to subscription, and stop some invalid marketing jobs, only keep the ones that actually work. I almost did nothing in December, and the revenue come to $243. During this process, I have learned lots of things, there are some of them that I think could help you as well. Web or App My App is an iOS app that only can running on Apple’s device such like iPhone/iPad or Mac with Apple silicon. Many people ask me why my product is an iOS app not a website, because they don’t have any Apple device. It's true that promoting an app is much harder than promoting a website. However I am now very glad I made an App and not a website! If I make a website, I don't think it's possible to make $100 in the first month. My App is about keyword research, to help people find some ideas from search keyword, because every keyword people searched in Google are representing a real need of them, also can be used in SEO field. However there are a lot of website tools about keyword research, some of them are famous like Ahrefs, SEMrush… I have no intention of competing with them. Actually I don’t have any chance. While in app store, there are little apps about keyword research, each of them have terrible data and user experience, that means if my app has better data and experience that could be my chance. In fact, the App store brings me 20 organic installs a day that Google would never have been able to bring me if I had a website, at least for the first few months. Furthermore, Apple nearly did everything for developer, I don’t need to care about user login, payment and so on, Apple did everything, I just need to call their API, that save lots of time, if I build a website, I need to implement login and payment by myself, that would add some extra work. Not to mention I'd need to buy servers and domains, that would cost me a lot of money. Although Apple will take 30% of the revenue, I can live with that in the early stages because the most important thing for me is to get the product to market as soon as possible. Actually thought Apple’s SMB program, the take rate is 15% now. So Web or App is not important in the early stage, time is important, if people need my product, it's easy to make a website one. More Users or More Valuable Users In November, I notice some users would like use my app, and they were meet paywall, but they never subscribe. I provided 7 day free trail, but it seem that they don’t like it. So I decide to change subscription to pay per use. Because as a user, I don’t like subscription as well, pay per use seem like more friendly. So I change from subscription to pay per use. People can afford $9.99 to subscribe monthly for unlimited use or pay $1.99 for each data they want(First purchase is $0.99 then $1.99). I was expecting more user to pay, but it was the complete opposite! Some users who would have paid a higher subscription fee are switching to a lower priced single payment. Users are encountering paywalls more often, and each time they need to make a decision about whether or not to pay, which increases the probability that they will abandon payment. This resulted in a 75% decrease in revenue in November. In fact, the mostly of my revenue comes from a handful of long-cycle subscribers, such as annual subscription. \\Few bring in most of the revenue,\\ that is the most important thing I learned. You don't need a lot of customers, you just need more valuable ones. That's why it's only right to design a mechanism to filter out high-value customers and focus on them, all the things you want do is just let more people into the filter, and from that point of view, subscription with free trial period is the best way, even if most people don't like it. The rule of 20/80 will always be there. The most important thing is always focus on the 20 percent things and people. Effort does not always guarantee rewards. Unless one engages in deep thinking, or most efforts are invalid. I have been working very hard to promote my product for a period of time. It’s about in November. I did a lot of job, such as write script to send message to my potential clients on Fiverr, post and write comments on others post on Reddit, find related questions and answer them on Quora, post and comments on Twitte, etc. During that period, I was exhausted every day, but the outcome did not meet my expectations. There is only little growth on App installation, even less revenue than before. That make me frustrated. I finally realized that If I need to put in a tremendous amount of effort just to make a little progress, there is must something wrong. So I stop 80% of promote work I have ever did, only keep app store search ad, which will bring a installation with less than $0.5 cost. Then I dive into long time and deeply thinking, I spent more time on reading books, investigate other product with great MRR, watch interviews with people who are already living the kind of life I aspire to live, for example, u/levelsio. These things have given me great inspiration, and my life has become easier. It seems that the life I anticipated when I resigned is getting closer. I also have a clearer understanding of my app. Meanwhile, MRR has been growing. This experience let me learn that effort does not always guarantee results. Many times, our efforts are just wishful thinking, they are invalid, do the right thing after deeply thinking is more important. What Next? My goal is reach $3K MRR, as same as my job payment, I will never stop to building things, and I will keep my currently lifestyle. I still don't know how to get more people to use my app, but levelsio's interviews give me some inspiration that I can verified something by manually instead of build a software. I plan to launch a trend analysis product based on the keyword data provided by my current app. I have always wanted to combine AI to build such a product, but I didn't know how to do it. Now I intend to manually complete it first and start software development once there are paying users. If you are interested to my App, you could try it.

Feedback request: Virtual Receptionist - Phone Answering Service
reddit
LLM Vibe Score0
Human Vibe Score0
AlexDataKnowlThis week

Feedback request: Virtual Receptionist - Phone Answering Service

Hi everyone! We develop an AI solution for the Enterprise segment, aimed at managing and automating interactions with Customers (e.g. self-service customer support via telephone) I do not refer in any way to company names, products, etc... to avoid any ambiguity or distraction. ➡️ My post is aimed at exploring the point of view of an SMB regarding the problem of managing interactions with customers, for example in managing telephone calls. ⭐In particular, what do you think of traditional virtual receptionist and phone answering services? The underlying issue is valuing every call, or interaction in general. ⭐A missed or poorly managed call can be a lost opportunity (a sale, a service) or even worse, a lost customer. But often outsourcing the service also means relying on people in a call center who manage your business in parallel with many others, and you cannot be an expert in everything. Therefore, the outsourced agents will be competent on 10, 15 points. This often results in little use for the caller, which does not resolve the issue for which he called. AI is making giant strides, as is speech synthesis and speech recognition. 🚀 What do you think if you were to use AI as a virtual receptionist, or in general for call management for your business? ⭐ Points in favor? What is the most critical aspect? Typical cases: Provide information Customer service Appointment made Order management ...other? The purpose of this post is to provide food for constructive reflection, combining different visions.

40% Of SMBs Still Can't Pay Their Rent, Extending High Delinquency From September Into October
reddit
LLM Vibe Score0
Human Vibe Score1
Aegidius25This week

40% Of SMBs Still Can't Pay Their Rent, Extending High Delinquency From September Into October

https://www.alignable.com/forum/q4s-off-to-a-rough-start-40-of-smbs-still-cant-pay-their-rent October 31, 2023: While the federal government reported a surge in economic growth for the U.S. last week, that news doesn't hold true for many small business owners. In fact, in October polling by Alignable, only 12% said their companies are experiencing significant growth this month. Beyond that, Alignable’s October Rent Report, released today, shows that a whopping 40% of SMBs couldn't even pay their October rent in full and on time. This marks the second consecutive month of a 40% rent delinquency rate -- extending 2023's record high from September through October. These findings are based on responses from 4,246 randomly selected small business owners surveyed from 10/1/23 to 10/30/23, as well as input from 44,000+ other respondents over the past year. As the chart below shows, October's SMB rent delinquency rate is 10 percentage points higher than it was in January, reflecting cumulative economic struggles: increased rents, high interest rates, still-stifling inflation, rising labor costs, and revenues that have declined since this time last year. Rent delinquency rates among small businesses during 2023 based on Alignable surveys So, Why's Rent Delinquency At 40% For A 2nd Month? Here’s the current list of problems contributing to two months' worth of the highest delinquency rate 2023 has seen so far: Consumer Spending Declines On Main Street: Quarterly, we ask about customer spending habits at retailers. This month, 45% of independent Mom and Pop Shops said spending has been down over the last 30 days. Some said it was due to more people spending money online with big retailers like Amazon. This figure is quite high, especially considering that back in July, only 24% reported a drop in consumer spending -- 21 percentage points less severe than it is now. Revenue Troubles: 42% are making half or less of the income they generated monthly prior to COVID. For businesses that are less than three years old, this situation is even worse: 53% of this group reports making half or less of what they generated this time last year. High Interest Rates: Over half of all SMB owners polled said the past 19 months of high interest rates have hurt their margins, reduced revenues, and put their expansion plans on hold, as they don't want to apply for loans. Increased Rent Prices: 50% say they’re being charged more for rent now than they were six months ago, with 15% saying rent has increased by 20% or more. At present, only 37% of pre-COVID businesses have recovered financially from the pandemic era, leaving 63% still striving to make up for time they lost due to COVID, inflationary pressures, and high interest rates. There's a slight silver lining here, though, as the 37% figure is three percentage points higher than it was in September. But, with that said, a recovery rate of 37% after more than three and a half years is still very low and speaks volumes about the ongoing list of troubles small business owners face looking into the rest of 2023. Tech, Manufacturing, Gyms, Beauty & Retail Struggle Examining the rent delinquency landscape in terms of sectors, there's quite a negative shift occurring among some industries in October. Let's look at the charts below to see what's really happening. Sectors most affected by rent delinquency include tech and retail Details on sectors affected by rent delinquency in October This is alarming for a few reasons: The countless technology layoffs at larger companies over the past year appear to be affecting the small companies now, too, who are often dependent on the larger ones as clients. Right now, 54% of science/technology small companies couldn't pay their October rent, up 10 percentage points from September and 16 percentage points since August. There are also some comments in the surveys of technology roles being reduced or replaced by ChatGPT and other AI, which can write software programs. Gyms have been struggling now for a while and now 50% of them can't afford the rent, up 8 percentage points from September. The biggest shift between October and September occurred among manufacturers, partially due to ongoing fluctuation in the price of gas and other inflationary issues. For quite some time, manufacturers were improving a lot in terms of their rent delinquency rates, but in October, they jumped 25 percentage points, doubling their rate, which is now 50%. This is also a record high for manufacturers in 2023. We hope this is just a blip, but we'll see in November. Also due, in part, to fluctuating gas prices and costs of vehicles, 45% of transportation companies couldn't pay October rent in full and on time. That's up 6 percentage points from last month. Sadly, 47% of salon owners couldn't cover October rent, after showing a lot of stability over the past few months. But that stability ended this month, as salons' rent delinquency rates jumped nine percentage points. Though rates have dropped three percentage points in October, a high percentage of retailers are still having trouble paying the rent. Last month, it was 47%. This month, it's better, but is still over 40%, landing at 44%. This is worrisome, especially since Q4 is a "make it or break it" time for many Main Street merchants. Looking more closely at the industries, there was some good news, in that a few others experienced lower delinquency rates in October, including restaurants, which dipped to 40% from 44% in September. Travel/lodging dropped seven percentage points to 38% (from 45% last month), as did education, which is also at 38%, down from 43%. When looking at rent delinquency from the vantage point of the states that are most affected, many surges can be seen between October and September, while a few states saw some dramatic, encouraging declines, too. Rent Troubles Increase For IL, VA, TX, MA, FL, & CO Looking at the states' charts, you can see how tumultuous the rent story has become this fall. Let's first talk about those with significant jumps in their delinquency rates. Here's the rundown: Illinois leads the list once again. After having a better month in September, its delinquency rate has soared, once more, landing at 54% for October (up from 46% last month). In fact, the 54% figure is the highest rate IL-based SMBs have seen in 2023. Virginia was in great shape last month, with a delinquency rate of just 19%. But Virginia-based small business owners have had a very rough month, at least in terms of rent. Now, 50% of them who took our poll say they couldn't cover rent (an increase of 31 percentage points). Texas is third on the list, with an 11-percentage-point lift from 38% in September to 49% in October. MA is next up at 48%, which marks the largest jump on the chart -- 32 percentage points from a low of just 16% in September. Small businesses in Florida have also experienced two challenging months in terms of rent delinquency. Right now, 45% of SMBs there couldn't afford to pay, up nine percentage points from September and 15 percentage points from August. Colorado's businesses regressed in October, hitting a new record high of 40%. That rent delinquency rate jumped 13 percentage points from September to October. While we just covered states with some very high delinquency rates, there were also several more positive swings that have occurred in October. Though encouraging, we'll have to see how long those delinquency rates continue. Here are the most remarkable: New York -- After reaching a record rate of 55% last month, New York's small business owners now report a more stable number: just 29%. That's down 26 percentage points. New Jersey -- New York's neighbor has an even more impressive story in October: only 20% of New Jersey's SMBs couldn't pay rent this month, a record low over at least the past 14 months, down 34 percentage points from a record high of 54%. Michigan -- Similarly, Michigan's small business owners boast a rate of just 20%, down from 45% in September.

My Manager Thinks ML Projects Takes 5 Minutes 🤦‍♀️
reddit
LLM Vibe Score0
Human Vibe Score1
SaraSavvy24This week

My Manager Thinks ML Projects Takes 5 Minutes 🤦‍♀️

Hey, everyone! I’ve got to vent a bit because work has been something else lately. I’m a BI analyst at a bank, and I’m pretty much the only one dealing with machine learning and AI stuff. The rest of my team handles SQL and reporting—no Python, no R, no ML knowledge AT ALL. You could say I’m the only one handling data science stuff So, after I did a Python project for retail, my boss suddenly decided I’m the go-to for all things ML. Since then, I’ve been getting all the ML projects dumped on me (yay?), but here’s the kicker: my manager, who knows nothing about ML, acts like he’s some kind of expert. He keeps making suggestions that make zero sense and setting unrealistic deadlines. I swear, it’s like he read one article and thinks he’s cracked the code. And the best part? Whenever I finish a project, he’s all “we completed this” and “we came up with these insights.” Ummm, excuse me? We? I must’ve missed all those late-night coding sessions you didn’t show up for. The higher-ups know it’s my work and give me credit, but my manager just can’t help himself. Last week, he set a ridiculous deadline of 10 days for a super complex ML project. TEN DAYS! Like, does he even know that data preprocessing alone can take weeks? I’m talking about cleaning up messy datasets, handling missing values, feature engineering, and then model tuning. And that’s before even thinking about building the model! The actual model development is like the tip of the iceberg. But I just nodded and smiled because I was too exhausted to argue. 🤷‍♀️ And then, this one time, they didn’t even invite me to a meeting where they were presenting my work! The assistant manager came to me last minute, like, “Hey, can you explain these evaluation metrics to me so I can present them to the heads?” I was like, excuse me, what? Why not just invite me to the meeting to present my own work? But nooo, they wanted to play charades on me So, I gave the most complicated explanation ever, threw in all the jargon just to mess with him. He came back 10 minutes later, all flustered, and was like, “Yeah, you should probably do the presentation.” I just smiled and said, “I know… data science isn’t for everyone.” Anyway, they called me in at the last minute, and of course, I nailed it because I know my stuff. But seriously, the nerve of not including me in the first place and expecting me to swoop in like some kind of superhero. I mean, at least give me a cape if I’m going to keep saving the day! 🤦‍♀️ Honestly, I don’t know how much longer I can keep this up. I love the work, but dealing with someone who thinks they’re an ML guru when they can barely spell Python is just draining. I have built like some sort of defense mechanism to hit them with all the jargon and watch their eyes glaze over How do you deal with a manager who takes credit for your work and sets impossible deadlines? Should I keep pushing back or just let it go and keep my head down? Any advice! TL;DR: My manager thinks ML projects are plug-and-play, takes credit for my work, and expects me to clean and process data, build models, and deliver results in 10 days. How do I deal with this without snapping? #WorkDrama

How I Started Learning Machine Learning
reddit
LLM Vibe Score0
Human Vibe Score1
TechPrimoThis week

How I Started Learning Machine Learning

Hello, everyone. As promised, I'll write a longer post about how I entered the world of ML, hoping it will help someone shape their path. I'll include links to all the useful materials I used alongside the story, which you can use for learning. I like to call myself an AI Research Scientist who enjoys exploring new AI trends, delving deeper into understanding their background, and applying them to real products. This way, I try to connect science and entrepreneurship because I believe everything that starts as scientific research ends up "on the shelves" as a product that solves a specific user problem. I began my journey in ML in 2016 when it wasn't such a popular field. Everyone had heard of it, but few were applying it. I have several years of development experience and want to try my hand at ML. The first problem I encountered was where to start - whether to learn mathematics, statistics, or something else. That's when I came across a name and a course that completely changed my career. Let's start You guessed it. It was Professor Andrew Ng and his globally popular Machine Learning course available on Coursera (I still have the certificate, hehe). This was also my first official online course ever. Since that course no longer exists as it's been replaced by a new one, I recommend you check out: Machine Learning (Stanford CS229) Machine Learning Specialization These two courses start from the basics of ML and all the necessary calculus you need to know. Many always ask questions like whether to learn linear algebra, statistics, or probability, but you don't need to know everything in depth. This knowledge helps if you're a scientist developing a new architecture, but as an engineer, not really. You need to know some basics to understand, such as how the backpropagation algorithm works. I know that Machine Learning (Stanford CS229) is a very long and arduous course, but it's the right start if you want to be really good at ML. In my time, I filled two thick notebooks by hand while taking the course mentioned above. TensorFlow and Keras After the course, I didn't know how to apply my knowledge because I hadn't learned specifically how to code things. Then, I was looking for ways to learn how to code it. That's when I came across a popular framework called Keras, now part of TensorFlow. I started with a new course and acquiring practical knowledge: Deep Learning Specialization Deep Learning by Ian Goodfellow Machine Learning Yearning by Andrew Ng These resources above were my next step. I must admit that I learned the most from that course and from the book Deep Learning by Ian Goodfellow because I like reading books (although this one is quite difficult to read). Learn by coding To avoid just learning, I went through various GitHub repositories that I manually retyped and learned that way. It may be an old-fashioned technique, but it helped me a lot. Now, most of those repositories don't exist, so I'll share some that I found to be good: Really good Jupyter notebooks that can teach you the basics of TensorFlow Another good repo for learning TF and Keras Master the challenge After mastering the basics in terms of programming in TF/Keras, I wanted to try solving some real problems. There's no better place for that challenge than Kaggle and the popular Titanic dataset. Here, you can really find a bunch of materials and simple examples of ML applications. Here are some of my favorites: Titanic - Machine Learning from Disaster Home Credit Default Risk House Prices - Advanced Regression Techniques Two Sigma: Using News to Predict Stock Movements I then decided to further develop my career in the direction of applying ML to the stock market, first using predictions on time series and then using natural language processing. I've remained in this field until today and will defend my doctoral dissertation soon. How to deploy models To continue, before I move on to the topic of specialization, we need to address the topic of deployment. Now that we've learned how to make some basic models in Keras and how to use them, there are many ways and services, but I'll only mention what I use today. For all my ML models, whether simple regression models or complex GPT models, I use FastAPI. It's a straightforward framework, and you can quickly create API endpoints. I'll share a few older and useful tutorials for beginners: AI as an API tutorial series A step-by-step guide Productizing an ML Model with FastAPI and Cloud Run Personally, I've deployed on various cloud providers, of which I would highlight GCP and AWS because they have everything needed for model deployment, and if you know how to use them, they can be quite cheap. Chose your specialization The next step in developing my career, besides choosing finance as the primary area, was my specialization in the field of NLP. This happened in early 2020 when I started working with models based on the Transformer architecture. The first model I worked with was BERT, and the first tasks were related to classifications. My recommendations are to master the Transformer architecture well because 99% of today's LLM models are based on it. Here are some resources: The legendary paper "Attention Is All You Need" Hugging Face Course on Transformers Illustrated Guide to Transformers - Step by Step Explanation Good repository How large language models work, a visual intro to transformers After spending years using encoder-based Transformer models, I started learning GPT models. Good open-source models like Llama 2 then appear. Then, I started fine-tuning these models using the excellent Unsloth library: How to Finetune Llama-3 and Export to Ollama Fine-tune Llama 3.1 Ultra-Efficiently with Unsloth After that, I focused on studying various RAG techniques and developing Agent AI systems. This is now called AI engineering, and, as far as I can see, it has become quite popular. So I'll write more about that in another post, but here I'll leave what I consider to be the three most famous representatives, i.e., their tutorials: LangChain tutorial LangGraph tutorial CrewAI examples Here I am today Thanks to the knowledge I've generated over all these years in the field of ML, I've developed and worked on numerous projects. The most significant publicly available project is developing an agent AI system for well-being support, which I turned into a mobile application. Also, my entire doctoral dissertation is related to applying ML to the stock market in combination with the development of GPT models and reinforcement learning (more on that in a separate post). After long 6 years, I've completed my dissertation, and now I'm just waiting for its defense. I'll share everything I'm working on for the dissertation publicly on the project, and in tutorials I'm preparing to write. If you're interested in these topics, I announce that I'll soon start with activities of publishing content on Medium and a blog, but I'll share all of that here on Reddit as well. Now that I've gathered years of experience and knowledge in this field, I'd like to share it with others and help as much as possible. If you have any questions, feel free to ask them, and I'll try to answer all of them. Thank you for reading.

How I Started Learning Machine Learning
reddit
LLM Vibe Score0
Human Vibe Score1
TechPrimoThis week

How I Started Learning Machine Learning

Hello, everyone. As promised, I'll write a longer post about how I entered the world of ML, hoping it will help someone shape their path. I'll include links to all the useful materials I used alongside the story, which you can use for learning. I like to call myself an AI Research Scientist who enjoys exploring new AI trends, delving deeper into understanding their background, and applying them to real products. This way, I try to connect science and entrepreneurship because I believe everything that starts as scientific research ends up "on the shelves" as a product that solves a specific user problem. I began my journey in ML in 2016 when it wasn't such a popular field. Everyone had heard of it, but few were applying it. I have several years of development experience and want to try my hand at ML. The first problem I encountered was where to start - whether to learn mathematics, statistics, or something else. That's when I came across a name and a course that completely changed my career. Let's start You guessed it. It was Professor Andrew Ng and his globally popular Machine Learning course available on Coursera (I still have the certificate, hehe). This was also my first official online course ever. Since that course no longer exists as it's been replaced by a new one, I recommend you check out: Machine Learning (Stanford CS229) Machine Learning Specialization These two courses start from the basics of ML and all the necessary calculus you need to know. Many always ask questions like whether to learn linear algebra, statistics, or probability, but you don't need to know everything in depth. This knowledge helps if you're a scientist developing a new architecture, but as an engineer, not really. You need to know some basics to understand, such as how the backpropagation algorithm works. I know that Machine Learning (Stanford CS229) is a very long and arduous course, but it's the right start if you want to be really good at ML. In my time, I filled two thick notebooks by hand while taking the course mentioned above. TensorFlow and Keras After the course, I didn't know how to apply my knowledge because I hadn't learned specifically how to code things. Then, I was looking for ways to learn how to code it. That's when I came across a popular framework called Keras, now part of TensorFlow. I started with a new course and acquiring practical knowledge: Deep Learning Specialization Deep Learning by Ian Goodfellow Machine Learning Yearning by Andrew Ng These resources above were my next step. I must admit that I learned the most from that course and from the book Deep Learning by Ian Goodfellow because I like reading books (although this one is quite difficult to read). Learn by coding To avoid just learning, I went through various GitHub repositories that I manually retyped and learned that way. It may be an old-fashioned technique, but it helped me a lot. Now, most of those repositories don't exist, so I'll share some that I found to be good: Really good Jupyter notebooks that can teach you the basics of TensorFlow Another good repo for learning TF and Keras Master the challenge After mastering the basics in terms of programming in TF/Keras, I wanted to try solving some real problems. There's no better place for that challenge than Kaggle and the popular Titanic dataset. Here, you can really find a bunch of materials and simple examples of ML applications. Here are some of my favorites: Titanic - Machine Learning from Disaster Home Credit Default Risk House Prices - Advanced Regression Techniques Two Sigma: Using News to Predict Stock Movements I then decided to further develop my career in the direction of applying ML to the stock market, first using predictions on time series and then using natural language processing. I've remained in this field until today and will defend my doctoral dissertation soon. How to deploy models To continue, before I move on to the topic of specialization, we need to address the topic of deployment. Now that we've learned how to make some basic models in Keras and how to use them, there are many ways and services, but I'll only mention what I use today. For all my ML models, whether simple regression models or complex GPT models, I use FastAPI. It's a straightforward framework, and you can quickly create API endpoints. I'll share a few older and useful tutorials for beginners: AI as an API tutorial series A step-by-step guide Productizing an ML Model with FastAPI and Cloud Run Personally, I've deployed on various cloud providers, of which I would highlight GCP and AWS because they have everything needed for model deployment, and if you know how to use them, they can be quite cheap. Chose your specialization The next step in developing my career, besides choosing finance as the primary area, was my specialization in the field of NLP. This happened in early 2020 when I started working with models based on the Transformer architecture. The first model I worked with was BERT, and the first tasks were related to classifications. My recommendations are to master the Transformer architecture well because 99% of today's LLM models are based on it. Here are some resources: The legendary paper "Attention Is All You Need" Hugging Face Course on Transformers Illustrated Guide to Transformers - Step by Step Explanation Good repository How large language models work, a visual intro to transformers After spending years using encoder-based Transformer models, I started learning GPT models. Good open-source models like Llama 2 then appear. Then, I started fine-tuning these models using the excellent Unsloth library: How to Finetune Llama-3 and Export to Ollama Fine-tune Llama 3.1 Ultra-Efficiently with Unsloth After that, I focused on studying various RAG techniques and developing Agent AI systems. This is now called AI engineering, and, as far as I can see, it has become quite popular. So I'll write more about that in another post, but here I'll leave what I consider to be the three most famous representatives, i.e., their tutorials: LangChain tutorial LangGraph tutorial CrewAI examples Here I am today Thanks to the knowledge I've generated over all these years in the field of ML, I've developed and worked on numerous projects. The most significant publicly available project is developing an agent AI system for well-being support, which I turned into a mobile application. Also, my entire doctoral dissertation is related to applying ML to the stock market in combination with the development of GPT models and reinforcement learning (more on that in a separate post). After long 6 years, I've completed my dissertation, and now I'm just waiting for its defense. I'll share everything I'm working on for the dissertation publicly on the project, and in tutorials I'm preparing to write. If you're interested in these topics, I announce that I'll soon start with activities of publishing content on Medium and a blog, but I'll share all of that here on Reddit as well. Now that I've gathered years of experience and knowledge in this field, I'd like to share it with others and help as much as possible. If you have any questions, feel free to ask them, and I'll try to answer all of them. Thank you for reading.

List of free educational ML resources I used to become a FAANG ML Engineer
reddit
LLM Vibe Score0
Human Vibe Score1
aifordevsThis week

List of free educational ML resources I used to become a FAANG ML Engineer

Full commentary and notes here ➡️: https://www.trybackprop.com/blog/top\ml\learning\resources Used these to brush up on math and teach myself AI/ML over the course of two years. I'm now a staff ML engineer at FAANG. Hope these help. Fundamentals Linear Algebra – 3Blue1Brown's Essence of Linear Algebra series, binged all these videos on a one hour train ride visiting my parents Multivariable Calculus – Khan Academy's Multivariable Calculus lessons were a great refresher of what I had learned in college. Looking back, I just needed to have reviewed Unit 1 – intro and Unit 2 – derivatives. Calculus for ML – this amazing animated video explains calculus and backpropagation Information Theory – easy-to-understand book on information theory called Information Theory: A Tutorial Introduction. Statistics and Probability – the StatQuest YouTube channel Machine Learning Stanford Intro to Machine Learning by Andrew Ng – Stanford's CS229, the intro to machine learning course, published their lectures on YouTube for free. I watched lectures 1, 2, 3, 4, 8, 9, 11, 12, and 13, and I skipped the rest since I was eager to move onto deep learning. The course also offers a free set of course notes, which are very well written. Caltech Machine Learning – Caltech's machine learning lectures on YouTube, less mathematical and more intuition based Deep Learning Andrej Karpathy's Zero to Hero Series – Andrej Karpathy, an AI researcher who graduated with a Stanford PhD and led Tesla AI for several years, released an amazing series of hands on lectures on YouTube. highly highly recommend Neural networks – Stanford's CS231n course notes and lecture videos were my gateway drug*, so to speak, into the world of deep learning. Transformers and LLMs Transformers – watched these two lectures: lecture from the University of Waterloo and lecture from the University of Michigan. I have also heard good things about Jay Alammar's The Illustrated Transformer guide ChatGPT Explainer – Wolfram's YouTube explainer video on ChatGPT Interactive LLM Visualization – This LLM visualization that you can play with in your browser is hands down the best interactive experience with an LLM. Financial Times' Transformer Explainer – The Financial Times released a lovely interactive article that explains the transformer very well. Residual Learning – 2023 Future Science Prize Laureates Lecture on residual learning. Efficient ML and GPUs How are Microchips Made? – This YouTube video by Branch Education is one of the best free educational videos on the internet, regardless of subject, but also, it's the best video on understanding microchips. CUDA – My L8 and L9 FAANG coworkers acquired their CUDA knowledge from this series of lectures. TinyML and Efficient Deep Learning Computing – 2023 lectures on efficient ML techniques online. Chip War – Chip War is a bestselling book published in 2022 about microchip technology whose beginning chapters on the invention of the microchip actually explain CPUs very well

[Help Needed] Developing an AI to Play Mini Metro – Struggling with Data Extraction & Strategy method
reddit
LLM Vibe Score0
Human Vibe Score1
Primary_Cheesecake63This week

[Help Needed] Developing an AI to Play Mini Metro – Struggling with Data Extraction & Strategy method

Hello everyone ! First of all, please excuse my English if i do mistakes, as it is not my native language and I am not necessarily comfortable with it :) Regarding this project, I will explain my initial intention. I know very little about coding, but I enjoy it and have had some Python lessons, along with a few small personal projects for fun, mostly using YouTube tutorials. Nothing too advanced... However, now I want to take it to the next level. Since I have some familiarity with coding, I’ve wanted to work on artificial intelligence for a while. I have never coded AI myself, but I enjoy downloading existing projects (for chess, checkers, cat-and-mouse games, etc.), testing their limits, and understanding how they work. One of my favorite strategy game genres is management games, especially Mini Metro. Given its relatively simple mechanics, I assumed there would already be AI projects for it. But to my surprise, I could only find mods that add maps ! I admit that I am neither the best nor the most patient researcher, so I haven’t spent hours searching, but the apparent lack of projects for this game struck me. Maybe the community is just small ? I haven't looked deeply into it. So, I got it into my head to create my own AI. After all, everything is on the internet, and perseverance is key ! However, perseverance alone is not enough when you are not particularly experienced, so I am turning to the community to find knowledgeable people who can help me. The First Obstacle: Getting Game Data I quickly realized that the biggest challenge is that Mini Metro does not have an accessible API (at least, not one I could find). This means I cannot easily extract game data. My initial idea was to have an AI analyze the game, think about the best move, and then write out the actions to be performed, instead of coding a bot that directly manipulates the game. But first, I needed a way to retrieve and store game data. Attempt #1: Image Recognition (Failed) Since there was no API, I tried using image recognition to gather game data. Unfortunately, it was a disaster. I used mss for screenshots ,Tesseract for OCR, andNumPy to manipulate images in the HSV color space but it produced unreliable results : It detected many false positives (labeling empty spaces as stations) It failed to consistently detect numbers (scores or resources like trains and lines) Dotted bridge indicators over rivers were misinterpreted as stations While I could detect stations, lines, and moving trains, the data was chaotic and unreliable Attempt #2: Manual Data Entry (Partially Successful but Impractical) Since image recognition was unreliable, I decided to manually update the game data in real-time. I created a script that : Displays an overlay when I press Shift+R. Allows me to manually input stations, lines, and other game elements. Saves the current state when I press Shift+R again, so I can resume playing. Implements a simple resource management system (trains, lines, etc.). This works better than image recognition because I control the input, but I’m running into serious limitations : Some game mechanics are hard to implement manually (adding a station in the middle of a line, extending the correct line when two lines overlap at a station) Keeping track of station demands (the shapes passengers want to travel to) becomes overwhelming as the game progresses Updating the score in real-time is practically impossible manually, and the score is essential for training an AI (for my reward systems) My Dilemma At this point, I am unsure of how to proceed. My questions for the community : Am I going in the right direction? Should I continue improving my manual tracking system or is it a dead end? Should I have persevered with image recognition instead? Is there a better way to extract game data that I haven’t thought of? I would appreciate any guidance or ideas. Thanks in advance ! if you need more info, i have posted my codes here : https://github.com/Dmsday/mini\metro\data\analyzer (for the image detection version I'm not sure that it's the latest version aka the most "functional" version that I could do because I think I deleted it out of boredom...)

How do byte-level language models work?
reddit
LLM Vibe Score0
Human Vibe Score1
Additional-Ad-7043This week

How do byte-level language models work?

I've recently been trying to pre-train my own small language model on the tiny-series datasets on huggingface: https://huggingface.co/collections/nampdn-ai/tiny-series-6503910fd491144159519c70. I also wanted to use a model similar to MEGABYTE: https://arxiv.org/pdf/2305.07185.pdf, but I don't understand how using bytes would work. The only implementation I could find: https://github.com/lucidrains/MEGABYTE-pytorch used str(chr(max(32, token))) to decode any token (byte) to a character and put the embedding size as 256. Firstly, why 256 and not 256-32 as any values below 32 are ignored? Also, many byte-level models including this and ByteT5 mention that they can process any text sequence even in a multilingual setting, however how would that be true if we are only using one byte, would we have to move to 2 bytes or use an UNK token, and if we did use 2 bytes that would make our embedding size around 65000 which defeats sort of the point as one of the advantages mentioned is that we are able to use a small embedding matrix? Furthermore, most language models add special tokens like bos, eos, unk and even for llama they use beginning of instruction, end of instruction, and more for system instructions, response, context... Should I use something like this as my dataset has some structures where there is a context, instruction and response, and if i did how would I add these if I'm using byte-level encodings? Final questions: Firstly, for the datasets mentioned including code,stories,webtext,... would I tokenise all of these datasets then concatenate them to then randomly sample from, or should i train seperately on each as some like code and webtext are much larger than the others? Finally, for the webtext part of the dataset, there is a passage of text then a passage analysing the text (main ideas,purpose,...), how should I encode this, should I use an extra ANALYSE token or just concatenate? Thank you for reading this far, I am sort of a beginner so if I said something stupid please point it out. Also, if there were unclear parts in my question I'm sorry as I struggled how to word these questions. Any help would be appreciated!

Scratch Machine Learning Algorithms Implementations
reddit
LLM Vibe Score0
Human Vibe Score1
ParkMountainThis week

Scratch Machine Learning Algorithms Implementations

Hi there, other Redditors! Like many of you, when I first started working in the AI field, I wanted to build some basic Machine Learning models from scratch in order to better understand how each algorithm works, improve my programming and math skills, or simply produce an eye-catching, difficult project to put in the résumé. After spending some time searching for resources that could help me guide my studies, I discovered that the majority of scratch implementations that are currently available are either i) outdated (having been implemented years ago using Python 2 or an earlier version of Python 3); ii) too difficult to understand (using a lot of difficult, unfriendly optimization techniques or with poorly written code); or iii) too simple (only covering binary classification). With that in mind, I made the decision to develop user-friendly, uncomplicated, organized, and simple implementations from scratch. Aside from all of that, I've always wanted to create an open-source project so that others, particularly novices and those with less than a year's experience (like me), can collaborate with others, contribute to public projects, and experience Git firsthand (some of these implementations were made by other contributors!). Here are some implementations that are available: Algorithms (Random Forest Classifier and Regressor, Decision Tree Classifier and Regressor, KMeans, KNN Classifier and Regressor, Gaussian Naive Bayes, Linear Regression, Logistic Regression, PCA, Perceptron, MLP Classifier and Regressor, SVM Classifier and Regressor); Regression and classification metrics; Distance metrics (such as Euclidean); Data split functions (such as KFold); Activation and loss functions; Scalers (such as MinMaxScaler) and encoders (such as One Hot Encoder); and a few things more! Project's link: https://github.com/rafaelgreca/scratchml Disclaimer: The goal of this library is to provide code that is simpler, easier to understand, and more approachable for artificial intelligence enthusiasts and beginners who want to contribute to an open-source repository or who want to learn more about how algorithms work. It is not meant to replace existing libraries that are better, more optimized, and have a wider variety of implemented algorithms (such as scikit-learn, PyTorch, Keras, and Tensorflow). If you want to use optimized implementations with accurate results, please use one of the previously mentioned libraries. P.S.: I accidentally deleted the other post, so I am posting again. :-)

Month of August in AI
reddit
LLM Vibe Score0
Human Vibe Score1
Difficult-Race-1188This week

Month of August in AI

🔍 Inside this Issue: 🤖 Latest Breakthroughs: This month it’s all about Agents, LangChain RAG, and LLMs evaluation challenges.* 🌐 AI Monthly News: Discover how these stories are revolutionizing industries and impacting everyday life: EU AI Act, California’s Controversial SB1047 AI regulation act, Drama at OpenAI, and possible funding at OpenAI by Nvidia and Apple.* 📚 Editor’s Special: This covers the interesting talks, lectures, and articles we came across recently. Follow me on Twitter and LinkedIn at RealAIGuys and AIGuysEditor to get insight on new AI developments. Please don't forget to subscribe to our Newsletter: https://medium.com/aiguys/newsletter Latest Breakthroughs Are Agents just simple rules? Are Agents just enhanced reasoning? The answer is yes and no. Yes, in the sense that agents have simple rules and can sometimes enhance reasoning capabilities compared to a single prompt. But No in the sense that agents can have a much more diverse functionality like using specific tools, summarizing, or even following a particular style. In this blog, we look into how to set up these agents in a hierarchal manner just like running a small team of Authors, researchers, and supervisors. How To Build Hierarchical Multi-Agent Systems? TextGrad. It is a powerful framework performing automatic “differentiation” via text. It backpropagates textual feedback provided by LLMs to improve individual components of a compound AI system. In this framework, LLMs provide rich, general, natural language suggestions to optimize variables in computation graphs, ranging from code snippets to molecular structures. TextGrad showed effectiveness and generality across various applications, from question-answering and molecule optimization to radiotherapy treatment planning. TextGrad: Improving Prompting Using AutoGrad The addition of RAG to LLMs was an excellent idea. It helped the LLMs to become more specific and individualized. Adding new components to any system leads to more interactions and its own sets of problems. Adding RAG to LLMs leads to several problems such as how to retrieve the best content, what type of prompt to write, and many more. In this blog, we are going to combine the LangChain RAG with DSPy. We deep dive into how to evaluate the RAG pipeline quantitatively using RAGAs and how to create a system where instead of manually tweaking prompts, we let the system figure out the best prompt. How To Build LangChain RAG With DSPy? As the field of natural language processing (NLP) advances, the evaluation of large language models (LLMs) like GPT-4 becomes increasingly important and complex. Traditional metrics such as accuracy are often inadequate for assessing these models’ performance because they fail to capture the nuances of human language. In this article, we will explore why evaluating LLMs is challenging and discuss effective methods like BLEU and ROUGE for a more comprehensive evaluation. The Challenges of Evaluating Large Language Models AI Monthly News AI Act enters into force On 1 August 2024, the European Artificial Intelligence Act (AI Act) enters into force. The Act aims to foster responsible artificial intelligence development and deployment in the EU. The AI Act introduces a uniform framework across all EU countries, based on a forward-looking definition of AI and a risk-based approach: Minimal risk: most AI systems such as spam filters and AI-enabled video games face no obligation under the AI Act, but companies can voluntarily adopt additional codes of conduct. Specific transparency risk: systems like chatbots must clearly inform users that they are interacting with a machine, while certain AI-generated content must be labelled as such. High risk: high-risk AI systems such as AI-based medical software or AI systems used for recruitment must comply with strict requirements, including risk-mitigation systems, high-quality of data sets, clear user information, human oversight, etc. Unacceptable risk: for example, AI systems that allow “social scoring” by governments or companies are considered a clear threat to people’s fundamental rights and are therefore banned. EU announcement: Click here https://preview.redd.it/nwyzfzgm4cmd1.png?width=828&format=png&auto=webp&s=c873db37ca0dadd5b510bea70ac9f633b96aaea4 California AI bill SB-1047 sparks fierce debate, Senator likens it to ‘Jets vs. Sharks’ feud Key Aspects of SB-1047: Regulation Scope: Targets “frontier” AI models, defined by their immense computational training requirements (over 10²⁶ operations) or significant financial investment (>$100 million). Compliance Requirements: Developers must implement safety protocols, including the ability to immediately shut down, cybersecurity measures, and risk assessments, before model deployment. Whistleblower Protections: Encourages reporting of non-compliance or risks by offering protection against retaliation. Safety Incident Reporting: Mandates reporting AI safety incidents within 72 hours to a newly established Frontier Model Division. Certification: Developers need to certify compliance, potentially under penalty of perjury in earlier drafts, though amendments might have altered this. Pros: Safety First: Prioritizes the prevention of catastrophic harms by enforcing rigorous safety standards, potentially safeguarding against AI misuse or malfunction. Incentivizes Responsible Development: By setting high standards for AI model training, the company encourages developers to think critically about the implications of their creations. Public Trust: Enhances public confidence in AI by ensuring transparency and accountability in the development process. Cons: Innovation Stagnation: Critics argue it might stifle innovation, especially in open-source AI, due to the high costs and regulatory burdens of compliance. Ambiguity: Some definitions and requirements might be too specific or broad, leading to legal challenges or unintended consequences. Global Competitiveness: There’s concern that such regulations could push AI development outside California or the U.S., benefiting other nations without similar restrictions. Implementation Challenges: The practicalities of enforcing such regulations, especially the “positive safety determination,” could be complex and contentious. News Article: Click here Open Letter: Click here https://preview.redd.it/ib96d7nk4cmd1.png?width=828&format=png&auto=webp&s=0ed5913b5dae72e203c8592393e469d9130ed689 MORE OpenAI drama OpenAI co-founder John Schulman has left the company to join rival AI startup Anthropic, while OpenAI president and co-founder Greg Brockman is taking an extended leave until the end of the year. Schulman, who played a key role in creating the AI-powered chatbot platform ChatGPT and led OpenAI’s alignment science efforts, stated his move was driven by a desire to focus more on AI alignment and hands-on technical work. Peter Deng, a product manager who joined OpenAI last year, has also left the company. With these departures, only three of OpenAI’s original 11 founders remain: CEO Sam Altman, Brockman, and Wojciech Zaremba, lead of language and code generation. News Article: Click here https://preview.redd.it/0vdjc18j4cmd1.png?width=828&format=png&auto=webp&s=e9de604c26aed3e47b50df3bdf114ef61f967080 Apple and Nvidia may invest in OpenAI Apple, which is planning to integrate ChatGPT into iOS, is in talks to invest. Soon after, Bloomberg also reported that Apple is in talks but added that Nvidia “has discussed” joining the funding round as well. The round is reportedly being led by Thrive Capital and would value OpenAI at more than $100 billion. News Article: Click here https://preview.redd.it/ude6jguh4cmd1.png?width=828&format=png&auto=webp&s=3603cbca0dbb1be3e6d0efcf06c3a698428bbdd6 Editor’s Special The AI Bubble: Will It Burst, and What Comes After?: Click here Eric Schmidt Full Controversial Interview on AI Revolution (Former Google CEO): Click here AI isn’t gonna keep improving Click here General Intelligence: Define it, measure it, build it: Click here

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.
reddit
LLM Vibe Score0
Human Vibe Score0.714
level6-killjoyThis week

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.

This is a recap covering the major news from last week. 🔥Top 3 news - OpenAI’s updates, Meta’s upcoming free LLM and EU Regulation 🗞️Interesting reads include PSA about protecting your keys, The GPT ouroboros, Reddit - OpenAI’s moat, and more.. 🧑‍🎓Learning includes a Step-by-step guide from a non-technical founder who launched his MVP, Chatbot for your Gdrive and more 🔥Top 3 AI news in the past week OpenAI: New Pricing, Models, & Functions OpenAI has been on a roll. Last week we saw the release of OpenAI best practice on using GPT. This week we saw some amazing updates. Three major buckets were: First, the price decreases for both embeddings and GPT-3.5 tokens. Second, new models for gpt-4 and gpt-3.5. A new longer context model for gpt-3.5. Third, a new function calling capability. Why is it important? Previously, the output from OpenAI was all text. So, calling an external API from GPT was quite difficult. You had to parse the text data and things were often incorrect. Langchain created the Agents and Tools feature to tackle this problem. It was still unreliable and prone to issues. Now you get native support to generate a fixed format output. You can use the output to generate functional calls and also pass functions which need to be called. For example, if your app has multiple API endpoints then you can use GPT to generate the API calls with parameters. You can also pass the endpoints as function calls to ensure the correct function is executed. This functionality can further be used to generate structured data (JSON) out of GPT. So, you can generate data from GPT and load it into your backend. What’s next? This functionality allows turning natural language responses into structured data. This can be used to create “intelligent” backends using LLMs. We might see implementations in no-code tools to allow more robust and natural-language tools for non-technical folks. The structured data process goes both ways. You can also feed structured data into GPT for better responses. This feature also has its share of issues. Function calling suffers from the same prompt injection issues. Malicious actors can pass malicious code in function or the responses. For example, creation of queries using functions might contain malicious code to delete data. Without proper user validation this code will be executed automatically and delete data. So, using LLM as the back-end layer needs proper security implementation. Meta's LLM: Commercial Use Ahead Llama has been a boon for the open source community. Many of the open source models rely on Llama. The issue is that Llama is research-only and cannot be used commercially. So, no one can use it to build any product. Meta is now working on the next version of the model. This model will be available for commercial use. This is in stark contrast to both OpenAI and Google. Both safe-guarde their models and make it available through API. Why is it important? Certain industries cannot use LLM APIs because of strict restrictions on data privacy. These companies would want to run their own instance of a foundational model. A commercially available foundational model is also going to help people who want to keep their “API call” costs next to 0. A commercially available free-for-all model will also help push the open source community further. Just like Llama. What’s next? Sam Altman has said OpenAI didn’t release GPT-3 as open-source because they didn’t think people would be able to run it. Now OpenAI is working on an open-source model. This is going to be weaker than GPT-4. Let the battle of LLMs begin. EU's Proposed Legislation and Its Impact on AI Usage The EU parliament voted to move ahead with the E.U. AI Act. This act aims to ensure consumer protection against the dangers of AI. Why is it important? OpenAI and Sam Altman want regulations for models. They have proposed a IAEA-type of agency to stop the proliferation of LLM models. As per OpenAI, all models should be regulated and monitored. The suggestion of a license based regulation has led to significant backlash. Many people have called it “regulatory capture” - with the aim of shutting down competing LLMs. Licensing based regulations might not really be effective. The EU is approaching regulation from a different angle. It doesn’t focus on how models are developed. Rather focuses on how AI will/can be used. They have broken down use cases into 4 categories - unacceptable (prohibited), high, medium and low risk. For example, Building a Pre-Crime software,on%20crimes%20not%20yet%20committed.) to predict crimes? Building a Social credit system? Unacceptable. Using tools to influence elections or recommendation algorithms? High (Highly regulated). Using generative AI tools to create text or images on news sites? Medium (Add label that the content is AI generated) AI providers also need to disclose their training source. To me this sounds like good legislation. What do you guys think? But, OpenAI has warned that EU regulations might force them to pull out completely. What’s next? The disclosure requirements might help various publishing companies. AI and media companies are in talks to pay for training data. Google has been leading the charge. Additionally, OpenAI and Deepmind will open their models for safety and research purposes to the UK government. 🗞️10 AI news highlights and interesting reads PSA: If you are using Repl to write code, you might want to check your OpenAI API keys. If you have left them embedded then people can pirate and steal the keys. LLMs rely on human annotation or human feedback to learn. And one way to generate human annotation is crowdsourcing. But what if the crowdsource human annotators use LLMs? Research shows 33-46% workers used LLMs. So, basically we go from Human -> AI -> Human -> AI. The AI ouroboros. Researchers also say generated data to train models might cause serious issue. All the talks about moats \- Reddit might be OpenAI’s \future\ moat. Given the amount of complaints about how Google search experience has deteriorated during the blackout, this might be true? Doctors are using ChatGPT but not to diagnose.Rather to be more empathetic. We discussed this just a month ago. And guess where the data for this study came from? Reddit AskDocs. Moat FTW?! Beatles to make a comeback…using Generative AI. SnapFusion - Text to Image diffusion on mobile phones. Large context lengths are important for better GPT experience. The secret sauce for 100k context length. There is a lot of bad AI research out there. Some border on snake oil. Most AI “research” should be double checked and challenged. A new research on huggingface said that GPT-4 can ace MIT curriculum. Now someone is replicating the results and say that GPT-4 can’t beat MIT. Are we seeing peak AI? Especially when people from Deepmind and Meta are involved? Mistral AI raised $113 million in seed round with no product. Some might say this funding is for the team and the team is really solid. The issue though is whether the valuation is justified when OpenAI and Google already have a head start. The AI Hype Wall of Shame. \- Collection of articles which mislead people about AI in various aspects. 🧑‍🎓3 Learning Resources Building and Launching a company using GPT-4 with prompts. (The author didn’t know how to code but created and launched the MVP in a month). Chatbot for your Gdrive - https://www.haihai.ai/gpt-gdrive/ Building ChatGPT plugin using Supabase - https://supabase.com/blog/building-chatgpt-plugins-template That’s it folks. Thank you for reading and have a great week ahead. If you are interested in a focused weekly recap delivered to your inbox on Mondays you can subscribe here. It is FREE!

Study Plan for Learning Data Science Over the Next 12 Months [D]
reddit
LLM Vibe Score0
Human Vibe Score1
daniel-dataThis week

Study Plan for Learning Data Science Over the Next 12 Months [D]

In this thread, I address a study plan for 2021. In case you're interested, I wrote a whole article about this topic: Study Plan for Learning Data Science Over the Next 12 Months Let me know your thoughts on this. &#x200B; https://preview.redd.it/emg20nzhet661.png?width=1170&format=png&auto=webp&s=cf09e4dc5e82ba2fd7b57c706ba2873be57fe8de We are ending 2020 and it is time to make plans for next year, and one of the most important plans and questions we must ask is what do we want to study?, what do we want to enhance?, what changes do we want to make?, and what is the direction we are going to take (or continue) in our professional careers?. Many of you will be starting on the road to becoming a data scientist, in fact you may be evaluating it, since you have heard a lot about it, but you have some doubts, for example about the amount of job offers that may exist in this area, doubts about the technology itself, and about the path you should follow, considering the wide range of options to learn. I’m a believer that we should learn from various sources, from various mentors, and from various formats. By sources I mean the various virtual platforms and face-to-face options that exist to study. By mentors I mean that it is always a good idea to learn from different points of view and learning from different teachers/mentors, and by formats I mean the choices between books, videos, classes, and other formats where the information is contained. When we extract information from all these sources we reinforce the knowledge learned, but we always need a guide, and this post aims to give you some practical insights and strategies in this regard. To decide on sources, mentors and formats it is up to you to choose. It depends on your preferences and ease of learning: for example, some people are better at learning from books, while others prefer to learn from videos. Some prefer to study on platforms that are practical (following online code), and others prefer traditional platforms: like those at universities (Master’s Degree, PHDs or MOOCs). Others prefer to pay for quality content, while others prefer to look only for free material. That’s why I won’t give a specific recommendation in this post, but I’ll give you the whole picture: a study plan. To start you should consider the time you’ll spend studying and the depth of learning you want to achieve, because if you find yourself without a job you could be available full time to study, which is a huge advantage. On the other hand, if you are working, you’ll have less time and you’ll have to discipline yourself to be able to have the time available in the evenings, mornings or weekends. Ultimately, the important thing is to meet the goal of learning and perhaps dedicating your career to this exciting area! We will divide the year into quarters as follows First Quarter: Learning the Basics Second Quarter: Upgrading the Level: Intermediate Knowledge Third Quarter: A Real World Project — A Full-stack Project Fourth Quarter: Seeking Opportunities While Maintaining Practice First Quarter: Learning the Basics &#x200B; https://preview.redd.it/u7t9bthket661.png?width=998&format=png&auto=webp&s=4ad29cb43618e7acf793259243aa5a60a8535f0a If you want to be more rigorous you can have start and end dates for this period of study of the bases. It could be something like: From January 1 to March 30, 2021 as deadline. During this period you will study the following: A programming language that you can apply to data science: Python or R. We recommend Python due to the simple fact that approximately 80% of data science job offers ask for knowledge in Python. That same percentage is maintained with respect to the real projects you will find implemented in production. And we add the fact that Python is multipurpose, so you won’t “waste” your time if at some point you decide to focus on web development, for example, or desktop development. This would be the first topic to study in the first months of the year. Familiarize yourself with statistics and mathematics. There is a big debate in the data science community about whether we need this foundation or not. I will write a post later on about this, but the reality is that you DO need it, but ONLY the basics (at least in the beginning). And I want to clarify this point before continuing. We could say that data science is divided in two big fields: Research on one side and putting Machine Learning algorithms into production on the other side. If you later decide to focus on Research then you are going to need mathematics and statistics in depth (very in depth). If you are going to go for the practical part, the libraries will help you deal with most of it, under the hood. It should be noted that most job offers are in the practical part. For both cases, and in this first stage you will only need the basics of: Statistics (with Python and NumPy) Descriptive statistics Inferential Statistics Hypothesis testing Probability Mathematics (with Python and NumPy) Linear Algebra (For example: SVD) Multivariate Calculus Calculus (For example: gradient descent) Note: We recommend that you study Python first before seeing statistics and mathematics, because the challenge is to implement these statistical and mathematical bases with Python. Don’t look for theoretical tutorials that show only slides or statistical and/or mathematical examples in Excel/Matlab/Octave/SAS and other different to Python or R, it gets very boring and impractical! You should choose a course, program or book that teaches these concepts in a practical way and using Python. Remember that Python is what we finally use, so you need to choose well. This advice is key so you don’t give up on this part, as it will be the most dense and difficult. If you have these basics in the first three months, you will be ready to make a leap in your learning for the next three months. Second Quarter: Upgrading the Level: Intermediate Knowledge &#x200B; https://preview.redd.it/y1y55vynet661.png?width=669&format=png&auto=webp&s=bd3e12bb112943025c39a8975faf4d64514df275 If you want to be more rigorous you can have start and end dates for this period of study at the intermediate level. It could be something like: From April 1 to June 30, 2021 as deadline. Now that you have a good foundation in programming, statistics and mathematics, it is time to move forward and learn about the great advantages that Python has for applying data analysis. For this stage you will be focused on: Data science Python stack Python has the following libraries that you should study, know and practice at this stage Pandas: for working with tabular data and make in-depth analysis Matplotlib and Seaborn: for data visualization Pandas is the in-facto library for data analysis, it is one of the most important (if not the most important) and powerful tools you should know and master during your career as a data scientist. Pandas will make it much easier for you to manipulate, cleanse and organize your data. Feature Engineering Many times people don’t go deep into Feature Engineering, but if you want to have Machine Learning models that make good predictions and improve your scores, spending some time on this subject is invaluable! Feature engineering is the process of using domain knowledge to extract features from raw data using data mining techniques. These features can be used to improve the performance of machine learning algorithms. Feature engineering can be considered as applied machine learning itself. To achieve the goal of good feature engineering you must know the different techniques that exist, so it is a good idea to at least study the main ones. Basic Models of Machine Learning At the end of this stage you will start with the study of Machine Learning. This is perhaps the most awaited moment! This is where you start to learn about the different algorithms you can use, which particular problems you can solve and how you can apply them in real life. The Python library we recommend you to start experimenting with ML is: scikit-learn. However it is a good idea that you can find tutorials where they explain the implementation of the algorithms (at least the simplest ones) from scratch with Python, since the library could be a “Black Box” and you might not understand what is happening under the hood. If you learn how to implement them with Python, you can have a more solid foundation. If you implement the algorithms with Python (without a library), you will put into practice everything seen in the statistics, mathematics and Pandas part. These are some recommendations of the algorithms that you should at least know in this initial stage Supervised learning Simple Linear Regression Multiple Linear Regression K-nearest neighbors (KNN) Logistic Regression Decision Trees Random Forest Unsupervised Learning K-Means PCA Bonus: if you have the time and you are within the time ranges, you can study these others Gradient Boosting Algorithms GBM XGBoost LightGBM CatBoost Note: do not spend more than the 3 months stipulated for this stage. Because you will be falling behind and not complying with the study plan. We all have shortcomings at this stage, it is normal, go ahead and then you can resume some concepts that did not understand in detail. The important thing is to have the basic knowledge and move forward! If at least you succeed to study the mentioned algorithms of supervised and unsupervised learning, you will have a very clear idea of what you will be able to do in the future. So don’t worry about covering everything, remember that it is a process, and ideally you should have some clearly established times so that you don’t get frustrated and feel you are advancing. So far, here comes your “theoretical” study of the basics of data science. Now we’ll continue with the practical part! Third Quarter: A Real World Project — A Full-stack Project &#x200B; https://preview.redd.it/vrn783vqet661.png?width=678&format=png&auto=webp&s=664061b3d33b34979b74b10b9f8a3d0f7b8b99ee If you want to be more rigorous you can have start and end dates for this period of study at the intermediate level. It could be something like: From July 1 to September 30, 2021 as deadline. Now that you have a good foundation in programming, statistics, mathematics, data analysis and machine learning algorithms, it is time to move forward and put into practice all this knowledge. Many of these suggestions may sound out of the box, but believe me they will make a big difference in your career as a data scientist. The first thing is to create your web presence: Create a Github (or GitLab) account, and learn Git*. Being able to manage different versions of your code is important, you should have version control over them, not to mention that having an active Github account is very valuable in demonstrating your true skills. On Github, you can also set up your Jupyter Notebooks and make them public, so you can show off your skills as well. This is mine for example: https://github.com/danielmoralesp Learn the basics of web programming*. The advantage is that you already have Python as a skill, so you can learn Flask to create a simple web page. Or you can use a template engine like Github Pages, Ghost or Wordpress itself and create your online portfolio. Buy a domain with your name*. Something like myname.com, myname.co, myname.dev, etc. This is invaluable so you can have your CV online and update it with your projects. There you can make a big difference, showing your projects, your Jupyter Notebooks and showing that you have the practical skills to execute projects in this area. There are many front-end templates for you to purchase for free or for payment, and give it a more personalized and pleasant look. Don’t use free sub-domains of Wordpress, Github or Wix, it looks very unprofessional, make your own. Here is mine for example: https://www.danielmorales.dev/ Choose a project you are passionate about and create a Machine Learning model around it. The final goal of this third quarter is to create ONE project, that you are passionate about, and that is UNIQUE among others. It turns out that there are many typical projects in the community, such as predicting the Titanic Survivors, or predicting the price of Houses in Boston. Those kinds of projects are good for learning, but not for showing off as your UNIQUE projects. If you are passionate about sports, try predicting the soccer results of your local league. If you are passionate about finance, try predicting your country’s stock market prices. If you are passionate about marketing, try to find someone who has an e-commerce and implement a product recommendation algorithm and upload it to production. If you are passionate about business: make a predictor of the best business ideas for 2021 :) As you can see, you are limited by your passions and your imagination. In fact, those are the two keys for you to do this project: Passion and Imagination. However don’t expect to make money from it, you are in a learning stage, you need that algorithm to be deployed in production, make an API in Flask with it, and explain in your website how you did it and how people can access it. This is the moment to shine, and at the same time it’s the moment of the greatest learning. You will most likely face obstacles, if your algorithm gives 60% of Accuracy after a huge optimization effort, it doesn’t matter, finish the whole process, deploy it to production, try to get a friend or family member to use it, and that will be the goal achieved for this stage: Make a Full-stack Machine Learning project. By full-stack I mean that you did all the following steps: You got the data from somewhere (scrapping, open data or API) You did a data analysis You cleaned and transformed the data You created Machine Learning Models You deployed the best model to production for other people to use. This does not mean that this whole process is what you will always do in your daily job, but it does mean that you will know every part of the pipeline that is needed for a data science project for a company. You will have a unique perspective! Fourth Quarter: Seeking Opportunities While Maintaining Practice &#x200B; https://preview.redd.it/qd0osystet661.png?width=1056&format=png&auto=webp&s=2da456b15985b2793041256f5e45bca99a23b51a If you want to be more rigorous you can have start and end dates for this period of study at the final level. It could be something like: From October 1 to December 31, 2021 as deadline. Now you have theoretical and practical knowledge. You have implemented a model in production. The next step depends on you and your personality. Let’s say you are an entrepreneur, and you have the vision to create something new from something you discovered or saw an opportunity to do business with this discipline, so it’s time to start planning how to do it. If that’s the case, obviously this post won’t cover that process, but you should know what the steps might be (or start figuring them out). But if you are one of those who want to get a job as a data scientist, here is my advice. Getting a job as a data scientist “You’re not going to get a job as fast as you think, if you keep thinking the same way”.Author It turns out that all people who start out as data scientists imagine themselves working for the big companies in their country or region. Or even remote. It turns out that if you aspire to work for a large company like data scientist you will be frustrated by the years of experience they ask for (3 or more years) and the skills they request. Large companies don’t hire Juniors (or very few do), precisely because they are already large companies. They have the financial muscle to demand experience and skills and can pay a commensurate salary (although this is not always the case). The point is that if you focus there you’re going to get frustrated! Here we must return to the following advise: “You need creativity to get a job in data science”. Like everything else in life we have to start at different steps, in this case, from the beginning. Here are the scenarios If you are working in a company and in a non-engineering role you must demonstrate your new skills to the company you are working for*. If you are working in the customer service area, you should apply it to your work, and do for example, detailed analysis of your calls, conversion rates, store data and make predictions about it! If you can have data from your colleagues, you could try to predict their sales! This may sound funny, but it’s about how creatively you can apply data science to your current work and how to show your bosses how valuable it is and EVANGELIZE them about the benefits of implementation. You’ll be noticed and they could certainly create a new data related department or job. And you already have the knowledge and experience. The key word here is Evangelize. Many companies and entrepreneurs are just beginning to see the power of this discipline, and it is your task to nurture that reality. If you are working in an area related to engineering, but that is not data science*. Here the same applies as the previous example, but you have some advantages, and that is that you could access the company’s data, and you could use it for the benefit of the company, making analyses and/or predictions about it, and again EVANGELIZING your bosses your new skills and the benefits of data science. If you are unemployed (or do not want, or do not feel comfortable following the two examples above)*, you can start looking outside, and what I recommend is that you look for technology companies and / or startups where they are just forming the first teams and are paying some salary, or even have options shares of the company. Obviously here the salaries will not be exorbitant, and the working hours could be longer, but remember that you are in the learning and practice stage (just in the first step), so you can not demand too much, you must land your expectations and fit that reality, and stop pretending to be paid $ 10,000 a month at this stage. But, depending of your country $1.000 USD could be something very interesting to start this new career. Remember, you are a Junior at this stage. The conclusion is: don’t waste your time looking at and/or applying to offers from big companies, because you will get frustrated. Be creative, and look for opportunities in smaller or newly created companies. Learning never stops While you are in that process of looking for a job or an opportunity, which could take half of your time (50% looking for opportunities, 50% staying in practice), you have to keep learning, you should advance to concepts such as Deep Learning, Data Engineer or other topics that you feel were left loose from the past stages or focus on the topics that you are passionate about within this group of disciplines in data science. At the same time you can choose a second project, and spend some time running it from end-to-end, and thus increase your portfolio and your experience. If this is the case, try to find a completely different project: if the first one was done with Machine Learning, let this second one be done with Deep learning. If the first one was deployed to a web page, that this second one is deployed to a mobile platform. Remember, creativity is the key! Conclusion We are at an ideal time to plan for 2021, and if this is the path you want to take, start looking for the platforms and media you want to study on. Get to work and don’t miss this opportunity to become a data scientist in 2021! Note: we are building a private community in Slack of data scientist, if you want to join us write to the email: support@datasource.ai I hope you enjoyed this reading! you can follow me on twitter or linkedin Thank you for reading!

Let’s Build One Person Business Using 100% AI
reddit
LLM Vibe Score0
Human Vibe Score1
AssistanceOk2217This week

Let’s Build One Person Business Using 100% AI

AI made it possible for 9-to-5 workers to start a one-person business without quitting their jobs. Full Article https://preview.redd.it/tynb9y6z695d1.png?width=1309&format=png&auto=webp&s=b490d3676a63adcc01faff8c476056cb7d420022 https://i.redd.it/9x3okti0795d1.gif The Opportunities for Starting a Business ○ There are huge opportunities to start your own business by leveraging valuable skills to attract paying audiences. ○ New software and AI platforms make it easier to distribute products/services and automate tasks that were previously time-consuming. Our One Person Book Publication House ○ This article explores building a one-person AI-powered business focused on publishing books. ○ Users input data on a topic, and AI generates a comprehensive book structure and content based on that. ○ The generated content can be formatted, designed, and published digitally or in print easily. Why Read This Article? ○ It presents an innovative AI-powered approach to streamline the book publishing process. ○ It provides technical implementation details using LLM, Python and the Streamlit library as a reference. ○ It highlights AI's potential in automating creative tasks like writing and content creation. Approaching the One Person Business ○ Reflect on areas where you overcame personal struggles and gained valuable skills. ○ Leverage that expertise to build an AI business serving others facing similar obstacles. ○ Use AI tools to create content, automate processes, and efficiently scale your offerings. The Publication Business Idea ○ Focus on writing and publishing small books using AI writing assistants. ○ AI can streamline research, writing drafts, outlines, and ideas across genres. ○ Concentrate efforts on editing, formatting, and marketing while AI handles writing. The Book Generation Process ○ Users input structured topic data like outlines, key points, and references. ○ Advanced AI language models generate flowing book content from that data. ○ Minimal human effort is needed beyond initial inputs and refinement. ○ AI systems automatically handle formatting, design, and publishing. Technical Implementation ○ Includes a Book class to represent a book's hierarchical structure in Python. ○ Functions to generate book structures and section content using AI models. ○ Integrates with a Streamlit app for user input and output. ○ Allows downloading the final book in Markdown format. Closing Thoughts ○ This AI-powered approach makes book writing and publishing more accessible to individuals. ○ AI handles the heavy lifting, with humans providing quality control through editing. ○ It opens up possibilities for innovative knowledge sharing as technology evolves.

Master AI Integration: How to Integrate AI in Your Application
reddit
LLM Vibe Score0
Human Vibe Score1
AssistanceOk2217This week

Master AI Integration: How to Integrate AI in Your Application

A Comprehensive Guide with Every Detail Spelled Out for Flawless AI Implementation Full Article &#x200B; https://preview.redd.it/m5b79j55f14d1.png?width=1328&format=png&auto=webp&s=8cf04c80cd21be1710dd117a9e74b07d0e8cbe6a In the ideal world, we'd design our software systems with AI in mind from the very beginning. But in the real world, that's not always possible. Many businesses have large, complex systems that have been running for years, and making significant changes to them is risky and expensive. What this Article is About? ● This article aims to convince you that even when changing existing systems is not an option, you can still seamlessly integrate AI into your business processes. It explores real-world scenarios and shows how a company (though simulated) has successfully incorporated AI without overhauling their existing infrastructure. &#x200B; https://i.redd.it/fayl1gcbf14d1.gif Why Read This Article? ● By reading this article, you will learn the critical skill of integrating AI into your existing business ecosystem without making significant changes to your stable workflows. This skill is becoming increasingly important as more and more companies recognize the value of AI while also acknowledging the challenges of overhauling their existing systems. What is Our Business Use Case? ● The article uses a simulated supply chain management company as a business use case. This company has multiple departments, each exposing its own REST API, and to get an inquiry answered, the request has to go through various departments, their respective APIs, and database calls. The article introduces AI capabilities to enhance the company's operations without modifying the existing system architecture. Our Supply Chain Management Company AI Integration Design ● The article describes the various components of the simulated supply chain management company, including the "Data Processing System," "Company Data Handling System," "AI Integration System," "Mapping System," and "System Admin Dashboard." Let's Get Cooking! ● This section provides the code and explanations for implementing the AI integration system in the simulated supply chain management company. It covers the following: ○ Dashboard & AI Integration System ○ Company Data Handling System ○ Data Processing System ○ Mapping System Let's Setup ● This section shows the expected output when setting up the simulated supply chain management system with AI integration. Let's Run it ● This section demonstrates how to run the system and ask questions related to supply chain management, showcasing the AI integration in action. https://i.redd.it/3e68mb57f14d1.gif Closing Thoughts The supply chain management project we have explored in this article serves as a powerful example of how to seamlessly integrate cutting-edge AI capabilities into existing business systems without the need for significant overhauls or disruptions. By leveraging the flexibility and power of modern AI technologies, we were able to enhance the functionality of a simulated supply chain management system while preserving its core operations and workflows. Throughout the development process, we placed a strong emphasis on minimizing the impact on the existing system architecture. Rather than attempting to replace or modify the established components, we introduced an “AI Integration System” that acts as a bridge between the existing infrastructure and the AI-powered capabilities. This approach allowed us to maintain the integrity of the existing systems while simultaneously leveraging the benefits of AI. One of the key advantages of this integration strategy is the ability to leverage the wealth of data already available within the existing systems. By accessing and processing this data through the AI models, we were able to generate more informed and intelligent responses to user queries, providing valuable insights and recommendations tailored to the specific supply chain activities and scenarios. As we look towards the future, the importance of seamlessly integrating AI into existing business ecosystems will only continue to grow. With the rapid pace of technological advancements and the increasing demand for intelligent automation and decision support, organizations that embrace this approach will be better positioned to capitalize on the opportunities presented by AI while minimizing disruptions to their operations. It is my hope that through this simulated real-world example, you have gained a deeper understanding of the potential for AI integration and the various strategies and best practices that can be employed to achieve successful implementation. By embracing this approach, businesses can unlock the transformative power of AI while preserving the investments and institutional knowledge embedded in their existing systems.

What if… Employers Employ AI Agents to Get 360° Feedback from Employees?
reddit
LLM Vibe Score0
Human Vibe Score0
AssistanceOk2217This week

What if… Employers Employ AI Agents to Get 360° Feedback from Employees?

AI Agent powered Comprehensive 360° Feedback Collection & Analysis Full Article &#x200B; https://i.redd.it/1ieczv6pud1d1.gif ⚪ What is this Article About? ● This article demonstrates how AI agents can be used in the real-world for gathering feedback from employees ● It explores using AI agents to collect insights on employee experiences, job satisfaction, and suggestions for improvement ● By leveraging AI agents and language models, organizations can better understand their workforce's needs and concerns ⚪Why Read this Article? ● Learn about the potential benefits of using AI agents for comprehensive feedback collection ● Understand how to build practical, real-world solutions by combining AI agents with other technologies ● Stay ahead of the curve by exploring cutting-edge applications of AI agents ⚪What are we doing in this Project? \> Part 1: AI Agents to Coordinate and Gather Feedback ● AI agents collaborate to collect comprehensive feedback from employees through surveys and interviews ● Includes a Feedback Collector Agent, Feedback Analyst Agent, and Feedback Reporter Agent \> Part 2: Analyze Feedback Data with Pandas AI and Llama3 ● Use Pandas AI and Llama3 language model to easily analyze the collected feedback data ● Extract insights, identify patterns, strengths, and areas for improvement from the feedback ⚪ Let's Design Our AI Agent System for 360° Feedback \> Feedback Collection System: ● Collect feedback from employees (simulated) ● Analyze the feedback data ● Report findings and recommendations \> Feedback Analysis System: ● Upload employee feedback CSV file ● Display uploaded data ● Perform natural language analysis and queries ● Generate automated insights and visual graphs ⚪ Let's get Cooking ● Explanation of the code for the AI agent system and feedback analysis system ● Includes code details for functions, classes, and streamlit interface ⚪ Closing Thoughts ● AI agents can revolutionize how businesses operate and tackle challenges ● Their ability to coordinate, collaborate, and perform specialized tasks is invaluable ● AI agents offer versatile and scalable solutions for optimizing processes and uncovering insights ⚪ Future Work ● This project is a demo to show the potential real-world use cases of AI Agents. To achieve the results seen here, I went through multiple iterations and changes. AI Agents are not fully ready yet (although they are making huge progress every day). AI Agents still need to go through an improvement cycle to reach their full potential in real-world settings. &#x200B;

Let’s Build Small AI Buzz, Offer ‘Claim Processing’ to Mid/Big Companies
reddit
LLM Vibe Score0
Human Vibe Score1
AssistanceOk2217This week

Let’s Build Small AI Buzz, Offer ‘Claim Processing’ to Mid/Big Companies

Discover How AI Can Transform Businesses, Every Details Spelled Out. Full Article https://preview.redd.it/jp0vc5g6e86d1.png?width=1421&format=png&auto=webp&s=efa43e2a9b04b6996b00adac4e4947a3b21c7e63 Artificial Intelligence (AI) is rapidly reshaping business landscapes, promising unprecedented efficiency and accuracy across industries. In this article, we delve into how Aniket Insurance Inc. (Imaginary) leverages AI to revolutionize its claim processing operations, offering insights into the transformative power of AI in modern business environments. ➡️ What’s This Article About? \* The article explores how Aniket Insurance Inc. uses AI to transform its claim processing. \* It details the three main workflows: User claim submission, Admin + AI claim processing, and Executive + AI claim analysis. https://preview.redd.it/ql0ec20ae86d1.png?width=769&format=png&auto=webp&s=4b6889dd85f848194d6adfc92c9c699138eb1fe7 ➡️ Why Read This Article \* Readers can see practical ways AI boosts efficiency in business, using Aniket Insurance as an example. \* AI speeds up routine tasks, like data entry, freeing up humans for more strategic work. It shows how AI-driven data analysis can lead to smarter business decisions. ➡️Let’s Design: Aniket Insurance Inc. has implemented AI architecture that encompasses three pivotal workflows: User Claim Submission Flow, Admin + AI Claim Processing Flow, and Executive + AI Claim Analysis Flow. Powered by AI models and integrated with store, this architecture ensures seamless automation and optimization of the entire claim processing lifecycle. By leveraging AI technologies like machine learning models and data visualization tools, Aniket Insurance how business can enhance operational efficiency, and strategic decision-making capabilities. https://preview.redd.it/qgdmzs3ee86d1.png?width=733&format=png&auto=webp&s=445295beb52a56d826e5527859cf62879116ddb0 ➡️Closing Thoughts: Looking ahead, the prospects of AI adoption across various industries are incredibly exciting. Imagine manufacturing plants where AI optimizes production lines, predicts maintenance needs, and ensures quality control. Envision healthcare facilities where AI assists in diagnosis, treatment planning, and drug discovery. Picture retail operations where AI personalizes product recommendations, streamlines inventory management, and enhances customer service. The possibilities are endless, as AI’s capabilities in pattern recognition, predictive modeling, and automation can be leveraged to tackle complex challenges and uncover valuable insights in virtually any domain. https://preview.redd.it/w3hr913ge86d1.png?width=754&format=png&auto=webp&s=d839a7703f5b28314a3278c8d628ae5f05d3668f

My Building Of Trading Order Management System Using AI Agents
reddit
LLM Vibe Score0
Human Vibe Score0
AniketWorkThis week

My Building Of Trading Order Management System Using AI Agents

Practical Guide : Automating Business Transactions with AI-Powered Workflows Full Article | Code https://preview.redd.it/hrkeo00yz4ie1.jpg?width=1911&format=pjpg&auto=webp&s=5bcb6f02c72bbce22fb691e4d8b799c414fed2a7 https://preview.redd.it/1cp0izzxz4ie1.png?width=1899&format=png&auto=webp&s=2598e25e17ab03a95f3009f5333f02b077ce30ca https://preview.redd.it/cjp1640yz4ie1.png?width=1899&format=png&auto=webp&s=13dad0ee8e0b1b22415a60a57b571058f0bdef33 TL;DR A practical implementation of an AI-powered B2B order management system using LangChain and LLM, demonstrating automated order processing, inventory management, and real-time communication between trading partners. https://i.redd.it/kxe4l69105ie1.gif Introduction In today’s fast-paced business environment, efficient order management is crucial for B2B operations. GlobalTrade Nexus AI showcases how artificial intelligence can streamline complex business transactions, reduce errors, and enhance communication between trading partners. What’s This Article About? This article presents a comprehensive B2B trading platform that leverages AI to automate order processing workflows. The system handles everything from order placement to fulfillment, featuring: Real-time inventory verification Automated shipping cost calculations Instant order validation Secure transaction processing Smart order cancellation capabilities State management across the entire order lifecycle The platform demonstrates how modern AI technologies can be integrated into traditional business processes to create a seamless, efficient trading environment. Tech stack Why Read It? As businesses increasingly embrace digital transformation, AI-powered solutions are becoming essential for maintaining competitive advantage. This article provides: A practical example of AI implementation in B2B commerce Insights into modern system architecture for business applications Real-world application of language models in business logic Demonstration of secure and scalable state management Blueprint for building similar AI-enhanced business systems Through our fictional companies’ implementation, readers can understand how AI can transform their business operations and prepare for the future of B2B commerce.

How I Built an Agentic Marketing Campaign Strategist
reddit
LLM Vibe Score0
Human Vibe Score1
AniketWorkThis week

How I Built an Agentic Marketing Campaign Strategist

Marketing at Scale: How One AI System Replaces Hundreds of Strategy Hours Article https://i.redd.it/uekqj3zmerme1.gif https://i.redd.it/30rk23zmerme1.gif https://preview.redd.it/fk1t53zmerme1.png?width=797&format=png&auto=webp&s=d07f473a9556fbd38885b3a2f862101d9b25424e https://preview.redd.it/n84113zmerme1.jpg?width=1914&format=pjpg&auto=webp&s=f42679269a1003e1c8d6501dd2d53e10db745bba https://preview.redd.it/l13ae3zmerme1.jpg?width=791&format=pjpg&auto=webp&s=ecab3c295c2a416bc0fed8c62fecbe3321e37093 TL;DR This article guides you through building an AI-powered marketing strategist using Python. It combines vector databases, language models, and PDF generation to create customized marketing strategies automatically. I’ll show you the complete system architecture, from storing marketing knowledge to generating professional strategy documents, with practical code examples you can implement today. Perfect for marketers and developers looking to leverage AI for business growth. Introduction Welcome to the exciting intersection of marketing and artificial intelligence! In today’s digital world, creating effective marketing campaigns requires deep expertise, market research, and creative thinking. But what if you could automate parts of this process? That’s exactly what I set out to build: an AI system that generates comprehensive marketing strategies tailored to specific products, audiences, and budgets. What’s This Article About? This article walks you through the creation of an AI-powered marketing strategist that combines the retrieval of relevant marketing knowledge with advanced language generation to produce detailed campaign strategies. The system I built uses Retrieval-Augmented Generation (RAG), which enhances AI outputs by grounding them in specific knowledge sources. Here’s how it works: You provide a simple campaign description (like “a new eco-friendly water bottle targeting millennials with a budget of $50,000”) The system searches a knowledge base of marketing principles and best practices It then uses a language model to craft a comprehensive strategy that includes campaign objectives, target audience analysis, channel selection, content ideas, budget allocation, and measurement KPIs Finally, it generates a professional PDF document with your complete marketing strategy The beauty of this approach is that it combines the creativity and adaptability of AI with established marketing frameworks, ensuring the strategies are both innovative and grounded in proven principles. Why Read It? AI is rapidly transforming how businesses operate, and marketing is at the forefront of this revolution. According to recent studies, companies that effectively leverage AI in their marketing efforts see significant improvements in customer engagement, conversion rates, and ROI. Even if you’re not building a system for a real company right now, understanding how to implement AI in marketing processes gives you valuable skills and insights. This article provides a practical example of how AI can: Save marketers countless hours of research and strategy development Ensure consistency in marketing approaches across different campaigns Generate creative ideas that might not have been considered otherwise Scale marketing expertise across an organization By following along, you’ll gain hands-on experience with technologies like vector databases, language models, and automated document generation — all skills that are increasingly valuable in today’s business environment.

How I Built A Simple ‘BPO’ Company, All AI Employees (All Local)
reddit
LLM Vibe Score0
Human Vibe Score1
AssistanceOk2217This week

How I Built A Simple ‘BPO’ Company, All AI Employees (All Local)

Disrupting the BPO Industry: My Journey Building a Fully Automated Company with AI Employees Full Article : https://medium.com/@learn-simplified/how-i-built-a-simple-bpo-company-all-ai-employees-all-local-631e48fa908a &#x200B; https://preview.redd.it/htjo1mancl2d1.png?width=1586&format=png&auto=webp&s=7e77f4c66e5ca55a8b0ea6969c43a458503ad921 ● What Are We Doing Today? We are building a BPO (Business Process Outsourcing) call center for an imaginary electric company called "Aniket Very General Electric Company". We will create different departments staffed by AI agents who can chat (and eventually speak in next part) with customers to answer questions, handle complaints, or provide services. ● Why Should You Read This Article? Learning how to build AI agents that can do tasks in real setting, co ordinate w/ human, AI, providing technical support will be a highly valuable skill. ● How Are We Going to Build Our All AI Employees Company? ○ We will explain what BPO and call centers are. ○ Our AI company will have departments like Customer Service, Tech Support, Billing & Payments, Outage Management, and Onboarding Customers. ○ We will use Docker containers to run the Dify AI platform as the base. ○ The AI agents will use the LLaMA-3 language model from Meta AI. ○ We may use Groq's AI accelerator chip to make LLaMA-3 faster. ○ Each department will have a knowledge base of text files that the AI agents can reference. ● Let's Get Cooking! This section provides setup instructions for installing Docker, Ollama (for running LLaMA-3), and the Dify AI platform. It also outlines the different AI agents we will create for departments like Reception, Customer Service, Billing, Tech Support, etc. ● Let's Design our Organization ○ We explain how each department's AI agents will have their own knowledge base, like an employee handbook. ○ The knowledge bases will contain policies, procedures, and other key information. ○ The AI agents can quickly reference this information to provide accurate and knowledgeable responses. ● Let's Meet Our AI Employees ○ We chose the LLaMA-3 70B model as the base for all AI agents across departments. ○ We give the AI agents customized prompts to define their personalities and roles. ○ The knowledge bases act as training materials tailored to each department. ○ In the future, AI agents could have additional tools like ticket systems and integrations. ● Let's Run Our BPO Organization Now that the AI workforce and knowledge bases are ready, we can open our BPO company and have the AI agents start handling customer inquiries across different departments like billing, tech support, outages, and new connections. ● Debugging This section highlights the importance of debugging, showing traces of how the language model understands customer queries and retrieves relevant context from knowledge bases to provide good responses. ● Future Work ○ Scale up to handle more customers using cloud services or distributed computing. ○ Move AI agents and knowledge bases to the cloud for accessibility and maintenance. ○ Fine-tune language models for better performance in each department. ○ Use scalable vector databases for faster knowledge retrieval. ○ Enable voice interfaces and computer vision for more natural interactions. ○ Implement continuous learning so AI agents can expand their knowledge over time. The article demonstrates the potential of building an actual AI-powered company and raises thought-provoking questions about the role of humans, ethics, and using AI to create a better world. &#x200B;

Let’s Build One Person Business Using 100% AI
reddit
LLM Vibe Score0
Human Vibe Score1
AssistanceOk2217This week

Let’s Build One Person Business Using 100% AI

AI made it possible for 9-to-5 workers to start a one-person business without quitting their jobs. Full Article https://preview.redd.it/tynb9y6z695d1.png?width=1309&format=png&auto=webp&s=b490d3676a63adcc01faff8c476056cb7d420022 https://i.redd.it/9x3okti0795d1.gif The Opportunities for Starting a Business ○ There are huge opportunities to start your own business by leveraging valuable skills to attract paying audiences. ○ New software and AI platforms make it easier to distribute products/services and automate tasks that were previously time-consuming. Our One Person Book Publication House ○ This article explores building a one-person AI-powered business focused on publishing books. ○ Users input data on a topic, and AI generates a comprehensive book structure and content based on that. ○ The generated content can be formatted, designed, and published digitally or in print easily. Why Read This Article? ○ It presents an innovative AI-powered approach to streamline the book publishing process. ○ It provides technical implementation details using LLM, Python and the Streamlit library as a reference. ○ It highlights AI's potential in automating creative tasks like writing and content creation. Approaching the One Person Business ○ Reflect on areas where you overcame personal struggles and gained valuable skills. ○ Leverage that expertise to build an AI business serving others facing similar obstacles. ○ Use AI tools to create content, automate processes, and efficiently scale your offerings. The Publication Business Idea ○ Focus on writing and publishing small books using AI writing assistants. ○ AI can streamline research, writing drafts, outlines, and ideas across genres. ○ Concentrate efforts on editing, formatting, and marketing while AI handles writing. The Book Generation Process ○ Users input structured topic data like outlines, key points, and references. ○ Advanced AI language models generate flowing book content from that data. ○ Minimal human effort is needed beyond initial inputs and refinement. ○ AI systems automatically handle formatting, design, and publishing. Technical Implementation ○ Includes a Book class to represent a book's hierarchical structure in Python. ○ Functions to generate book structures and section content using AI models. ○ Integrates with a Streamlit app for user input and output. ○ Allows downloading the final book in Markdown format. Closing Thoughts ○ This AI-powered approach makes book writing and publishing more accessible to individuals. ○ AI handles the heavy lifting, with humans providing quality control through editing. ○ It opens up possibilities for innovative knowledge sharing as technology evolves.

My Building Of Trading Order Management System Using AI Agents
reddit
LLM Vibe Score0
Human Vibe Score0
AniketWorkThis week

My Building Of Trading Order Management System Using AI Agents

Practical Guide : Automating Business Transactions with AI-Powered Workflows Full Article | Code https://preview.redd.it/hrkeo00yz4ie1.jpg?width=1911&format=pjpg&auto=webp&s=5bcb6f02c72bbce22fb691e4d8b799c414fed2a7 https://preview.redd.it/1cp0izzxz4ie1.png?width=1899&format=png&auto=webp&s=2598e25e17ab03a95f3009f5333f02b077ce30ca https://preview.redd.it/cjp1640yz4ie1.png?width=1899&format=png&auto=webp&s=13dad0ee8e0b1b22415a60a57b571058f0bdef33 TL;DR A practical implementation of an AI-powered B2B order management system using LangChain and LLM, demonstrating automated order processing, inventory management, and real-time communication between trading partners. https://i.redd.it/kxe4l69105ie1.gif Introduction In today’s fast-paced business environment, efficient order management is crucial for B2B operations. GlobalTrade Nexus AI showcases how artificial intelligence can streamline complex business transactions, reduce errors, and enhance communication between trading partners. What’s This Article About? This article presents a comprehensive B2B trading platform that leverages AI to automate order processing workflows. The system handles everything from order placement to fulfillment, featuring: Real-time inventory verification Automated shipping cost calculations Instant order validation Secure transaction processing Smart order cancellation capabilities State management across the entire order lifecycle The platform demonstrates how modern AI technologies can be integrated into traditional business processes to create a seamless, efficient trading environment. Tech stack Why Read It? As businesses increasingly embrace digital transformation, AI-powered solutions are becoming essential for maintaining competitive advantage. This article provides: A practical example of AI implementation in B2B commerce Insights into modern system architecture for business applications Real-world application of language models in business logic Demonstration of secure and scalable state management Blueprint for building similar AI-enhanced business systems Through our fictional companies’ implementation, readers can understand how AI can transform their business operations and prepare for the future of B2B commerce.

Browser Agents Real Example
reddit
LLM Vibe Score0
Human Vibe Score1
No_Information6299This week

Browser Agents Real Example

I made a Browser Price Matching Tool that uses browser automation and some clever skills to adjust your product prices based on real-time web searches data. If you're into scraping, automation, or just love playing with the latest in ML-powered tools like OpenAI's GPT-4, this one's for you. What My Project Does The tool takes your current product prices (think CSV) and finds similar products online (targeting Amazon for demo purposes). It then compares prices, allowing you to adjust your prices competitively. The magic happens in a multi-step pipeline: Generate Clean Search Queries: Uses a learned skill to convert messy product names (like "Apple iPhone14!<" or "Dyson! V11!!// VacuumCleaner") into clean, Google-like search queries. Browser Data Extraction: Launches asynchronous browser agents (leveraging Playwright) to search for those queries on Amazon, retrieves the relevant data, and scrapes the page text. Parse & Structure Results: Another custom skill parses the browser output to output structured info: product name, price, and a short description. Enrich Your Data: Finally, the tool combines everything to enrich your original data with live market insights! Full code link: Full code File Rundown learn\skill.py Learns how to generate polished search queries from your product names with GPT-4o-mini. It outputs a JSON file: makequery.json. learn\skill\select\best\product.py Trains another skill to parse web-scraped data and select the best matching product details. Outputs select_product.json. make\query.json The skill definition file for generating search queries (produced by learnskill.py). select\product.json The skill definition file for extracting product details from scraped results (produced by learnskillselectbest_product.py). product\price\matching.py The main pipeline script that orchestrates the entire process—from loading product data, running browser agents, to enriching your CSV. Setup & Installation Install Dependencies: pip install python-dotenv openai langchain\_openai flashlearn requests pytest-playwright Install Playwright Browsers: playwright install Configure OpenAI API: Create a .env file in your project directory with:OPENAI\API\KEY="sk-your\api\key\_here" Running the Tool Train the Query Skill: Run learnskill.py to generate makequery.json. Train the Product Extraction Skill: Run learnskillselectbestproduct.py to generate select_product.json. Execute the Pipeline: Kick off the whole process by running productpricematching.py. The script will load your product data (sample data is included for demo, but easy to swap with your CSV), generate search queries, run browser agents asynchronously, scrape and parse the data, then output the enriched product listings. Target Audience I built this project to automate price matching—a huge pain point for anyone running an e-commerce business. The idea was to minimize the manual labor of checking competitor prices while integrating up-to-date market insights. Plus, it was a fun way to combine automation,skill training, and browser automation! Customization Tweak the concurrency in productpricematching.py to manage browser agent load. Replace the sample product list with your own CSV for a real-world scenario. Extend the skills if you need more data points or different parsing logic. Ajudst skill definitions as needed Comparison With existing approaches you need to manually write parsing loginc and data transformation logic - here ai does it for you. If you like the tutorial - leave a star github

How I Built an Agentic Marketing Campaign Strategist
reddit
LLM Vibe Score0
Human Vibe Score1
AniketWorkThis week

How I Built an Agentic Marketing Campaign Strategist

Marketing at Scale: How One AI System Replaces Hundreds of Strategy Hours Article https://i.redd.it/uekqj3zmerme1.gif https://i.redd.it/30rk23zmerme1.gif https://preview.redd.it/fk1t53zmerme1.png?width=797&format=png&auto=webp&s=d07f473a9556fbd38885b3a2f862101d9b25424e https://preview.redd.it/n84113zmerme1.jpg?width=1914&format=pjpg&auto=webp&s=f42679269a1003e1c8d6501dd2d53e10db745bba https://preview.redd.it/l13ae3zmerme1.jpg?width=791&format=pjpg&auto=webp&s=ecab3c295c2a416bc0fed8c62fecbe3321e37093 TL;DR This article guides you through building an AI-powered marketing strategist using Python. It combines vector databases, language models, and PDF generation to create customized marketing strategies automatically. I’ll show you the complete system architecture, from storing marketing knowledge to generating professional strategy documents, with practical code examples you can implement today. Perfect for marketers and developers looking to leverage AI for business growth. Introduction Welcome to the exciting intersection of marketing and artificial intelligence! In today’s digital world, creating effective marketing campaigns requires deep expertise, market research, and creative thinking. But what if you could automate parts of this process? That’s exactly what I set out to build: an AI system that generates comprehensive marketing strategies tailored to specific products, audiences, and budgets. What’s This Article About? This article walks you through the creation of an AI-powered marketing strategist that combines the retrieval of relevant marketing knowledge with advanced language generation to produce detailed campaign strategies. The system I built uses Retrieval-Augmented Generation (RAG), which enhances AI outputs by grounding them in specific knowledge sources. Here’s how it works: You provide a simple campaign description (like “a new eco-friendly water bottle targeting millennials with a budget of $50,000”) The system searches a knowledge base of marketing principles and best practices It then uses a language model to craft a comprehensive strategy that includes campaign objectives, target audience analysis, channel selection, content ideas, budget allocation, and measurement KPIs Finally, it generates a professional PDF document with your complete marketing strategy The beauty of this approach is that it combines the creativity and adaptability of AI with established marketing frameworks, ensuring the strategies are both innovative and grounded in proven principles. Why Read It? AI is rapidly transforming how businesses operate, and marketing is at the forefront of this revolution. According to recent studies, companies that effectively leverage AI in their marketing efforts see significant improvements in customer engagement, conversion rates, and ROI. Even if you’re not building a system for a real company right now, understanding how to implement AI in marketing processes gives you valuable skills and insights. This article provides a practical example of how AI can: Save marketers countless hours of research and strategy development Ensure consistency in marketing approaches across different campaigns Generate creative ideas that might not have been considered otherwise Scale marketing expertise across an organization By following along, you’ll gain hands-on experience with technologies like vector databases, language models, and automated document generation — all skills that are increasingly valuable in today’s business environment.

I'm Building an "AiExecutiveSuperAgent_Systems_Interface" between humanity and the Ai world, as well as each other... Let's Talk?
reddit
LLM Vibe Score0
Human Vibe Score1
Prudent_Ad_3114This week

I'm Building an "AiExecutiveSuperAgent_Systems_Interface" between humanity and the Ai world, as well as each other... Let's Talk?

Ok... So look... This one is pretty crazy... I'm building an Ai Interface that knows me better than I know myself - Check, lots of people have this, either in reality with employees and family members, or with ai intelligence. But it doesn't just know Me... It knows how to talk with Me. It understands my language, because I've trained it to. I've also trained it to translate that to all my clients and HumanAgents, soon to become RobotAgents... The RESULT: I can literally just spend 1-18 hours talking to it, and things get DONE. Most of that time, I just say EXECUTE, or ENGAGE, or DRAFT, or DISPATCH. I feel like a secret agent communicating in codes with his agency 😂 Not great for the paranoiac in me, but it's easy to get that part under control, ya'll. It's like having a team of 10,000 people, all available 24/7, all perfectly synchronised to each other's communication styles, preferences and ultimately: WHAT DO YOU NEED ME TO DO. At the end of the it all, having run my single COMMAND through a thousand of those people, a Document is prepared that outlines the next 3 stages of the plan, along with instructions to the whole team for how to ENACT it. Sounds rather grand and wonderful... Even when I simply use it to help me come up with a filing system for my creative work... \\\\\\\\\\\\\\\\\\\\\\ Here's my current VISION, why I'm doing this AND why I'm doing it publicly despite it being top secret. VISION To create an army of User-Owned and Operated "AiSuperAgencies" which gather intelligence on the user, securely file and analyse it, and then construct a sub-army of agents and tools that work together to produce the desired output, for any Function in the Personal and Professional Lives of EVERYONE, EVERYWHERE, in 3-5 Years. To start, I'm building it for me and the 5-10 cleaners who've made it to Level 1 in my access system. They were sick of toxic employers, tyrannical agencies and greedy customers. They gathered around us (many came in, many went out, few stayed, took about a year for our core team of 3 Level 2 Cleaners. My goal has always been to never employ anyone. Just me, my Partner and the Cleaners. All Shared Owners in the system for delivering the right cleaner to the right house in our town, at the right time and without any dramas or arguments... I have a personal talent for resolving disputes, which has made working for and buying from my business a mostly enjoyable and upbeat experience, with a touch of mystery and a feeling that you're part of something big! It is a business that ran on Me. I put in my time, every day, building automated tool after automated tool. Hiring a contractor to do a job, scratching my head when it didn't add enough value to pay for itself, then just doing it myself again. I wanted to solve that problem. I'm trusting that the few who hear about it who actually see the potential, will just come join us, no dramas, just cool people partnering up! And those that don't, won't. No one could steal it, because it's Mine, and I'll just change the keys anyway loser! Enjoy digging through my past, you lunatic! I'm out here living Now. Anyways... It's lonely around here. I have a cleaning business that I run from my laptop, which means I can live anywhere, but I still had this big problem of time... NOT ENOUGH Oh Wait. It's Here.

How I landed an internship in AI
reddit
LLM Vibe Score0
Human Vibe Score1
Any-Reserve-4403This week

How I landed an internship in AI

For motivational purposes only! I see a lot of posts on here from people without “traditional” machine learning, data science, etc.. backgrounds asking how they can break into the field, so I wanted to share my experience. EDIT Learning Resources and Side Project Ideas * My background: I graduated from a decent undergraduate school with a degree in Political Science several years ago. Following school I worked in both a client services role at a market research company and an account management role at a pretty notable fintech start-up. Both of these roles exposed me to ML, AI and more sophisticated software concepts in general, and I didn’t really care for the sales side of things, so I decided to make an attempt at switching careers into something more technical. While working full time I began taking night classes at a local community college, starting with pre calculus all the way up to Calc 2 and eventually more advanced classes like linear algebra and applied probability. I also took some programming courses including DSA. I took these classes for about two years while working, and on the side had been working through various ML books and videos on YouTube. What worked the best for me was Hands-on Machine Learning with Scikit Learn, Keara’s and Tensorflow. I eventually had enough credits where I was able to begin applying to MS in Data Science programs and was fortunate enough to get accepted into one and also get a position in their Robotics Lab doing Computer Vision work. When it came time to apply for internships, it was a BLOODBATH. I must have applied to over 100 roles with my only responses being video interviews and OA’s. Finally I got an interview for an AI Model Validation internship with a large insurance company and after completing the interviews was told I performed well but they were still interviewing several candidates. I ended up getting the offer and accepting the role where I’ll be working on a Computer Vision model and some LLM related tasks this summer and could not be more fortunate / excited. A couple things stood out to them during the interview process. 1, the fact that I was working and taking night classes with the intent to break into the field. It showed a genuine passion as opposed to someone who watched a YouTube video and claims they are now an expert. 2, side projects. I not only had several projects, but I had some that were relevant to the work I’d be doing this summer from the computer vision standpoint. 3, business sense. I emphasized during my interviews how working in a business role prior to beginning my masters would give me a leg up as intern because I would be able to apply the work of a data scientist to solving actual business challenges. For those of you trying to break into the field, keep pushing, keep building, and focus on what makes you unique and able to help a company! Please feel free to contact me if you would like any tips I can share, examples of projects, or anything that would be helpful to your journey.

How to Start Research in Computer Science & AI in 2025 – A Modernized Framework
reddit
LLM Vibe Score0
Human Vibe Score1
somdipdeyThis week

How to Start Research in Computer Science & AI in 2025 – A Modernized Framework

Over a decade ago, I wrote two articles: "A Beginner’s Guide to Computer Science Research" and "How to Start a Research Work in Computer Science: A Framework for Beginners" \- that have been used at several universities around the world for the same purpose. These articles aimed to help students and early-career researchers navigate the complexities of academic research in computer science. However, since 2014, the research landscape has changed dramatically with the rise of AI, automation, and powerful collaborative tools. Now, in 2025, starting research in computer science and AI is more accessible than ever. With AI-powered research assistants, open-access repositories, and real-time collaborative platforms, researchers can work more efficiently and focus on innovation. I recently published an updated guide in The Times of India, presenting a modern “Eight-Step Approach to Research” framework that integrates the latest methodologies and tools for AI and CS research. This framework is designed to help students and researchers independently explore their chosen topics while leveraging cutting-edge technology. If you’re curious about how to streamline your research workflow, enhance your literature review process, and effectively collaborate in the AI research space, check out the article here: 🔗 How to Start a Research Work in Computer Science and AI in 2025 – An Updated Framework Block Diagram of “Eight-Step Approach to Research” in 2025 Would love to hear thoughts from the ML research community—what tools and techniques do you use to make research more efficient in 2025? Let’s discuss! 🚀

MarkDrop
reddit
LLM Vibe Score0
Human Vibe Score1
Willing-Ear-8271This week

MarkDrop

I’m excited to share my Python package, Markdrop, which has hit 5.01k+ downloads in just a month, so updated it just now! 🚀 It’s a powerful tool for converting PDF documents into structured formats like Markdown (.md) and HTML (.html) while automatically processing images and tables into descriptions for downstream use. Here's what Markdrop does: Key Features: PDF to Markdown/HTML Conversion: Converts PDFs into clean, structured Markdown files (.md) or HTML outputs, preserving the content layout. AI-Powered Descriptions: Replaces tables and images with descriptive summaries generated by LLM, making the content fully textual and easy to analyze. Earlier I added support of 6 different LLM Clients, but to improve the inference time, now this supports only GEMINI\API\KEY and OPENAI\API\KEY. Downloadable Tables: Can add accurate download buttons in HTML for tables, allowing users to download them as Excel files. Seamless Table and Image Handling: Extracts tables and images, generating detailed summaries for each, which are then embedded into the final Markdown document. At the end, one can have a .md file that contains only textual data, including the AI-generated summaries of tables, images, graphs, etc. This results in a highly portable format that can be used directly for several downstream tasks, such as: Can be directly integrated into a RAG pipeline for enhanced content understanding and querying on documents containg useful images and tabular data. Ideal for automated content summarization and report generation. Facilitates extracting key data points from tables and images for further analysis. The .md files can serve as input for machine learning tasks or data-driven projects. Ideal for data extraction, simplifying the task of gathering key data from tables and images. The downloadable table feature is perfect for analysts, reducing the manual task of copying tables into Excel. Markdrop streamlines workflows for document processing, saving time and enhancing productivity. You can easily install it via: pip install markdrop There’s also a Colab demo available to try it out directly: Open in Colab. Github Repo If you've used Markdrop or plan to, I’d love to hear your feedback! Share your experience, any improvements, or how it helped in your workflow. Check it out on PyPI and let me know your thoughts!

Backend dev wants to learn ML
reddit
LLM Vibe Score0
Human Vibe Score1
chipmuxThis week

Backend dev wants to learn ML

Hello ML Experts, I am staff engineer, working in a product based organization, handling the backend services. I see myself becoming Solution Architect and then Enterprise Architect one day. With the AI and ML trending now a days, So i feel ML should be an additional skill that i should acquire which can help me leading and architecting providing solutions to the problems more efficiently, I think however it might not replace the traditional SWEs working on backend APIs completely, but ML will be just an additional diamention similar to the knowledge of Cloud services and DevOps. So i would like to acquire ML knowledge, I dont have any plans to be an expert at it right now, nor i want to become a full time data scientist or ML engineer as of today. But who knows i might diverge, but thats not the plan currently. I did some quick promting with ChatGPT and was able to comeup with below learning path for me. So i would appreciate if some of you ML experts can take a look at below learning path and provide your suggestions 📌 PHASE 1: Core AI/ML & Python for AI (3-4 Months) Goal: Build a solid foundation in AI/ML with Python, focusing on practical applications. 1️⃣ Python for AI/ML (2-3 Weeks) Course: [Python for Data Science and Machine Learning Bootcamp]() (Udemy) Topics: Python, Pandas, NumPy, Matplotlib, Scikit-learn basics 2️⃣ Machine Learning Fundamentals (4-6 Weeks) Course: Machine Learning Specialization by Andrew Ng (C0ursera) Topics: Linear & logistic regression, decision trees, SVMs, overfitting, feature engineering Project: Build an ML model using Scikit-learn (e.g., predicting house prices) 3️⃣ Deep Learning & AI Basics (4-6 Weeks) Course: Deep Learning Specialization by Andrew Ng (C0ursera) Topics: Neural networks, CNNs, RNNs, transformers, generative AI (GPT, Stable Diffusion) Project: Train an image classifier using TensorFlow/Keras 📌 PHASE 2: AI/ML for Enterprise & Cloud Applications (3-4 Months) Goal: Learn how AI is integrated into cloud applications & enterprise solutions. 4️⃣ AI/ML Deployment & MLOps (4 Weeks) Course: MLOps Specialization by Andrew Ng (C0ursera) Topics: Model deployment, monitoring, CI/CD for ML, MLflow, TensorFlow Serving Project: Deploy an ML model as an API using FastAPI & Docker 5️⃣ AI/ML in Cloud (Azure, AWS, OpenAI APIs) (4-6 Weeks) Azure AI Services: Course: Microsoft AI Fundamentals (C0ursera) Topics: Azure ML, Azure OpenAI API, Cognitive Services AWS AI Services: Course: [AWS Certified Machine Learning – Specialty]() (Udemy) Topics: AWS Sagemaker, AI workflows, AutoML 📌 PHASE 3: AI Applications in Software Development & Future Trends (Ongoing Learning) Goal: Explore AI-powered tools & future-ready AI applications. 6️⃣ Generative AI & LLMs (ChatGPT, GPT-4, LangChain, RAG, Vector DBs) (4 Weeks) Course: [ChatGPT Prompt Engineering for Developers]() (DeepLearning.AI) Topics: LangChain, fine-tuning, RAG (Retrieval-Augmented Generation) Project: Build an LLM-based chatbot with Pinecone + OpenAI API 7️⃣ AI-Powered Search & Recommendations (Semantic Search, Personalization) (4 Weeks) Course: [Building Recommendation Systems with Python]() (Udemy) Topics: Collaborative filtering, knowledge graphs, AI search 8️⃣ AI-Driven Software Development (Copilot, AI Code Generation, Security) (Ongoing) Course: AI-Powered Software Engineering (C0ursera) Topics: AI code completion, AI-powered security scanning 🚀 Final Step: Hands-on Projects & Portfolio Once comfortable, work on real-world AI projects: AI-powered document processing (OCR + LLM) AI-enhanced search (Vector Databases) Automated ML pipelines with MLOps Enterprise AI Chatbot using LLMs ⏳ Suggested Timeline 📅 6-9 Months Total (10-12 hours/week) 1️⃣ Core ML & Python (3-4 months) 2️⃣ Enterprise AI/ML & Cloud (3-4 months) 3️⃣ AI Future Trends & Applications (Ongoing) Would you like a customized plan with weekly breakdowns? 🚀

Browser Agents Real Example
reddit
LLM Vibe Score0
Human Vibe Score1
No_Information6299This week

Browser Agents Real Example

I made a Browser Price Matching Tool that uses browser automation and some clever skills to adjust your product prices based on real-time web searches data. If you're into scraping, automation, or just love playing with the latest in ML-powered tools like OpenAI's GPT-4, this one's for you. What My Project Does The tool takes your current product prices (think CSV) and finds similar products online (targeting Amazon for demo purposes). It then compares prices, allowing you to adjust your prices competitively. The magic happens in a multi-step pipeline: Generate Clean Search Queries: Uses a learned skill to convert messy product names (like "Apple iPhone14!<" or "Dyson! V11!!// VacuumCleaner") into clean, Google-like search queries. Browser Data Extraction: Launches asynchronous browser agents (leveraging Playwright) to search for those queries on Amazon, retrieves the relevant data, and scrapes the page text. Parse & Structure Results: Another custom skill parses the browser output to output structured info: product name, price, and a short description. Enrich Your Data: Finally, the tool combines everything to enrich your original data with live market insights! Full code link: Full code File Rundown learn\skill.py Learns how to generate polished search queries from your product names with GPT-4o-mini. It outputs a JSON file: makequery.json. learn\skill\select\best\product.py Trains another skill to parse web-scraped data and select the best matching product details. Outputs select_product.json. make\query.json The skill definition file for generating search queries (produced by learnskill.py). select\product.json The skill definition file for extracting product details from scraped results (produced by learnskillselectbest_product.py). product\price\matching.py The main pipeline script that orchestrates the entire process—from loading product data, running browser agents, to enriching your CSV. Setup & Installation Install Dependencies: pip install python-dotenv openai langchain\_openai flashlearn requests pytest-playwright Install Playwright Browsers: playwright install Configure OpenAI API: Create a .env file in your project directory with:OPENAI\API\KEY="sk-your\api\key\_here" Running the Tool Train the Query Skill: Run learnskill.py to generate makequery.json. Train the Product Extraction Skill: Run learnskillselectbestproduct.py to generate select_product.json. Execute the Pipeline: Kick off the whole process by running productpricematching.py. The script will load your product data (sample data is included for demo, but easy to swap with your CSV), generate search queries, run browser agents asynchronously, scrape and parse the data, then output the enriched product listings. Target Audience I built this project to automate price matching—a huge pain point for anyone running an e-commerce business. The idea was to minimize the manual labor of checking competitor prices while integrating up-to-date market insights. Plus, it was a fun way to combine automation,skill training, and browser automation! Customization Tweak the concurrency in productpricematching.py to manage browser agent load. Replace the sample product list with your own CSV for a real-world scenario. Extend the skills if you need more data points or different parsing logic. Ajudst skill definitions as needed Comparison With existing approaches you need to manually write parsing loginc and data transformation logic - here ai does it for you. If you like the tutorial - leave a star github

I'm Building an "AiExecutiveSuperAgent_Systems_Interface" between humanity and the Ai world, as well as each other... Let's Talk?
reddit
LLM Vibe Score0
Human Vibe Score1
Prudent_Ad_3114This week

I'm Building an "AiExecutiveSuperAgent_Systems_Interface" between humanity and the Ai world, as well as each other... Let's Talk?

Ok... So look... This one is pretty crazy... I'm building an Ai Interface that knows me better than I know myself - Check, lots of people have this, either in reality with employees and family members, or with ai intelligence. But it doesn't just know Me... It knows how to talk with Me. It understands my language, because I've trained it to. I've also trained it to translate that to all my clients and HumanAgents, soon to become RobotAgents... The RESULT: I can literally just spend 1-18 hours talking to it, and things get DONE. Most of that time, I just say EXECUTE, or ENGAGE, or DRAFT, or DISPATCH. I feel like a secret agent communicating in codes with his agency 😂 Not great for the paranoiac in me, but it's easy to get that part under control, ya'll. It's like having a team of 10,000 people, all available 24/7, all perfectly synchronised to each other's communication styles, preferences and ultimately: WHAT DO YOU NEED ME TO DO. At the end of the it all, having run my single COMMAND through a thousand of those people, a Document is prepared that outlines the next 3 stages of the plan, along with instructions to the whole team for how to ENACT it. Sounds rather grand and wonderful... Even when I simply use it to help me come up with a filing system for my creative work... \\\\\\\\\\\\\\\\\\\\\\ Here's my current VISION, why I'm doing this AND why I'm doing it publicly despite it being top secret. VISION To create an army of User-Owned and Operated "AiSuperAgencies" which gather intelligence on the user, securely file and analyse it, and then construct a sub-army of agents and tools that work together to produce the desired output, for any Function in the Personal and Professional Lives of EVERYONE, EVERYWHERE, in 3-5 Years. To start, I'm building it for me and the 5-10 cleaners who've made it to Level 1 in my access system. They were sick of toxic employers, tyrannical agencies and greedy customers. They gathered around us (many came in, many went out, few stayed, took about a year for our core team of 3 Level 2 Cleaners. My goal has always been to never employ anyone. Just me, my Partner and the Cleaners. All Shared Owners in the system for delivering the right cleaner to the right house in our town, at the right time and without any dramas or arguments... I have a personal talent for resolving disputes, which has made working for and buying from my business a mostly enjoyable and upbeat experience, with a touch of mystery and a feeling that you're part of something big! It is a business that ran on Me. I put in my time, every day, building automated tool after automated tool. Hiring a contractor to do a job, scratching my head when it didn't add enough value to pay for itself, then just doing it myself again. I wanted to solve that problem. I'm trusting that the few who hear about it who actually see the potential, will just come join us, no dramas, just cool people partnering up! And those that don't, won't. No one could steal it, because it's Mine, and I'll just change the keys anyway loser! Enjoy digging through my past, you lunatic! I'm out here living Now. Anyways... It's lonely around here. I have a cleaning business that I run from my laptop, which means I can live anywhere, but I still had this big problem of time... NOT ENOUGH Oh Wait. It's Here.

Built a Free AI Fitness Planner - From Passion to Product with No Traditional Coding
reddit
LLM Vibe Score0
Human Vibe Score1
jhojnac2This week

Built a Free AI Fitness Planner - From Passion to Product with No Traditional Coding

I posted this in r/entrepreneur as well but figured this is a great place too. I am looking to get your thoughts on this project and maybe some ideas as well. I wanted to share my journey of creating a free ai-powered workout planning tool with bolt. new and very minimal coding skills. It has taken me probably 4 days in total to complete and get to a point I am happy with. Many improvements coming but want to get it out there for some feedback and testing. I have been going to the gym for years and at this point my routines have gotten stale. I end up doing the same sets of exercises and repetitions over and over. I figured why not let chat gpt or some AI software help me develop or at least recommend different exercises. I was then was recommended youtube videos on creating your own web application without any coding. I will say it does take some coding knowledge, not that I am editing it myself, but I know what its trying to do and can prompt it correctly. I am still struggling with some things like integrating stripe for subscriptions so I only have it set up for donations currently. I dont mind it being free as I would like everyone the opportunity to help develop their own workouts. current cost breakdown to create: bolt. new credits - $100/month (gonna drop to the $20 now that its complete) supabase database - $35/month netlify domain - $11.99/year If anyone is interested or has questions feel free to let me know. It is called fitfocuscalendar. com this can all be done even cheaper using their free options but might take a lot more time depending on the complexity of the application as there are not a lot of free credits to code with each month and the supabase free database plan it pretty limited on size. title was AI generated.

Building a No-Code AI Customer Service Tool While Working 9-5 | All real - No BS
reddit
LLM Vibe Score0
Human Vibe Score1
Content_Limit_9723This week

Building a No-Code AI Customer Service Tool While Working 9-5 | All real - No BS

I want to share my journey of building Chaterimo, my first revenue-generating side project that I've been working on for the past 1.5 years alongside my day job. What started as a solution to make AI chatbots more accessible has grown to over 300 signups, 30 paying customers, and 50,000+ customer queries handled. The Problem I Wanted to Solve: It started with my father's business struggling with customer service - hiring staff was expensive and they would eventually leave, creating a constant cycle of training new people. I decided to help by building a livechat chatbot powered by AI to handle customer queries. The first version was basic (running on ChatGPT-3 with 4k tokens), but it worked! Seeing its success at my father's business, I realized this could help many other businesses too. As I kept improving it and adding features, I expanded to focus on e-commerce stores facing similar challenges. What Makes Chaterimo Different: True no-code setup: Install and run in seconds Choice of AI Models: ChatGPT by default, with options for Claude and the latest Gemini Flexible API Integration: Bring your own API keys for cheaper, unlimited messaging Smart Context Understanding: Can search Google or scan the current webpage to provide relevant answers Lead Generation: Capture and manage potential customer information Rich Integrations: Works with Shopify, Facebook Messenger, and Make for automation Customizable Bot Personality: Edit your chatbot's role and behavior through system prompts The Journey: This is my first side project that's actually generating revenue ($500+ MRR), unlike my previous "just for fun" projects. The past 1.5 years have been a learning experience, balancing development with a full-time job. What started as a simple idea has evolved based on real user feedback and needs. Current Metrics: 300+ total signups 30 paying customers 50,000+ customer queries successfully handled by AI $500+ monthly recurring revenue All while maintaining a 9-5 job Some Things I've Learned: Focus on making things simpler, not adding more features Listen to users - they'll tell you what they really need Flexibility matters - letting users use their own API keys was a game-changer Building something you believe in makes all the difference I'm still actively improving Chaterimo based on feedback. If you're running a website or e-commerce store and want to try it out, I'd love to hear your thoughts. What's Next: I'm focused on making the onboarding even smoother and adding more customization options while keeping the core simplicity that makes Chaterimo work. Would love to hear your thoughts or answer any questions! Has anyone else built successful side projects while working full-time? What were your biggest learnings?

Built a multi-agent AI mental health assistant (7 agents, backend automated, no-code stack)
reddit
LLM Vibe Score0
Human Vibe Score1
CapitalCategory4044This week

Built a multi-agent AI mental health assistant (7 agents, backend automated, no-code stack)

Been working on this little side project and finally got it to a working version. It’s an AI-powered mental health assistant — not just a chatbot, but a system that can retrieve user history, analyze input, access data in real-time, and suggest personalized treatment plans. UI Chat Tech stack: Loveable + Momen How it’s structured: It uses 7 specialized AI agents, each responsible for a niche task — chat, generate professional responses, summarize user info, classify intent, etc. Agent Team The main agent (the chat one) will call other agents in the backend via automated workflows. It keeps track of user data (symptoms, conversations, medical history) and updates it in real time — all triggered automatically. Everything runs in the backend to reduce manual steps and minimize errors. How it’s built: Started by drafting the UI with Loveable AI — it auto-generated a 7-page interface from a product brief, which saved me time. (Didn’t use it for the live app though — good for prototyping, but I wanted more control for complex backend workflows.) Rebuilt the UI and database in Momen, since I needed deeper control over data flow and backend logic. The entire AI agent system and backend workflows were built in Momen as well. So I can make the agents collaborate with each other. The main chat agent invokes backend workflows to call other agents when needed. Entire flow looks like this: the user sends a message, the system: → pulls in the latest user data→ triggers the right agent(s) based on the input→ responds in real-time→ quietly summarizes and updates everything in the background. FlowChart It’s still an MVP, but the multi-agent setup + automated backend feels pretty scalable.This was a super fun build and I learned a lot about orchestrating AI workflows. Would love any feedback or thoughts on how to improve this.

I created leadsnavi that helps small businesses find quality leads without breaking the bank
reddit
LLM Vibe Score0
Human Vibe Score1
BrightCook5861This week

I created leadsnavi that helps small businesses find quality leads without breaking the bank

Hey Redditors, I’m excited to share LeadsNavi, a tool I built specifically to help small businesses and B2B professionals automatically generate leads and reach potential customers in a smarter way. After talking to a lot of small business owners, I realized how tough it is to juggle lead generation with limited resources. So, I decided to create a tool that could simplify the process and make it more accessible to those who don’t have the budget to invest in expensive solutions. What Exactly Is LeadsNavi? LeadsNavi is an intuitive, cost-effective platform that automates the process of lead generation. It's designed to make it easy for small businesses and entrepreneurs to identify quality leads and grow their customer base without the need for manual prospecting. Here’s what makes it stand out: Automatic Lead Tracking: Tracks visitors to your website and matches them with company data, so you get real insights into who’s interested in your business. AI-Powered Lead Recommendations: Based on your website’s traffic, LeadsNavi uses AI to suggest similar companies that could be interested in your product or service, helping you find new leads faster and more accurately. Affordable & Scalable: For only $49/month, you can use a highly effective tool that scales with your business. It’s designed to be affordable even for small businesses. CRM Integration: Connect your CRM to directly import leads and sync your outreach efforts. How Does It Work? LeadsNavi uses advanced algorithms to track website visitors' IP addresses and match them with a comprehensive business database. It provides details like company names, contact information, and helps you identify potential leads for follow-up. The best part? It works automatically, saving you hours of manual work and effort. Lead Identification: Get insights into which companies are visiting your website. AI-Driven Lead Recommendations: The AI analyzes your site’s traffic and suggests other companies in the same industry or with similar needs that might be a great fit for your product or service. Data-Enriched Leads: Gather real-time, actionable data on these leads to make your outreach more targeted. Easy Setup: Simply integrate with your website and CRM to start getting quality leads in minutes. Who’s It For? Small Businesses: You don’t have to be a marketing expert to generate quality leads. B2B Sales Teams: Perfect for anyone looking to target other businesses with a streamlined and automated approach. Entrepreneurs & Startups: Focus on scaling your business without worrying about lead generation overhead. Why Try It? LeadsNavi gives you the power to focus on what really matters—connecting with potential customers and scaling your business. If you’ve been struggling with finding quality leads, or if you’re just getting started, I believe LeadsNavi can help you save time, effort, and money. I’m offering a 14-day free trial, so you can see the tool in action before committing to anything. Give it a try and let me know what you think! I’d love to hear your thoughts, suggestions, and how it works for your business. https://preview.redd.it/fdwil4rssgle1.png?width=1867&format=png&auto=webp&s=eb73b41a2b7665ae1b651fe2a6b7459df6990530

I built a library to visualize and edit audio filters
reddit
LLM Vibe Score0
Human Vibe Score1
AlexStreletsThis week

I built a library to visualize and edit audio filters

Hey everyone! TLDR: No fancy AI Agents or trendy micro-SaaS here — just an old-school library. Scroll down for the demo link! 🙃 App Demo The Story Behind Several years ago, I deep-dived into reverse engineering the parameter system used in VAG (Volkswagen, Audi, Porsche, etc) infotainment units. I managed to decode their binary format for storing settings for each car type and body style. To explain it simply - their firmware contains equalizer settings for each channel of the on-board 5.1 speaker system based on cabin volume and other parameters, very similar to how home theater systems are configured (gains, delays, limiters, etc). I published this research for the car enthusiast community. While the interest was huge, the reach remained small since most community members weren't familiar with hex editors. Only a few could really replicate what I documented. After some time, I built a web application that visualized these settings and allowed to unpack, edit and repack that data back into the binary format. Nowadays The original project was pretty messy (spaghetti code, honestly) and had a very narrow focus. But then I realized the visualization library itself could be useful for any audio processing software. When I first tried to visualize audio filters with that project, I hit a wall. Most charting libraries are built for business data, all those "enterprise-ready visualization solutions". But NONE of them is designed for audio-specific needs. D3.js is the only real option here — it’s powerful but requires days of digging through docs just to get basic styling right. And if you want interactive features like drag-and-drop? Good luck with that. (Fun fact: due to D3's multiple abstraction layers, just the same filter calculations in DSSSP are 1.4-2x faster than D3's implementation). So, I built a custom vector-based graph from scratch with a modern React stack. The library focuses on one thing - audio filters. No unnecessary abstractions, no enterprise bloat, just fast and convenient (I hope!) tools for tools for audio processing software. Core Features Logarithmic frequency response visualization Interactive biquad filter manipulation Custom audio calculation engine Drag-and-drop + Mouse wheel controls Flexible theming API Technical Details Built with React + SVG (no Canvas) Zero external dependencies besides React Full TypeScript support Live Demo & Docs & GitHub This is the first public release, landing page is missing, and the backlog is huge, and docs do not cover some aspects. (You know, there's never a perfect timimng - I just had to stop implementing my ideas and make it community driven). I'd love to see what you could build with these components. What's missing? What could be improved? I'm still lacking the understanding of how it could gain some cash flow, while staying open-source. Any ideas?

How I got 1000 users on day one.
reddit
LLM Vibe Score0
Human Vibe Score1
Human-Grape-8319This week

How I got 1000 users on day one.

This might sound like a small number, depending on who you ask, but you know it’s a start. I’ll just share my learnings so far. Introduction: The product is simple: you type what you want to build, like, let's say, a SaaS idea, and it generates the code using a framework of your choice (like NextJS). Currently, it only generates front-end code. The marketing strategy was mainly focused on social media. My social media stats are as follows: I have a whopping 14 followers on Twitter, and 10 of them are bot accounts, and on LinkedIn, it’s about 400 or so. Launching on LinkedIn: LinkedIn is unique in two different ways: The algorithm is friendly to the little guy. Your network (the people) aren’t always friendly to the little guy. Let me elaborate. This is something I learned today, actually. When I posted for the first time and asked about three of my friends to repost it, within the first hour there were about 200 views, and the click-through rate was around 40%. This was really good, given that it was in the morning. I don’t know the exact factors, but I did have a video in my post, and those three reposts probably amplified it. However, people don’t seem to like or comment on it as much as you would think. Most of my connections are CS students because I am a recent grad, so it seems like most people can relate to this product, but none of them would even put a comment or a like. At the same time, I see people liking posts from big brands like OpenAI, Microsoft, etc. I am really confused, to be honest. However, throughout the day, the view count was going up, and people were coming. Launching on Twitter: Twitter didn’t really work for me at all. I think you need a decent audience. But there are tweets like “What startup are you working on?” type questions, and from that, I find you get a couple of views on your profile. Even though Twitter didn’t really help with the views, one guy tweeted, “Keep posting on Twitter and one day this might become something like Notion.” That really made my day, to be honest. Launching on Discord: This worked really well, to be honest, especially given that I was in a lot of Discord servers where there are software devs. If you use the right language that resonates with them, it’s a home run. Not much to say, but don’t use marketing lingo; people don’t like it there. Instagram and TikTok didn’t really work. Mainly, I think my video didn’t really resonate much. Finally, Facebook Launch: The Facebook reels didn’t really do the trick. Then I posted in a bunch of groups, and still, it didn’t really do anything. But then I sent cold DMs on Facebook, and that had a pretty high open rate because I sent them to people who I saw commented on posts related to what my product was solving. Obviously, after a while, Facebook blocks the ability to send DMs. That’s all for now. Thanks I’ll post my promo video in the comment section just so that you know the video and why it might have resonated with some platforms. Also this is the first time I made a video and I’m actually proud of making that more than the product itself. To summarize, for this idea LinkedIn worked really well, because of the algorithm not the ppl commenting and liking which is what I thought should be the way. Followed by Discord groups and Facebook DMs. The video I made seemed to resonate really well with the LinkedIn audience (the engagement was around 60%) despite falling in TikTok and other video sharing platforms.

I am building a free app to fight propaganda online. Need your help
reddit
LLM Vibe Score0
Human Vibe Score-1
Used_Park_1937This week

I am building a free app to fight propaganda online. Need your help

So long story short I was recently bombarded with an insane amounts of right-wing propaganda online and Russian bots so I decided I can somehow simplify the validation of the facts that are posted by people online. I have made a research of tools that are currently available for that and I've found that ChatGpt is quite biased with some topics and it lacks data about latest issues even when search mode is enabled. At least it takes some time for it to update it's memory or something Then I've checked a Google Fact Check tools which were even worse. Bad UI... I had no idea how it validates things because usually proof links that Google provides are quite random. So I decided to build a tool that simplifies the research. The idea was simple - you have a claim - you paste it in app and it goes on web collects articles on topic then processes them with ai and makes a decision based on those articles. The app is running right now. It is completely free and no ads. Also it generates a small article where it tells about how it made a conclusion. Also I need your help - How can I make this app better? Also feel free to test your claims so that I can have more real data to tune this app on. I believe that this is something that is necessary for all of us especially in those crazy times! Here is the link - truthorfake.com https://preview.redd.it/vp5fbzpyfkne1.png?width=854&format=png&auto=webp&s=84df7913382634614814866c8df743d8378d28dd

How me and my team made 15+ apps and not made a single sale in 2023
reddit
LLM Vibe Score0
Human Vibe Score0.818
MichaelbetterecycleThis week

How me and my team made 15+ apps and not made a single sale in 2023

Hey, my name is Michael, I am in Auckland NZ. This year was the official beginning of my adult life. I graduated from university and started a full-time job. I’ve also really dug into indiehacking/bootstrapping and started 15 projects (and it will be at least 17 before the year ends). I think I’ve learned a lot but I consciously repeated mistakes. Upto (Nov) Discord Statuses + Your Location + Facebook Poke https://preview.redd.it/4nqt7tp2tf5c1.png?width=572&format=png&auto=webp&s=b0223484bc54b45b5c65e0b1afd0dc52f9c02ad1 This was the end of uni, I often messaged (and got messaged) requests of status and location to (and from my) friends. I thought, what if we make a social app that’s super basic and all it does is show you where your friends are? To differentiate from snap maps and others we wanted something with more privacy where you select the location. However, never finished the codebase or launched it. This is because I slowly started to realize that B2C (especially social networks) are way too hard to make into an actual business and the story with Fistbump would repeat itself. However, this decision not to launch it almost launched a curse on our team. From that point, we permitted ourselves to abandon projects even before launching. Lessons: Don’t do social networks if your goal is 10k MRR ASAP. If you build something to 90% competition ship it or you will think it’s okay to abandon projects Insight Bites (Nov) Youtube Summarizer Extension &#x200B; https://preview.redd.it/h6drqej4tf5c1.jpg?width=800&format=pjpg&auto=webp&s=0f211456c390ac06f4fcb54aa51f9d50b0826658 Right after Upto, we started ideating and conveniently the biggest revolution in the recent history of tech was released → GPT. We instantly began ideating. The first problem we chose to use AI for is to summarize YouTube videos. Comical. Nevertheless, I am convinced we have had the best UX because you could right-click on a video to get a slideshow of insights instead of how everyone else did it. We dropped it because there was too much competition and unit economics didn’t work out (and it was a B2C). PodPigeon (Dec) Podcast → Tweet Threads https://preview.redd.it/0ukge245tf5c1.png?width=2498&format=png&auto=webp&s=23303e1cab330578a3d25cd688fa67aa3b97fb60 Then we thought, to make unit economics work we need to make this worthwhile for podcasters. This is when I got into Twitter and started seeing people summarize podcasts. Then I thought, what if we make something that converts a podcast into tweets? This was probably one of the most important projects because it connected me with Jason and Jonaed, both of whom I regularly stay in contact with and are my go-to experts on ideas related to content creation. Jonaed was even willing to buy Podpigeon and was using it on his own time. However, the unit economics still didn’t work out (and we got excited about other things). Furthermore, we got scared of the competition because I found 1 - 2 other people who did similar things poorly. This was probably the biggest mistake we’ve made. Very similar projects made 10k MRR and more, launching later than we did. We didn’t have a coherent product vision, we didn’t understand the customer well enough, and we had a bad outlook on competition and a myriad of other things. Lessons: I already made another post about the importance of outlook on competition. Do not quit just because there are competitors or just because you can’t be 10x better. Indiehackers and Bootstrappers (or even startups) need to differentiate in the market, which can be via product (UX/UI), distribution, or both. Asking Ace Intro.co + Crowdsharing &#x200B; https://preview.redd.it/0hu2tt16tf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3d397568ef2331e78198d64fafc1a701a3e75999 As I got into Twitter, I wanted to chat with some people I saw there. However, they were really expensive. I thought, what if we made some kind of crowdfunding service for other entrepreneurs to get a private lecture from their idols? It seemed to make a lot of sense on paper. It was solving a problem (validated via the fact that Intro.co is a thing and making things cheaper and accessible is a solid ground to stand on), we understood the market (or so we thought), and it could monetize relatively quickly. However, after 1-2 posts on Reddit and Indiehackers, we quickly learned three things. Firstly, no one cares. Secondly, even if they do, they think they can get the same information for free online. Thirdly, the reasons before are bad because for the first point → we barely talked to people, and for the second people → we barely talked to the wrong people. However, at least we didn’t code anything this time and tried to validate via a landing page. Lessons Don’t give up after 1 Redditor says “I don’t need this” Don’t be scared to choose successful people as your audience. Clarito Journaling with AI analyzer https://preview.redd.it/8ria2wq6tf5c1.jpg?width=1108&format=pjpg&auto=webp&s=586ec28ae75003d9f71b4af2520b748d53dd2854 Clarito is a classic problem all amateur entrepreneurs have. It’s where you lie to yourself that you have a real problem and therefore is validated but when your team asks you how much you would pay you say I guess you will pay, maybe, like 5 bucks a month…? Turns out, you’d have to pay me to use our own product lol. We sent it off to a few friends and posted on some forums, but never really got anything tangible and decided to move away. Honestly, a lot of it is us in our own heads. We say the market is too saturated, it’ll be hard to monetize, it’s B2C, etc. Lessons: You use the Mom Test on other people. You have to do it yourself as well. However, recognizing that the Mom Test requires a lot of creativity in its investigation because knowing what questions to ask can determine the outcome of the validation. I asked myself “Do I journal” but I didn’t ask myself “How often do I want GPT to chyme in on my reflections”. Which was practically never. That being said I think with the right audience and distribution, this product can work. I just don’t know (let alone care) about the audience that much (and I thought I was one of them)/ Horns & Claw Scrapes financial news texts you whether you should buy/sell the stock (news sentiment analysis) &#x200B; https://preview.redd.it/gvfxdgc7tf5c1.jpg?width=1287&format=pjpg&auto=webp&s=63977bbc33fe74147b1f72913cefee4a9ebec9c2 This one we didn’t even bother launching. Probably something internal in the team and also seemed too good to be true (because if this works, doesn’t that just make us ultra-rich fast?). I saw a similar tool making 10k MRR so I guess I was wrong. Lessons: This one was pretty much just us getting into our heads. I declared that without an audience it would be impossible to ship this product and we needed to start a YouTube channel. Lol, and we did. And we couldn’t even film for 1 minute. I made bold statements like “We will commit to this for at least 1 year no matter what”. Learnery Make courses about any subject https://preview.redd.it/1nw6z448tf5c1.jpg?width=1112&format=pjpg&auto=webp&s=f2c73e8af23b0a6c3747a81e785960d4004feb48 This is probably the most “successful” project we’ve made. It grew from a couple of dozen to a couple of hundred users. It has 11 buy events for $9.99 LTD (we couldn’t be bothered connecting Stripe because we thought no one would buy it anyway). However what got us discouraged from seriously pursuing it more is, that this has very low defensibility, “Why wouldn’t someone just use chatGPT?” and it’s B2C so it’s hard to monetize. I used it myself for a month or so but then stopped. I don’t think it’s the app, I think the act of learning a concept from scratch isn’t something you do constantly in the way Learnery delivers it (ie course). I saw a bunch of similar apps that look like Ass make like 10k MRR. Lessons: Don’t do B2C, or if you do, do it properly Don’t just Mixpanel the buy button, connect your Stripe otherwise, it doesn’t feel real and you won’t get momentum. I doubt anyone (even me) will make this mistake again. I live in my GPT bubble where I make assumptions that everyone uses GPT the same way and as much as I do. In reality, the argument that this has low defensibility against GPT is invalid. Platforms that deliver a differentiated UX from ChatGPT to audiences who are not tightly integrated into the habit of using ChatGPT (which is like - everyone except for SOME tech evangelists). CuriosityFM Make podcasts about any subject https://preview.redd.it/zmosrcp8tf5c1.jpg?width=638&format=pjpg&auto=webp&s=d04ddffabef9050050b0d87939273cc96a8637dc This was our attempt at making Learnery more unique and more differentiated from chatGPT. We never really launched it. The unit economics didn’t work out and it was actually pretty boring to listen to, I don’t think I even fully listened to one 15-minute episode. I think this wasn’t that bad, it taught us more about ElevenLabs and voice AI. It took us maybe only 2-3 days to build so I think building to learn a new groundbreaking technology is fine. SleepyTale Make children’s bedtime stories https://preview.redd.it/14ue9nm9tf5c1.jpg?width=807&format=pjpg&auto=webp&s=267e18ec6f9270e6d1d11564b38136fa524966a1 My 8-year-old sister gave me that idea. She was too scared of making tea and I was curious about how she’d react if she heard a bedtime story about that exact scenario with the moral that I wanted her to absorb (which is that you shouldn’t be scared to try new things ie stop asking me to make your tea and do it yourself, it’s not that hard. You could say I went full Goebbels on her). Zane messaged a bunch of parents on Facebook but no one really cared. We showed this to one Lady at the place we worked from at Uni and she was impressed and wanted to show it to her kids but we already turned off our ElevenLabs subscription. Lessons: However, the truth behind this is beyond just “you need to be able to distribute”. It’s that you have to care about the audience. I don’t particularly want to build products for kids and parents. I am far away from that audience because I am neither a kid anymore nor going to be a parent anytime soon, and my sister still asked me to make her tea so the story didn’t work. I think it’s important to ask yourself whether you care about the audience. The way you answer that even when you are in full bias mode is, do you engage with them? Are you interested in what’s happening in their communities? Are you friends with them? Etc. User Survey Analyzer Big User Survey → GPT → Insights Report Me and my coworker were chatting about AI when he asked me to help him analyze a massive survey for him. I thought that was some pretty decent validation. Someone in an actual company asking for help. Lessons Market research is important but moving fast is also important. Ie building momentum. Also don’t revolve around 1 user. This has been a problem in multiple projects. Finding as many users as possible in the beginning to talk to is key. Otherwise, you are just waiting for 1 person to get back to you. AutoI18N Automated Internationalization of the codebase for webapps This one I might still do. It’s hard to find a solid distribution strategy. However, the idea came from me having to do it at my day job. It seems a solid problem. I’d say it’s validated and has some good players already. The key will be differentiation via the simplicity of UX and distribution (which means a slightly different audience). In the backlog for now because I don’t care about the problem or the audience that much. Documate - Part 1 Converts complex PDFs into Excel https://preview.redd.it/8b45k9katf5c1.jpg?width=1344&format=pjpg&auto=webp&s=57324b8720eb22782e28794d2db674b073193995 My mom needed to convert a catalog of furniture into an inventory which took her 3 full days of data entry. I automated it for her and thought this could have a big impact but there was no distribution because there was no ICP. We tried to find the ideal customers by talking to a bunch of different demographics but I flew to Kazakhstan for a holiday and so this kind of fizzled out. I am not writing this blog post linearity, this is my 2nd hour and I am tired and don’t want to finish this later so I don’t even know what lessons I learned. Figmatic Marketplace of high-quality Figma mockups of real apps https://preview.redd.it/h13yv45btf5c1.jpg?width=873&format=pjpg&auto=webp&s=aaa2896aeac2f22e9b7d9eed98c28bb8a2d2cdf1 This was a collab between me and my friend Alex. It was the classic Clarito where we both thought we had this problem and would pay to fix it. In reality, this is a vitamin. Neither I, nor I doubt Alex have thought of this as soon as we bought the domain. We posted it on Gumroad, sent it to a bunch of forums, and called it a day. Same issue as almost all the other ones. No distribution strategy. However, apps like Mobin show us that this concept is indeed profitable but it takes time. It needs SEO. It needs a community. None of those things, me and Alex had or was interested in. However shortly after HTML → Figma came out and it’s the best plugin. Maybe that should’ve been the idea. Podcast → Course Turns Podcaster’s episodes into a course This one I got baited by Jason :P I described to him the idea of repurposing his content for a course. He told me this was epic and he would pay. Then after I sent him the demo, he never checked it out. Anyhow during the development, we realized that doesn’t actually work because A podcast doesn’t have the correct format for the course, the most you can extract are concepts and ideas, seldom explanations. Most creators want video-based courses to be hosted on Kajabi or Udemy Another lesson is that when you pitch something to a user, what you articulate is a platform or a process, they imagine an outcome. However, the end result of your platform can be a very different outcome to what they had in mind and there is even a chance that what they want is not possible. You need to understand really well what the outcome looks like before you design the process. This is a classic problem where we thought of the solution before the problem. Yes, the problem exists. Podcasters want to make courses. However, if you really understand what they want, you can see how repurposing a podcast isn’t the best way to get there. However I only really spoke to 1-2 podcasters about this so making conclusions is dangerous for this can just be another asking ace mistake with the Redditor. Documate Part 2 Same concept as before but now I want to run some ads. We’ll see what happens. https://preview.redd.it/xb3npj0ctf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3cd4884a29fd11d870d010a2677b585551c49193 In conclusion https://preview.redd.it/2zrldc9dtf5c1.jpg?width=1840&format=pjpg&auto=webp&s=2b3105073e752ad41c23f205dbd1ea046c1da7ff It doesn’t actually matter that much whether you choose to do a B2C, or a social network or focus on growing your audience. All of these can make you successful. What’s important is that you choose. If I had to summarize my 2023 in one word it’s indecision. Most of these projects succeeded for other people, nothing was as fundamentally wrong about them as I proclaimed. In reality that itself was an excuse. New ideas seduce, and it is a form of discipline to commit to a single project for a respectful amount of time. https://preview.redd.it/zy9a2vzdtf5c1.jpg?width=1456&format=pjpg&auto=webp&s=901c621227bba0feb4efdb39142f66ab2ebb86fe Distribution is not just posting on Indiehackers and Reddit. It’s an actual strategy and you should think of it as soon as you think of the idea, even before the Figma designs. I like how Denis Shatalin taught me. You have to build a pipeline. That means a reliable way to get leads, launch campaigns at them, close deals, learn from them, and optimize. Whenever I get an idea now I always try to ask myself “Where can I find 1000s leads in one day?” If there is no good answer, this is not a good project to do now. &#x200B; https://preview.redd.it/2boh3fpetf5c1.jpg?width=1456&format=pjpg&auto=webp&s=1c0d5d7b000716fcbbb00cbad495e8b61e25be66 Talk to users before doing anything. Jumping on designing and coding to make your idea a reality is a satisfying activity in the short term. Especially for me, I like to create for the sake of creation. However, it is so important to understand the market, understand the audience, understand the distribution. There are a lot of things to understand before coding. https://preview.redd.it/lv8tt96ftf5c1.jpg?width=1456&format=pjpg&auto=webp&s=6c8735aa6ad795f216ff9ddfa2341712e8277724 Get out of your own head. The real reason we dropped so many projects is that we got into our own heads. We let the negative thoughts creep in and kill all the optimism. I am really good at coming up with excuses to start a project. However, I am equally as good at coming up with reasons to kill a project. And so you have this yin and yang of starting and stopping. Building momentum and not burning out. I can say with certainty my team ran out of juice this year. We lost momentum so many times we got burnt out towards the end. Realizing that the project itself has momentum is important. User feedback and sales bring momentum. Building also creates momentum but unless it is matched with an equal force of impact, it can stomp the project down. That is why so many of our projects died quickly after we launched. The smarter approach is to do things that have a low investment of momentum (like talking to users) but result in high impact (sales or feedback). Yes, that means the project can get invalidated which makes it more short-lived than if we built it first, but it preserves team life energy. At the end of 2023 here is a single sentence I am making about how I think one becomes a successful indiehacker. One becomes a successful Indiehacker when one starts to solve pain-killer problems in the market they understand, for an audience they care about and consistently engage with for a long enough timeframe. Therefore an unsuccessful Indiehacker in a single sentence is An unsuccessful Indiehacker constantly enters new markets they don’t understand to build solutions for people whose problems they don’t care about, in a timeframe that is shorter than than the time they spent thinking about distribution. However, an important note to be made. Life is not just about indiehacking. It’s about learning and having fun. In the human world, the best journey isn’t the one that gets you the fastest to your goals but the one you enjoy the most. I enjoyed making those silly little projects and although I do not regret them, I will not repeat the same mistakes in 2024. But while it’s still 2023, I have 2 more projects I want to do :) EDIT: For Devs, frontend is always react with vite (ts) and backend is either node with express (ts) or python. For DB either Postgres or mongo (usually Prisma for ORM). For deployment all of it is on AWS (S3, EC2). In terms of libraries/APIs Whisper.cpp is best open source for transcription Obviously the gpt apis Eleven labs for voice related stuff And other random stuff here and there

How I built my SaaS and earned $273 MRR in the first month
reddit
LLM Vibe Score0
Human Vibe Score1
Ok_Damage_1764This week

How I built my SaaS and earned $273 MRR in the first month

Hi everyone! I’m Alex Varga, an indie developer. Last year, I focused on accelerating my development speed and launched 10 projects in 12 months. One of them called Bulk Image Generation started growing through SEO, so I decided to focus on it. After one month of SEO efforts, it’s generating $273 MRR. I hope my experience will be useful to others. Concept bulkimagegeneration.com website helps to generate up to 100 images in 15 seconds using AI I was using Google, started with keywords like "Bulk Image ..." a lot of them are Bulk Image Resizer, Downloader etc. But there was no Bulk Image Generator. I thought: yeah, this domain is available, let's buy. So I bought bulkimagegeneration.com and bulkimagegenerator.com So, the app concept is to help people generate images with AI at scale: let\`s say 100 images in 15 seconds. Marketing Gap https://preview.redd.it/4luzib02bbie1.png?width=1905&format=png&auto=webp&s=cbe845107aca46ae5729dfe121fefd5e9cdab9ac Most builders create a product first and figure out how to sell it later. I took a completely different approach with Bulk Image Generator. I identified a market gap and secured a domain name that matched exactly what people were searching for and launched app. https://preview.redd.it/h6vwur34bbie1.png?width=1905&format=png&auto=webp&s=9a163ff6f503be4c175c6e5e82e2003b32df1fe0 Growth Strategy SEO has become the main acquisition channel, so I’ve decided to focus even more on it with this experiment. Almost every day, I publish either a new article or a free micro-app (as a lead magnet) for Bulk Image Generator. I also tried Google Ads, spent $20, and got a $0.35 CPC. https://preview.redd.it/3rhnzvs6bbie1.png?width=1905&format=png&auto=webp&s=f9819d1e82d3e2429d6ccb7b00dcac86a7a351c2 In comparison, the Free Image to Text Prompt Converter (one of the lead magnets) has a $0.011 CPC, which is more than 30 times cheaper than Google Ads. So I decided not to focus now on paid ads. https://preview.redd.it/p333fyl9bbie1.png?width=1905&format=png&auto=webp&s=2e96532d7709b44b7459e7ccf37ef9a0fa784728 After using our free tools, some users explore our main product - a bulk image generation service. Users pay a monthly subscription to get credits, which they can spend on image generation, face swaps, and bulk background removal. Currently, this app generates around $250 in Monthly Recurring Revenue: https://preview.redd.it/9wcm0tjfbbie1.png?width=1905&format=png&auto=webp&s=41bcdd4f7594b09087c51cc5044e4b9c94c129c8 SEO Keyword Research I use Semrush or similar tools to find keywords with a search volume greater than 300 and then write articles targeting those keywords. If the topic has enough potential, I might create a free tool (e.g., a Free Image to Text Prompt Converter) to attract more users. Occasions matter. For instance, I wrote an article about creating images for Super Bowl ads, which led to one paying user who replicated the exact creatives showcased in the article https://preview.redd.it/shpax6mlbbie1.png?width=1905&format=png&auto=webp&s=d491385761df126424c2f9ba14c5da15f8cbb603 AI Tools Aggregators This can be an excellent acquisition channel. When BulkImageGeneration.com was featured in an article on Toolify.ai, I immediately gained three paying users (\~$60). I took 2 more AI Aggregators, and on average I had CPC = $0.2, which is a fair price and usually it has ROAs > 100%. However, some major aggregators are expensive ($300–400 per placement). I want to try it once I reach $500+ MRR. Next Steps bulkimagegeneration.com currently ranks #1 in search results for relevant keywords (e.g., “bulk image generation,” “bulk image generator”). I plan to keep producing content targeting niche keywords and timely occasions. buy more places in AI Aggregators I also want to reach out to YouTubers and ask them to include Bulk in their reviews for free

ChatPDF and PDF.ai are making millions using open source tech... here's the code
reddit
LLM Vibe Score0
Human Vibe Score1
Level-Thought6152This week

ChatPDF and PDF.ai are making millions using open source tech... here's the code

Why "copy" an existing product? The best SaaS products weren’t the first of their kind - think Slack, Shopify, Zoom, Dropbox, or HubSpot. They didn’t invent team communication, e-commerce, video conferencing, cloud storage, or marketing tools; they just made them better. What is a "Chat with PDF" SaaS? These are AI-powered PDF assistants that let you upload a PDF and ask questions about its content. You can summarize articles, extract key details from a contract, analyze a research paper, and more. To see this in action or dive deeper into the tech behind it, check out this YouTube video. Let's look at the market Made possible by advances in AI like ChatGPT and Retrieval-Augmented Generation (RAG), PDF chat tools started gaining traction in early 2023 and have seen consistent growth in market interest, which is currently at an all-time high (source:google trends) Keywords like "chat PDF" and "PDF AI" get between 1 to 10 million searches every month (source:keyword planner), with a broad target audience that includes researchers, students, and professionals across various industries. Leaders like PDF.ai and ChatPDF have already gained millions of users within a year of launch, driven by the growing market demand, with paid users subscribing at around $20/month. Alright, so how do we build this with open source? The core tech for most PDF AI tools are based on the same architecture. You generate text embeddings (AI-friendly text representations; usually via OpenAI APIs) for the uploaded PDF’s chapters/topics and store them in a vector database (like Pinecone). Now, every time the user asks a question, a similarity search is performed to find the most similar PDF topics from the vector database. The selected topic contents are then sent to an LLM (like ChatGPT) along with the question, which generates a contextual answer! Here are some of the best open source implementations for this process: GPT4 & LangChain Chatbot for large PDF docs by Mayo Oshin MultiPDF Chat App by Alejandro AO PDFToChat by Hassan El Mghari Worried about building signups, user management, payments, etc.? Here are my go-to open-source SaaS boilerplates that include everything you need out of the box: SaaS Boilerplate by Remi Wg Open SaaS by wasp-lang A few ideas to stand out from the noise: Here are a few strategies that could help you differentiate and achieve product market fit (based on the pivot principles from The Lean Startup by Eric Ries): Narrow down your target audience for a personalized UX: For instance, an exam prep assistant for students with study notes and quiz generator; or a document due diligence and analysis tool for lawyers. Add unique features to increase switching cost: You could autogenerate APIs for the uploaded PDFs to enable remote integrations (eg. support chatbot knowledge base); or build in workflow automation features for bulk analyses of PDFs. Offer platform level advantages: You could ship a native mobile/desktop apps for a more integrated UX; or (non-trivial) offer private/offline support by replacing the APIs with local open source deployments (eg. llama for LLM, an embedding model from the MTEB list, and FAISS for vector search). TMI? I’m an ex-AI engineer and product lead, so don’t hesitate to reach out with any questions! P.S. I've started a free weekly newsletter to share open-source/turnkey resources behind popular products (like this one). If you’re a founder looking to launch your next product without reinventing the wheel, please subscribe :)

My Marketing App made $10,000 in 2024. Here is how I target to make $100,000 in 2025:
reddit
LLM Vibe Score0
Human Vibe Score1
MonkDiThis week

My Marketing App made $10,000 in 2024. Here is how I target to make $100,000 in 2025:

You totally get me, I think. It’s a bizarre feeling when you build something, and people appreciate it and are even ready to pay! Pleasant though) In early 2024 my mate and I created a marketing tool that generates ads, content and strategy blocks with a click – Aiter.io. Users can just insert a URL, hit the button and everything is ready. TBH, I built this tool because I’m too lazy to chat with ChatGPT) https://preview.redd.it/ew2kud7ceyde1.png?width=1140&format=png&auto=webp&s=f3fe5b67075858cea3d52278e8063113efa3b97e In 2024 we made $10,000, here is what worked for us: AI directories. Still is the best channel of traffic and clients for us. We listed on TAAFT and other directories scrape TAAFT, so, eventually, we became listed on all major ones. I wrote a Reddit post earlier that explained this process in detail. Email marketing. Gosh, I thought it was dead – I have never been so wrong! We set up automatic emails that share marketing insights and they have a \~25% open rate + consistently convert people. It works great. Product marketing. Having a free version really helps with word-of-mouth and leads, which can be converted via email. Also, we consistently worked on product improvement. I’d say, that our free updates give people a feeling that the devs care about their stuff that’s why they are more confident investing in it. Google Ads. TBH, we had a shitty landing page all the time because were busy with the product. So, Google Ads didn’t work well for us. But we’ve launched the 2.0 version which has a better landing page, and will try it again. Influencers. Worked well for us, but we didn’t pay a dime for this. They just found our tool on directories and created videos about Aiter, so it was a sporadic marketing channel for us. We hope to change it in 2025. We see that our product works and attracts the audience, so we want to deliver and get more in 2025. Here is the plan: Product: add ad banners and video generation. So far, we generate only text data and it’s not so valuable in the time of ChatGPT and Claude. But to generate a high-quality ad banner is still challenging, so we put this on our roadmap. Another feature – one-click market analysis to get marketing insights. Become a TOP50 tool on TAAFT. We’ve become a top tool in our category (content generation) but will need to promote our profile on the profile far more aggressively to get into TOP50 Email marketing. We are fools because we almost didn’t have product emails that explain how it works. Will fix it. Also, we are considering participating more in paid newsletters, like collaborating with Substack influencers. Youtube marketing. Search for low-tail marketing keywords on YouTube and create videos on them, placing my product in them. Blog. Our new platform is Webflow which gives a lot of flexibility in terms of blogging. So, we will repeat the YouTube strategy with blogging. Paid marketing. With an updated landing page, we hope that paid campaigns will work better. We plan to launch campaigns that target different jobs to be done and customer objections to find the right message. Product Management. For 2025, our two key product metrics are retention and product activation rate. For this, we plan to simplify onboarding and make it simpler as well as conduct a lot of in-depth interviews to understand how we can retain users better. Funding. All of this exciting stuff requires money, so we are in the process of securing funding (fingers crossed). Having an indie project is exciting and invigorating. With all these activities, I hope we will achieve the goal of $100,000 in 2025. And what are your goals and marketing steps for 2025? Or maybe you could share some exciting marketing ideas I overlooked?

Introducing Stratify: Your Ultimate AI Strategy Builder for Business Success
reddit
LLM Vibe Score0
Human Vibe Score0
vsengarThis week

Introducing Stratify: Your Ultimate AI Strategy Builder for Business Success

Hello, I’m thrilled to announce the launch of my new startup, Stratify! 🔍 What is Stratify? Stratify is an AI Strategy Builder designed to help businesses of all sizes develop, implement, and optimize their strategic plans using cutting-edge artificial intelligence. Whether you're a startup looking to scale or an established company aiming to innovate, Stratify provides the tools and insights you need to stay ahead in today's competitive landscape. 🌟 Key Features: Automated Strategy Development: Leverage AI to analyze market trends, competitor data, and internal metrics to create comprehensive strategic plans tailored to your business goals. Real-Time Analytics & Insights: Monitor your strategy's performance with real-time data dashboards, enabling you to make informed decisions quickly. Scenario Planning: Use AI-driven simulations to forecast different business scenarios and understand potential outcomes, helping you prepare for uncertainties. Collaborative Tools: Facilitate team collaboration with integrated communication features, ensuring everyone is aligned and contributing to the strategy development process. Customizable Templates: Access a library of industry-specific strategy templates that can be customized to fit your unique business needs. 💡 Why Stratify? In today's fast-paced business environment, creating and adapting effective strategies can be challenging. Many companies struggle with data overload, lack of actionable insights, and inefficient planning processes. Stratify addresses these pain points by harnessing the power of AI to streamline strategy building, making it more efficient, data-driven, and adaptable. 🚀 Our Journey So Far: Founded: August 2024 Milestones Achieved: Developed and tested our MVP with a select group of beta users What's Next: Launching our public beta in Q4 2024 Expanding our feature set based on user feedback Growing our team with experts in AI, business strategy, and customer success 🤝 How You Can Help: We’re eager to connect with early adopters, business strategists, and industry experts who can benefit from or contribute to Stratify. Here’s how you can get involved: Join Our Beta Program: Be among the first to experience Stratify and provide valuable feedback. Share Your Insights: Help us refine our features by sharing your business strategy challenges and needs. Spread the Word: If you know someone who could benefit from an AI-driven strategy builder, please share our mission and be an affiliate to earn rewards! 🌐 Learn More: Visit our website at AI-Powered Brand Strategy & Content Creation | Stratify (brandprovoke.com) and follow us for the latest updates and insights. 🙏 Thank You! A heartfelt thank you to the Reddit community for your support and encouragement. We’re excited to embark on this journey and look forward to your feedback and suggestions! Looking forward to your thoughts and questions!

Just completed a new type of language learning website - read popular stories scaled to different reading levels
reddit
LLM Vibe Score0
Human Vibe Score1
creedaaronThis week

Just completed a new type of language learning website - read popular stories scaled to different reading levels

As a language learner and software developer, I bootstrapped my project superlang.com over the past year working on the side. There is a mobile friendly web app now, and iOS/Android apps coming in a few months. A year ago I discovered the concept of "comprehensible input" as a way to help me learn German. Even if it's not a silver bullet, it sounded pretty great. Rather than drilling vocab or looking at grammar charts, I could "just read" and acquire the language. I picked up some fairy tales in German, and stories like Alice in Wonderland. Unfortunately, I couldn't really read them. I had to stop every sentence to look up words and try and decipher sentence constructions. Then I turned to some purpose built simple stories for German beginners. But there was a different problem... these were not really stories with any real plot. I could only read so many "Hans goes to the market" type stories before losing interest. My idea was to try to get the best of both worlds somehow. What if I could take a real story, say Alice in Wonderland (or even War and Peace), and dial the difficulty down to my level without losing the plotline. That way, beginners can start right away with something basically comprehensible. Then, you could also re-read the same story at increasing difficulty levels as you gain confidence. As a cherry on top, more illustrations would help with comprehension so each page could have a picture. Is it revolutionary? Maybe, maybe not. I am building off a well established idea of "graded readers" which are simplified stories meant for learning languages. And there are somewhat similar ideas out there now that AI is good at simplifying text, but none that really take this idea where it needs to be with many preloaded stories, multiple difficulty levels, high quality human verified text, and all the bells and whistles. I spent a year building Superlang and it is ready to put out there. Some quick notes: There are 3 languages so far, intended for native English speakers: German, French, and Spanish There are 3 difficulty levels you can set on each story: beginner (roughly A1-A2), intermediate (roughly A2-B1), and advanced (the same level as the original story, but typically B2+) There is premium version as producing the content was somewhat expensive. You can still do a lot of reading on the free version. I have done no marketing yet, except for this post :) The implementation is a combination of AI, and human proofreading and reviewing. In particular, the simplification of stories is very heavily AI driven. The illustrations for each page are AI as well. For translation, as many of you may be aware new LLM models are typically better than Google translate, but still far from perfect. I am very much a proponent of keeping real people in the loop, and so I have real people proofread the translations. That's why there are only about 700 pages of content so far and not tens of thousands. Let me know what you think, and if you find it helpful! Alice in Wonderland - beginner level German Romeo and Juliet - beginner level Spanish

I made a super niche app for sailors and scaled it to 500k downloads
reddit
LLM Vibe Score0
Human Vibe Score0.5
TechPrimoThis week

I made a super niche app for sailors and scaled it to 500k downloads

I started developing this app in 2016, and it was my first app ever. I already had several years of programming experience. Since I was studying maritime navigation, I came up with the idea of creating a maritime app to help students with various nautical calculations and learn maritime regulations. Although I had no experience in mobile app development, I chose the Ionic framework and started development gradually. First Version The first version took me about four months to develop because I literally had to learn everything from scratch: how to develop mobile apps, how to publish them, and everything needed to enable downloads on the app stores. Many of you might recognize me from my story about developing Sintelly and its late monetization. I made the same mistake with this maritime app. At that time, in my country, there was no possibility of earning through in-app purchases, only through ad displays. Since the app was predominantly downloaded in countries like India, the Philippines, and Indonesia, the ad revenue was quite low, and after some time, I removed the ads. Abandonment and Realization As I started developing other apps, this one fell into obscurity. I even just remembered that I needed to renew the domain, which resulted in losing it. The domain buyer tried to sell it back to me for years for $20k, which was absurd. All this led me to rebrand and start working on this app again. Interestingly, during these 8 years, the app never showed a declining trend in installations or active users. I'll share some numbers to give you insight: Total installations (Android + iOS): 501,000 Active installations (Android): 48,000 Monthly active users: 20,000 Average rating: Android 4.8, iOS 4.7 When I considered these numbers, I realized they weren't bad at all and that I was far ahead of most competitors. This led to my decision to rebrand and create a new website. I quickly built the website using WordPress and published lots of existing content from the app. What surprises me is that today, after a year and a half, the website has about 8-10k monthly organic visits. Choosing a Direction Based on all this, I decided it was time to create a Premium version and start selling the app. Since I've been working with AI for many years (which I've written about here), I started thinking about using AI to help seafarers speed up some of their tasks. This led to the idea of creating a multi-agent system equipped with numerous tools to help seafarers. I developed various agents with functionalities, including retrieving maritime weather information, locating and tracking ships, doing various nautical calculations, calculating the shortest maritime routes and unit conversions, and learning about all courses and maritime regulations. All this required considerable work, but thanks to tools like Cursor and Claude, I implemented it in less than four weeks. Last week, I published this new version and started selling subscriptions, and I can already boast that I've earned slightly over $100. This isn't much, but I'm happy to see my first app generating some income, which I always thought impossible. Along this journey, I learned many lessons, and the most important one is to never give up or write off a product. With a little effort, everything can be brought back to life and secure at least some passive income, enough for your morning coffee. Additionally, I learned how to develop mobile apps, which has shaped my career since then. If it weren't for this app, I probably would never have become a developer. I have numerous plans for what to add next and how to improve. I'll base everything on AI features and push the app in that direction.

Finally launched my own app in the app store!
reddit
LLM Vibe Score0
Human Vibe Score0.429
ranftThis week

Finally launched my own app in the app store!

After reading on the sidelines here for about a year I just launched Kalo. My app is the 100th million ai powered calorie-counting app, hahaha. I know I know. Here it comes: Kalo Screenshots Despite being in a crowded space, Kalo has some caveats I am a bit proud of: \- I am a daily user of my app. Everything that bugs me will be gone ASAP. \- I have already lost 10kg with Kalo. I can't do any sports due to an energy-debilitating sickness (hello my me/cfs friends 👋), so this is huge. \- I HATE nudging. Hence, Kalo has no streaks, no notifications to rip off your valuable time. It’s just a tool to track calories and learn to get a feel for it. \- Ease of daily use and doing anything so it doesn't feel like a grind is Kalo's mission. I already implemented a lot of ways to quickly access tracking and leaving the app. \- Next feature will be tracking your own progress with some proper research based analytics is the one next step, that Im working on. \- Data: Minimal footprint as possible. Anything is currently saved only on the device, especially all health data. Check Kalo out here: https://apps.apple.com/de/app/kalo/id6739449751?l=en-GB Tech used to make it possible: There are some terrific security functions in here, and a robust paywall integration, both of which I could never have done without the MVP help of \- Claude and GPT \- Claude's Project function was basically my base project folder here. Claude is perfect when it comes to traditional features. Anything more recent than iOS14 can become a very difficult endeavour \- GPT 4o was great for error logging overview and general sorting measures. Claude's message restriction could be fended of many times here. \- GPT 1o became available more recently and its coding is a lot more robust than 4o. This helped me to not clog Claude with tedious bug fixing. Also it helped when Claude ran away in terrible directions Pre knowledge: I was a digital product designer way back, so I know a thing or two about making things easier to use, especially when it comes to the ease of daily use. Marketing: Will be my biggest focus now. I am quite shit at it, which means It can only get better. It's gonna be some rough weather to get eyes on my app. If anyone thinks they can help or knows how to, any tips are appreciated. Thats it for now. I'll try and keep you updated. I am happy. Let's see if this app will make me happy on a nicer bed, or a jet ski. Again, happy to get your impression of Kalo: https://apps.apple.com/de/app/kalo/id6739449751?l=en-GB

My Side Projects: From CEO to 4th Developer (Thanks, AI 🤖)
reddit
LLM Vibe Score0
Human Vibe Score1
tilopediaThis week

My Side Projects: From CEO to 4th Developer (Thanks, AI 🤖)

Hey Reddit 👋, I wanted to share a bit about some side projects I’ve been working on lately. Quick background for context: I’m the CEO of a mid-to-large-scale eCommerce company pulling in €10M+ annually in net turnover. We even built our own internal tracking software that’s now a SaaS (in early review stages on Shopify), competing with platforms like Lifetimely and TrueROAS. But! That’s not really the point of this post — there’s another journey I’ve been on that I’m super excited to share (and maybe get your feedback on!). AI Transformed My Role (and My Ideas List) I’m not a developer by trade — never properly learned how to code, and to be honest, I don’t intend to. But, I’ve always been the kind of guy who jots down ideas in a notes app and dreams about execution. My dev team calls me their “4th developer” (they’re a team of three) because I have solid theoretical knowledge and can kinda read code. And then AI happened. 🛠️ It basically turned my random ideas app into an MVP generation machine. I thought it’d be fun to share one of the apps I’m especially proud of. I am also planning to build this in public and therefore I am planning to post my progress on X and every project will have /stats page where live stats of the app will be available. Tackling My Task Management Problem 🚀 I’ve sucked at task management for YEARS, I still do! I’ve tried literally everything — Sheets, Todoist, Asana, ClickUp, Notion — you name it. I’d start… and then quit after a few weeks - always. What I struggle with the most is delegating tasks. As a CEO, I delegate a ton, and it’s super hard to track everything I’ve handed off to the team. Take this example: A few days ago, I emailed an employee about checking potential collaboration opportunities with a courier company. Just one of 10s of tasks like this I delegate daily. Suddenly, I thought: “Wouldn’t it be AMAZING if just typing out this email automatically created a task for me to track?” 💡 So… I jumped in. With the power of AI and a few intense days of work, I built a task manager that does just that. But of course, I couldn’t stop there. Research & Leveling It Up 📈 I looked at similar tools like TickTick and Todoist, scraped their G2 reviews (totally legally, promise! 😅), and ran them through AI for a deep SWOT analysis. I wanted to understand what their users liked/didn’t like and what gaps my app could fill. Some of the features people said they were missing didn’t align with the vision for my app (keeping it simple and personal), but I found some gold nuggets: Integration with calendars (Google) Reminders Customizable UX (themes) So, I started implementing what made sense and am keeping others on the roadmap for the future. And I’ve even built for that to, it still doesn’t have a name, however the point is you select on how many reviews of a specific app you want to make a SWOT analysis on and it will do it for you. Example for Todoist in comments. But more on that, some other time, maybe other post ... Key Features So Far: Here’s what’s live right now: ✅ Email to Task: Add an email as to, cc, or bcc — and it automatically creates a task with context, due dates, labels, etc. ✅ WhatsApp Reminders: Get nudged to handle your tasks via WhatsApp. ✅ WhatsApp to Task: Send a message like /task buy groceries — bam, it’s added with full context etc.. ✅ Chrome Extension (work-in-progress): Highlight text on any page, right-click, and send it straight to your task list. Next Steps: Build WITH the Community 👥 Right now, the app is 100% free while still in the early stages. But hey, API calls and server costs aren’t cheap, so pricing is something I’ll figure out with you as we grow. For now, my goal is to hit 100 users and iterate from there. My first pricing idea is, without monthly subscription, I don’t want to charge someone for something he didn’t use. So I am planning on charging "per task", what do you think? Here’s what I have planned: 📍 End of Year Goal: 100 users (starting from… 1 🥲). 💸 Revenue Roadmap: When we establish pricing, we’ll talk about that. 🛠️ Milestones: Post on Product Hunt when we hit 100 users. Clean up my self-written spaghetti code (hire a pro dev for review 🙃). Hire a part-time dev once we hit MRR that can cover its costs. You can check how are we doing on thisisatask.me/stats Other Side Projects I’m Working On: Because… what’s life without taking on too much, right? 😂 Full list of things I’m building: Internal HRM: Not public, tried and tested in-house. Android TV App: Syncs with HRM to post announcements to office TVs (streamlined and simple). Stats Tracker App: Connects to our internal software and gives me real-time company insights. Review Analyzer: Scrapes SaaS reviews (e.g., G2) and runs deep analysis via AI. This was originally for my Shopify SaaS but is quickly turning into something standalone. Coming soon! Mobile app game: secret for now. Let’s Build This Together! Would love it if you guys checked out https://thisisatask.me and gave it a spin! Still super early, super raw, but I’m pumped to hear your thoughts. Also, what’s a must-have task manager feature for you? Anything that frustrates you with current tools? I want to keep evolving this in public, so your feedback is gold. 🌟 Let me know, Reddit! Are you with me? 🙌

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I very rarely have stuff to post on Reddit, but I share how my project is going on, just random stuff, and memes on X. In case few might want to keep up 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2B products beats building B2C products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

ChatPDF and PDF.ai are making millions using open source tech... here's the code
reddit
LLM Vibe Score0
Human Vibe Score1
Level-Thought6152This week

ChatPDF and PDF.ai are making millions using open source tech... here's the code

Why "copy" an existing product? The best SaaS products weren’t the first of their kind - think Slack, Shopify, Zoom, Dropbox, or HubSpot. They didn’t invent team communication, e-commerce, video conferencing, cloud storage, or marketing tools; they just made them better. What is a "Chat with PDF" SaaS? These are AI-powered PDF assistants that let you upload a PDF and ask questions about its content. You can summarize articles, extract key details from a contract, analyze a research paper, and more. To see this in action or dive deeper into the tech behind it, check out this YouTube video. Let's look at the market Made possible by advances in AI like ChatGPT and Retrieval-Augmented Generation (RAG), PDF chat tools started gaining traction in early 2023 and have seen consistent growth in market interest, which is currently at an all-time high (source:google trends) Keywords like "chat PDF" and "PDF AI" get between 1 to 10 million searches every month (source:keyword planner), with a broad target audience that includes researchers, students, and professionals across various industries. Leaders like PDF.ai and ChatPDF have already gained millions of users within a year of launch, driven by the growing market demand, with paid users subscribing at around $20/month. Alright, so how do we build this with open source? The core tech for most PDF AI tools are based on the same architecture. You generate text embeddings (AI-friendly text representations; usually via OpenAI APIs) for the uploaded PDF’s chapters/topics and store them in a vector database (like Pinecone). Now, every time the user asks a question, a similarity search is performed to find the most similar PDF topics from the vector database. The selected topic contents are then sent to an LLM (like ChatGPT) along with the question, which generates a contextual answer! Here are some of the best open source implementations for this process: GPT4 & LangChain Chatbot for large PDF docs by Mayo Oshin MultiPDF Chat App by Alejandro AO PDFToChat by Hassan El Mghari Worried about building signups, user management, payments, etc.? Here are my go-to open-source SaaS boilerplates that include everything you need out of the box: SaaS Boilerplate by Remi Wg Open SaaS by wasp-lang A few ideas to stand out from the noise: Here are a few strategies that could help you differentiate and achieve product market fit (based on the pivot principles from The Lean Startup by Eric Ries): Narrow down your target audience for a personalized UX: For instance, an exam prep assistant for students with study notes and quiz generator; or a document due diligence and analysis tool for lawyers. Add unique features to increase switching cost: You could autogenerate APIs for the uploaded PDFs to enable remote integrations (eg. support chatbot knowledge base); or build in workflow automation features for bulk analyses of PDFs. Offer platform level advantages: You could ship a native mobile/desktop apps for a more integrated UX; or (non-trivial) offer private/offline support by replacing the APIs with local open source deployments (eg. llama for LLM, an embedding model from the MTEB list, and FAISS for vector search). TMI? I’m an ex-AI engineer and product lead, so don’t hesitate to reach out with any questions! P.S. I've started a free weekly newsletter to share open-source/turnkey resources behind popular products (like this one). If you’re a founder looking to launch your next product without reinventing the wheel, please subscribe :)

How should I implement this local ai into my project?
reddit
LLM Vibe Score0
Human Vibe Score1
TrustingmeerkatThis week

How should I implement this local ai into my project?

I’m currently building a system that takes the text content of news articles about cocktail competitions and then attempts to extract a JSON object from it via using phi4 local ai model. I’m developing it alongside Claude in a project and we’ve built so far a series of qualifying questions that is prompted to phi4 and it’s answered are formatted to JSON I’m attempting to one shot each answer with the specific question and content of the article by asking the same question to phi4 3 times and picking majority answer. Then, the flow of questions are conditional so that the ai is provided a set of questions based on previous answers. I’m getting decent results and anecdotally it’s about 50% correct. So I think I need to begin prompt engineering to get better. Except, I’m wondering if there’s a way to automate these iterations a bit? Currently I’m pasting code and results into 01 preview and asking for detailed analysis, then passing this back into Claude for code revisions all manually. I guess I should design an accuracy test (again with ai) across 10 or so random articles at a time and a/b test until we get something we’re happy with? Does anyone else have any suggestions? I also previously attempted to one shot the entire JSON object rather than elect to flow through a bunch of questions except that didn’t work so well and decided to pivot rather than keep trying to optimise it.

I grew my mobile app to 1.4 million downloads
reddit
LLM Vibe Score0
Human Vibe Score1
TechPrimoThis week

I grew my mobile app to 1.4 million downloads

I started developing the app in early 2017, well before the AI era, when mobile apps were at their peak popularity. My idea was to create an app for emotional and psychological support in the form of helpful articles and various quizzes, such as personality assessments and life satisfaction tests. I named the app "Emotional Intelligence" because this keyword showed good ASO potential for positioning at the top of mobile stores. This proved to be accurate, and the app quickly gained traction in terms of downloads. A major problem I faced then was monetization. Unfortunately, in my country, it wasn't possible to sell through Google Play then, so I could only display ads. I started with Google AdMob, earning $2000 monthly after just a few months. The app then got about 1500 organic downloads daily and quickly surpassed 500,000. Three years after launching the app, I decided it was time for branding to build recognition. By combining the words "sentiment" and "intelligence," I came up with "Sintelly." I then pushed the app toward a social network, which differed from the right move. Adding features like discussion forums for problems, likes, and comments would result in even more growth, but the opposite happened. The app started declining, and I began investing in advertising campaigns. I managed to maintain a balance between income and expenses but without any profit. Then COVID-19 hit, and everything went downhill. I had to give up development and find a job as a developer to ensure my livelihood. Two years passed since I gave up, and that's when ChatGPT started gaining popularity. This immediately showed me how to steer the app towards active support for well-being questions. As I'm not an expert in psychology, I found several external psychotherapists who helped me put together CBT therapy, which I then implemented through a chatbot. This is how the new Sintelly app was born, with its main feature being a chatbot system composed of 17 AI agents that adapt to the user and guide them through a five-phase CBT therapy (I'll write a post about the technology). In addition to the agents, I added various exercises and tests to provide better personalization for the user. Initially, I made all of this free, which was also a mistake. I followed the principle of first showing what the app can do and gathering enough new users before starting to charge. I started selling subscriptions at the beginning of July, and since then, the app has had stable growth. If you want a check app, here is the link. Lessons learned: If things are working, don't touch them Start selling immediately upon app release; there's no need to wait Regularly test prices and types of subscriptions Onboarding is the most essential part of the app because most users buy subscriptions during onboarding It's essential to listen to user feedback. From day one, have a website and work on content to generate organic visits and redirect users from the web to the mobile app Stats: Over 1.4 million downloads 4.4 rating Only 40,000 active users (I had a massive loss during the period when I gave up) 280 active subscribers $3000 monthly revenue Next steps: Work on improving the Agent AI approach Setting up email campaigns and transactional emails Introducing in-app and push notifications Introducing gamification Potential for B2B I hope you can extract useful information from my example and avoid repeating my mistakes. I'm interested in your thoughts and if you have any recommendations for the next steps. I'm always looking to learn and improve.

I am building my agency to help founders build AI startups after 2 successful AI SaaS exits and 4 failures
reddit
LLM Vibe Score0
Human Vibe Score1
_Gautam19This week

I am building my agency to help founders build AI startups after 2 successful AI SaaS exits and 4 failures

Hey everyone, I have been building AI products before ChatGPT was launched. In these years, I have managed to launch, scale and exit 2 SaaS products successfully. Today I am launching a new service offering - Query Labs - Helping you build AI agents for your startups. Like all my previous products, I will be building this in public and share my learning along the way. Here's what I have built so far : Microsponsors ( Fail ) My first product ever. I tried to create a marketplace for newsletter writers to find sponsorship opportunity. Got a few very big newsletter listed on the marketplace as well. However, building marketplace is tough. I found it very difficult to bring in sponsors. Ended up shutting it down, AI Query (Exit - Pre revenue ) It was the second half of 2022 and GPT-3 was the most advance AI on the market. I decided to build a tool that can help developers and non-technical folks write SQL queries by just asking in plain english. I got my first taste of success with this. Had a decent offer even before I figured out monetisation. Accepted the offer to focus on my next product which had already started gaining traction AI Excel Bot ( Exit - Revenue Generating ) AI Excel Bot was my wild success. I had worked hard on the SEO for the site, along with the UI / UX to make it the best AI to write excel formulas and general excel task. There was already a large competitor in the market. However, the reality is that you don't need to be the top player. There is always room for multiple players to survive in a large market. You just need to find the good differentiating factor For AI Excel Bot, the differentiator was the chrome extension, that helped users access it anywhere on the internet. Scaled the product to more than 40k users at the time of exit. However, in the end I decided to exit and focus on my software service business that needed more time. Tutore AI ( Fail ) I wanted to build something useful for students to help them learn better. Tutore was my idea to build AI tools for students. I did launch quickly with multiple tools. However, wasn't motivated enough to continue with the grind. I have decided to sell the product. Have had some meetings with potential buyers but didn't agree on price. Prompt Hackers ( 1k users but no revenue ) Prompt Hackers is a directory of AI prompts for all the use cases you can image. I focused a lot on bringing traffic and newsletter subscription from the day 1. I have never had a problem bringing initial set of users to my products. Prompt Hackers was getting close to 20k page views a month. At the same time we had close to 1k newsletter subscribers. Since our target customers were people choosing to use ChatGPT / Bard instead of some specific software for their task, I built a Prompt Generation and Prompt Optimisation AI. Along with this I also created features to build private prompt library. To make the experience even better, I launched a Chrome Extension that helps users access the prompt generation AI and their prompt library while using ChatGPT. However, I couldn't figure out monetisation. I still get close to 4k page views per month with no marketing at all. There are users who use the AI tools and the prompt library feature daily. But, since I couldn't figure out monetisation, I decided to not put time into the project. There you go. These are all the products I have built in the last 3 years. I have been heavy investing myself in the latest tech in LLMs and AI agents. I know the biggest challenge for AI founders is the AI agents and backend pipelines. That's why I am launching Query Labs. To help you build the best AI implementation for your innovative AI startup. I would love to hear feedback from the community. I will be sharing my learning with my new service along the way. Thanks!

I just built footyGPT.com , an AI assistant for football stats, data and data-backed predictions
reddit
LLM Vibe Score0
Human Vibe Score1
footyamigoThis week

I just built footyGPT.com , an AI assistant for football stats, data and data-backed predictions

Hey Amigos! I am based in London and I build lots of fun tools and websites in the sports data space to help sports bettors make better data-driven decisions. One of my recently launched ones is footyGPT.com I have tried to launch this earlier in 2023 but my team were not able to achieve the end goal (I am a non-technical founder) Funny story, I bought the domain footygpt(dot)com back in 2023 and let it expire because i felt like since we weren't able to acheieve the end goal, there was no need to keep the domain. But 2 months ago I started working on it one more time with a different team, and since AI has become "more friendly" than it was early in 2023, it was a much smoother ride and i was so scared someone else had already bought the domain, only to find out it was still available to buy! I was so scared I lost such a clean domain. Below is an example of how footyGPT works: Ask FootyGPT for Predictions for a Given Fixture how footyGPT works \(case study\) Ask FootyGPT to Scrape and Summarize Data from a Stats Website using footyGPT to scrape football stats websites Ask FootyGPT for Sentiments and Online Opinion About a Given Match or Events. ask footygpt for up to date sentiments and opinions about football matches Ask FootyGPT to Find Teams with Certain Streaks, Trends, or Stats https://reddit.com/link/1frdk04/video/9oyaj5ldqjrd1/player If you are into football (soccer) please give it a spin and let me know your thoughts! Cheers Amigos

What are Boilerplates?
reddit
LLM Vibe Score0
Human Vibe Score1
Inner_Lengthiness697This week

What are Boilerplates?

What are Boilerplates? Boilerplate originally referred to the rolled steel used to make boilers for steam engines in the 19th century. Over time, the term evolved to describe any standardized piece of text or code that can be reused without significant changes. Interest in SaaS has been on the rise, and many more people now want to build products. However, building products from scratch takes a lot of time, and it can be extremely frustrating. Enter SaaS Boilerplates With the standardization of stacks and basic systems that govern SaaS tools, it has become evident that there was a need, and the time was ripe for SaaS Boilerplates. SaaS Boilerplates come with landing pages, website components, authentication modules, payment modules, and various other standard features that can save developers a significant amount of time and cost. The market is flooded with Boilerplates for various tech stacks, such as NextJS, Laravel, Swift, NuxtJS, and so forth. Pros and Cons of Boilerplates Pros Save a significant amount of time and money Reduce frustration for developers as the redundant tasks are taken care of Boilerplates often follow best practices For anywhere between $49 and $299, they provide terrific value for those looking to build something very quickly Most importantly, Boilerplates also enable aspiring founders and builders with limited technical resources or abilities to ship their products faster and more cheaply. They are beacons of hope for non-technical founders looking to build a product quickly. Cons Limited flexibility May become outdated fairly quickly Setting them up still requires time Similar landing pages and design themes can make the product look like a clone Marc Lou’s Shipfast For most of us, Marc Lou popularized the idea of SaaS Boilerplate. Marc Lou launched Shipfast in August 2023. He had built 27 projects prior to this and Shipfast was nothing but all his basic code organised properly. At that time, there were no solid NextJS boilerplates, and Shipfast just took off. He got traction via Product Hunt, Twitter and Hacker News and soon Shipfast went viral. Shipfast now generates $130K/mo, just 9 months after its launch. Marc has been building Shipfast in public, which has led to a lot of interest in SaaS Boilerplates. The market is now flooded with boilerplates for every major tech stack. Marc reaped the benefits of the first mover’s advantage as well as the social proof via his Shipfast community. I don’t think any other boilerplates are as successful as Shipfast, but there are quite a few good ones out there. Shipixen* has grossed over $20K in the 5 months Makerkit* does \~$3500/mo Moreover, there are many open-source boilerplates available for popular stacks such as NextJS. The Evolution of Boilerplates Boilerplates are quickly turning into no-code/low-code code generation tools. For instance, Shipixen allows you to generate custom code for landing pages, waitlist pages and blogs using a simple User Interface. Boilerplates are perfectly posied to sit between code and no-code. Allow the flexibility of code with the interface of a no-code tool — that will be the core value proposition of SaaS boilerplates. Should you build a Boilerplate? Well, the market is flooded, but I believe there’s still an opportunity to leverage boilerplates. You can build boilerplates for certain types of apps or tools, such as Chrome extensions Boilerplates can act as a great lead funnel for building out a great productized services business No-code/low-code code generation boilerplates can become a big thing if you can help build complex tools Niche tech stack boilerplates may still be lucrative Known strategies for successfully building a boilerplate 👇🏻 Shipfast thrives because of social proof and community SaaSRock generates most of its traffic from its Gumroad listings and blogs Usenextbase and Shipixen are being built in public Many boilerplates start with waitlists They have a very clear value proposition around saving time and cost Design & No-Code Boilerplates Here is the corrected version with improved grammar and clarity: While SaaS (code) boilerplates have become fairly popular, other types of boilerplates are emerging in the market, such as design boilerplates and no-code boilerplates. To be honest, design boilerplates have been around for a while. You will find numerous landing page packs, component libraries, and so forth. Makers are now building kits that leverage standard libraries and technologies such as Tailwind CSS, Daisy UI, and more. Nick Buzz from the famous baked.design has this *50 Landing Page Design Kit* in Tailwind CSS & Figma which is wildly popular. Lastly, there is a trend of no-code boilerplates as well. Mohit is building a Bubble Boilerplate for the popular no-code platform — Bubble. All in all, I think that people want to build products and build them fast. Boilerplates help them save a significant amount of time and cost. More importantly, boilerplates are impulse purchases for people who have not shipped but who want to ship. Introducing BuilderKit.ai We have been building AI SaaS tools for quite a while now. 10+ products across text, image, speech, RAG — we have built em all. We figured that it seems easy but actually building these so called AI Wrappers can be time consuming and frustrating — there is a lot of nuance to it. So we built BuidlerKit.ai — a NextJS SaaS Boilerpalte It takes care of everything from landing pages, authentication, dashboarding, emails, SEO to payments — everything that you need to build your tool. It also comes with 8+ production-ready apps. Moreover, the BuilderKit community is an exclusive community of AI SaaS builders (Pro Only Access) The Pre Orders are now live at https://www.builderkit.ai (First 100 Customers get $100 Off — I think we have already done \~20 odd orders since the announcement yesterday, Grab your seat asap!) Starter Plan $49, Pro Plan @ $99

Things I did to promote my product, and how they turned out
reddit
LLM Vibe Score0
Human Vibe Score1
laike9mThis week

Things I did to promote my product, and how they turned out

(I will share more updates in the future, you can find me on Twitter and/or Mastodon) Ask any ten indie developers about the toughest part of their job, and nine will likely say "marketing." I recently got a taste of this firsthand when I launched Xylect. Here's a rundown of my promotional attempts - hopefully, my experiences can help fellow developers out there. Podcast Community (✅ Success) I kicked things off by promoting Xylect in my podcast listener group. It wasn't a blockbuster, but I managed to sell a few copies and got some invaluable feedback from friends. Shoutout to those early supporters! Reddit r/macapps (✅ Success) Having had some luck promoting open-source projects on Reddit before, I decided to make r/macapps my first stop in the English-speaking world. I made an app to help you automate boring tasks with one click This post turned out to be a hit! I sold about ten copies and got a ton of useful feedback. Users pointed out compatibility issues with PopClip and suggested improvements for the website. One Italian user even requested localization, which I happily added. https://preview.redd.it/y4fuwh6hleqd1.png?width=959&format=png&auto=webp&s=7bb1b68cbf8a4f94998999e0832b9b7bd85bac67 https://preview.redd.it/8uu4cmyhleqd1.png?width=683&format=png&auto=webp&s=8f1744636aee8074b0e7491a334ef06076b143b0 I also got an intriguing email from a French user - more on that later. More Reddit Posts (❌ Failure) Riding high on my r/macapps success, I branched out to r/SideProject, r/Entrepreneur, and r/indiehackers. These subreddits frown upon direct self-promotion, so I took a softer approach with an article: The unexpected emotional cost of being an indiehacker While the article was heartfelt, it fell flat. Across all three posts, I got a grand total of three comments - two of which were complaints about the font size on mobile. Needless to say, I didn't sell a single copy. Hacker News (❌ Failure) As one of the tech world's major forums, I had to give Hacker News a shot. I wasn't too optimistic, given my past experiences there. Posting on HN feels like a mix of luck and dark magic. As expected, my post vanished without a trace - no comments, no sales. I might give it another go someday. If you're curious, you can check out my previous HN submissions. Tools Directory Websites (❌ Failure) These sites have a simple premise: you list your app, they display it. Seemed like an easy way to get some backlinks, right? Well, I learned the hard way that it's not that simple. I stumbled upon a Reddit post where someone claimed to have made a killing with their directory site in just a few days. The catch? Each listing cost $19. The site had a handful of apps listed, so I thought, "Why not? Early bird gets the worm." I paid up and listed Xylect. Spoiler alert: all I got was $19 poorer 🥲 Lesson learned: These directory sites won't magically sell your product. At best, they're just glorified backlinks. There might be some value in paid promotions on these platforms, but I can't speak to that from experience. V2EX (❌ Failure) After striking out in the English-speaking world, I turned my attention to the Chinese market, starting with V2EX (think of it as China's hybrid of HN and Reddit). This turned out to be my most unexpected flop. Here's the post: [\[Launch Discount\] Mac's most powerful AI search (Perplexity + Wikipedia + Google), boost your efficiency tenfold with one click. No API key required, no prompt needed, no token limit 🔥 - V2EX](https://www.v2ex.com/t/1064930?p=1#reply36) I'd seen decent engagement on other promo posts, so I had high hopes. I posted late at night (US time) and went to bed dreaming of waking up to a flood of comments. Reality check: The next morning, I had exactly one reply - from Kilerd, a loyal podcast listener showing some love. I was baffled. After re-reading my post, I realized I'd missed a crucial element: promo codes. A quick scan of popular posts confirmed my suspicion. Nearly every successful promo post was offering codes, and most comments were just base64-encoded email addresses. Talk about a facepalm moment. I scrambled to add a note about an upcoming free trial and invited users to drop their emails. This got the ball rolling with some code requests, but by then, the damage was done. The post fizzled out, and I didn't sell a single copy 🫠 A French Friend's Newsletter (✅ Success) At this point, my promotional efforts were looking pretty grim. My sales chart had a depressing stretch of flatline. But then, a glimmer of hope appeared in my inbox. Remember that French user I mentioned earlier? He ran a newsletter called vvmac and offered to feature Xylect if I added French support and sent him a free license. It was an offer I couldn't refuse. What followed was a crash course in French localization (thank you, Claude!) and the start of an incredible partnership. This guy was the most thorough beta tester I've ever encountered. We exchanged over sixty emails, covering everything from translations to UI tweaks to bug fixes. His response time was lightning-fast - I'd fix a bug, and five minutes later, he'd confirm it was sorted. The result? A much-improved Xylect and a glowing feature in his newsletter. https://preview.redd.it/ylcq2wxoleqd1.png?width=991&format=png&auto=webp&s=ee395110f50417d5c7f61318f27bf3dc30247809 I'm still in awe of his dedication. He single-handedly transformed Xylect from a buggy mess into a polished product. I'll be forever grateful for his help. The newsletter feature led to a few more sales, but honestly, that felt like a bonus at that point. Influencers (❌ Failure) I knew from the start that to really make waves, I'd need influencer backing. So, I added a note offering free licenses to content creators willing to collaborate. https://preview.redd.it/tyb2m1rqleqd1.png?width=799&format=png&auto=webp&s=56eabf126e772515322595613c546e6ba69fb431 I did get one taker: Hey, I'll be honest, I am not a huge content creator but I think I put a lot of effort in evaluating and figuring out which apps work... So I was wondering if I could get a license in case you are willing to share it. Thank you for considering. Have a great weekend. But I knew I needed to aim higher. With the new French localization, I thought I'd try my luck with some French-speaking Mac YouTubers. I crafted emails highlighting how Xylect could help their French audience with English content. https://preview.redd.it/07oqzemrleqd1.png?width=542&format=png&auto=webp&s=3d160c1d149f28e9029816a277c6ab2496fcd57e After days of silence, I got one reply. It was... not what I was hoping for: Hi, Thank you for your proposal. I can help you to promote your service on Tiktok, Instagram et YouTube, with unique short video. Price for this project is 3500€. Unless I've completely lost my marbles, there's no way I'm dropping 3500€ on promotion. Sure, given their follower count (YouTube: 348K, TikTok: 2.7M, Instagram: 400K), it's not an outrageous ask. For some products, it might even be worth it. But for Xylect? No way. I also reached out to a Chinese influencer on Xiaohongshu, but they weren't interested. Back to the drawing board. Conclusion If you've made it this far, you've probably realized this isn't exactly a success story. My search for effective promotional channels came up largely empty-handed. I'd naively thought that my success with open-source projects would translate seamlessly to the indie dev world. Boy, was I wrong. As I mentioned in my previous article, open-source projects create a dynamic where users feel indebted to developers for their free labor. But in the commercial world of indie development, that dynamic completely flips. While this experience was often frustrating, it was also enlightening - which was kind of the point. As my first foray into indie development, my main goal was to learn the ropes and understand the process. Making money would've been nice, sure, but it wasn't my primary focus. Thanks for sticking with me through this post. I will share more updates in the future, you can follow me on  Twitter and/or Mastodon.

How I went from $27 to $3K as a solopreneur still in a 9-5
reddit
LLM Vibe Score0
Human Vibe Score1
jottrledThis week

How I went from $27 to $3K as a solopreneur still in a 9-5

My journey started back in November 2023. I was scrolling through Twitter and YouTube and saw a word that I had never come across before. Solopreneur. The word caught my eye. Mainly because I was pretty sure I knew what it meant even though it's not a word you'll find in the dictionary. I liked what it was describing. A solo entrepreneur. A one man business. It completely resonated with me. As a software engineer by trade I'm used to working alone, especially since the pandemic hit and we were forced to work remotely. See, I always wanted to ditch the 9-5 thing but thought that was too big and too scary for a single person to do. Surely you would need a lot of money to get started, right? Surely you would need investors? The whole concept seemed impossible to me. That was until I found all the success stories. I became obsessed with the concept of solopreneurship. As I went further down the rabbit hole I found people like Justin Welsh, Kieran Drew and Marc Louvion to name a few. All of whom have one person businesses making huge money every year. So I thought, if they can do it, why can't I? People like this have cleared the pathway for those looking to escape the 9-5 grind. I decided 2024 would be the year I try this out. My main goal for the year? Build a one man business, earn my first $ online and learn a sh\*t ton along the way. My main goal in general? Build my business to $100K per year, quit my 9-5 and live with freedom. From December 2023 to February 2024 I began brainstorming ideas. I was like a lost puppy looking for his ball. How on earth did people find good ideas? I began writing everything and anything that came to mind down in my notes app on my phone. By February I would have approximately 70 ideas. Each as weird and whacky as the other. I was skeptical though. If I went through all the trouble of building a product for one of these ideas how would I know if anyone would even be interested in using it? I got scared and took a break for a week. All these ideas seemed too big and the chance that they would take off into the atmosphere was slim (in my mind anyways). I was learning more and more about solopreneurship as the weeks went on so I decided to build a product centered around everything I was learning about. The idea was simple. Enter a business idea and use AI to give the user details about how to market it, who their target customers were, what to write on their landing page, etc. All for a measly $27 per use. I quickly built it and launched on March 3rd 2024. I posted about it on Indie Hackers, Reddit and Hacker News. I was so excited about the prospect of earning my first internet $! Surely everyone wanted to use my product! Nope...all I got was crickets. I was quickly brought back down to earth. That was until 5 days later. I looked at my phone and had a new Stripe notification! Cha-ching! My first internet $. What a feeling! That was goal number 1 complete. It would be another 6 days before I would get my second sale...and then another 15 days to get my third. It was an emotional rollercoaster. I went from feeling like quitting the 9-5 was actually possible to thinking that maybe the ups and downs aren't worth it. On one hand I had made my first internet dollar so I should my ecstatic, and don't get me wrong, I was but I wanted more. More validation that I could do this long term. By May I was starting to give up on the product. I had learned so much in the past few months about marketing, SEO, building an audience, etc. and I wanted to build something that I thought could have more success so I focused on one critical thing that I had learned about. What was it? Building a product that had SEO potential. A product that I knew hundreds of people were looking for. See this was my thinking - If I could find a keyword that people were searching for on Google hundreds/thousands of times every month and it was easy to rank high on search engines then I would go all in (in SEO land this equates to a Keyword that has a Keyword Difficulty of = 500). I began researching and found that the keyword "micro saas ideas" was being searched for around 600 times each month. Micro Saas was something that really interested me. It was perfect for solopreneurs. Small software products that 1 person could build. What's not to like if you're in the game of software and solopreneurship? Researching keywords like this became like a game for me. I was hooked. I was doing it every day, finding gems that were being searched for hundreds and thousands of times every month that still had potential. That's when I came up with my next product idea. I decided to create a database of Micro Saas Ideas all with this sort of SEO potential. See if you can build a product that you know people are looking for then that's all the validation you need. So I put this theory to the test. I created a database of Micro Saas Ideas with SEO Potential and launched it in June 2024. This time it was different. I made $700 in the first week of launching. A large contrast to my previous failed attempt at becoming the worlds greatest solopreneur. Since launch I have grown the product to $3K and I couldn't be happier. I know what you're saying, $3K isn't a lot. But it's validation. It's validation that I can earn $ online. Validation that I can grow a business and it gives me hope that one day I'll be able to quit that 9-5 grind. My plan is to keep growing the business. I expect there to be a few challenges up ahead but I'll tackle them as I go and learn from the failures and successes. I have a newsletter where I share Micro Saas Ideas with SEO potential every week which I'll leave below in the first comment. Feel free to come along for the ride. If not I hope this post brings you some value If you're thinking about starting as a solopreneur, stop thinking and start doing, you won't regret it.

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I very rarely have stuff to post on Reddit, but I share how my project is going on, just random stuff, and memes on X. In case few might want to keep up 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2B products beats building B2C products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

How me and my team made 15+ apps and not made a single sale in 2023
reddit
LLM Vibe Score0
Human Vibe Score0.818
MichaelbetterecycleThis week

How me and my team made 15+ apps and not made a single sale in 2023

Hey, my name is Michael, I am in Auckland NZ. This year was the official beginning of my adult life. I graduated from university and started a full-time job. I’ve also really dug into indiehacking/bootstrapping and started 15 projects (and it will be at least 17 before the year ends). I think I’ve learned a lot but I consciously repeated mistakes. Upto (Nov) Discord Statuses + Your Location + Facebook Poke https://preview.redd.it/4nqt7tp2tf5c1.png?width=572&format=png&auto=webp&s=b0223484bc54b45b5c65e0b1afd0dc52f9c02ad1 This was the end of uni, I often messaged (and got messaged) requests of status and location to (and from my) friends. I thought, what if we make a social app that’s super basic and all it does is show you where your friends are? To differentiate from snap maps and others we wanted something with more privacy where you select the location. However, never finished the codebase or launched it. This is because I slowly started to realize that B2C (especially social networks) are way too hard to make into an actual business and the story with Fistbump would repeat itself. However, this decision not to launch it almost launched a curse on our team. From that point, we permitted ourselves to abandon projects even before launching. Lessons: Don’t do social networks if your goal is 10k MRR ASAP. If you build something to 90% competition ship it or you will think it’s okay to abandon projects Insight Bites (Nov) Youtube Summarizer Extension &#x200B; https://preview.redd.it/h6drqej4tf5c1.jpg?width=800&format=pjpg&auto=webp&s=0f211456c390ac06f4fcb54aa51f9d50b0826658 Right after Upto, we started ideating and conveniently the biggest revolution in the recent history of tech was released → GPT. We instantly began ideating. The first problem we chose to use AI for is to summarize YouTube videos. Comical. Nevertheless, I am convinced we have had the best UX because you could right-click on a video to get a slideshow of insights instead of how everyone else did it. We dropped it because there was too much competition and unit economics didn’t work out (and it was a B2C). PodPigeon (Dec) Podcast → Tweet Threads https://preview.redd.it/0ukge245tf5c1.png?width=2498&format=png&auto=webp&s=23303e1cab330578a3d25cd688fa67aa3b97fb60 Then we thought, to make unit economics work we need to make this worthwhile for podcasters. This is when I got into Twitter and started seeing people summarize podcasts. Then I thought, what if we make something that converts a podcast into tweets? This was probably one of the most important projects because it connected me with Jason and Jonaed, both of whom I regularly stay in contact with and are my go-to experts on ideas related to content creation. Jonaed was even willing to buy Podpigeon and was using it on his own time. However, the unit economics still didn’t work out (and we got excited about other things). Furthermore, we got scared of the competition because I found 1 - 2 other people who did similar things poorly. This was probably the biggest mistake we’ve made. Very similar projects made 10k MRR and more, launching later than we did. We didn’t have a coherent product vision, we didn’t understand the customer well enough, and we had a bad outlook on competition and a myriad of other things. Lessons: I already made another post about the importance of outlook on competition. Do not quit just because there are competitors or just because you can’t be 10x better. Indiehackers and Bootstrappers (or even startups) need to differentiate in the market, which can be via product (UX/UI), distribution, or both. Asking Ace Intro.co + Crowdsharing &#x200B; https://preview.redd.it/0hu2tt16tf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3d397568ef2331e78198d64fafc1a701a3e75999 As I got into Twitter, I wanted to chat with some people I saw there. However, they were really expensive. I thought, what if we made some kind of crowdfunding service for other entrepreneurs to get a private lecture from their idols? It seemed to make a lot of sense on paper. It was solving a problem (validated via the fact that Intro.co is a thing and making things cheaper and accessible is a solid ground to stand on), we understood the market (or so we thought), and it could monetize relatively quickly. However, after 1-2 posts on Reddit and Indiehackers, we quickly learned three things. Firstly, no one cares. Secondly, even if they do, they think they can get the same information for free online. Thirdly, the reasons before are bad because for the first point → we barely talked to people, and for the second people → we barely talked to the wrong people. However, at least we didn’t code anything this time and tried to validate via a landing page. Lessons Don’t give up after 1 Redditor says “I don’t need this” Don’t be scared to choose successful people as your audience. Clarito Journaling with AI analyzer https://preview.redd.it/8ria2wq6tf5c1.jpg?width=1108&format=pjpg&auto=webp&s=586ec28ae75003d9f71b4af2520b748d53dd2854 Clarito is a classic problem all amateur entrepreneurs have. It’s where you lie to yourself that you have a real problem and therefore is validated but when your team asks you how much you would pay you say I guess you will pay, maybe, like 5 bucks a month…? Turns out, you’d have to pay me to use our own product lol. We sent it off to a few friends and posted on some forums, but never really got anything tangible and decided to move away. Honestly, a lot of it is us in our own heads. We say the market is too saturated, it’ll be hard to monetize, it’s B2C, etc. Lessons: You use the Mom Test on other people. You have to do it yourself as well. However, recognizing that the Mom Test requires a lot of creativity in its investigation because knowing what questions to ask can determine the outcome of the validation. I asked myself “Do I journal” but I didn’t ask myself “How often do I want GPT to chyme in on my reflections”. Which was practically never. That being said I think with the right audience and distribution, this product can work. I just don’t know (let alone care) about the audience that much (and I thought I was one of them)/ Horns & Claw Scrapes financial news texts you whether you should buy/sell the stock (news sentiment analysis) &#x200B; https://preview.redd.it/gvfxdgc7tf5c1.jpg?width=1287&format=pjpg&auto=webp&s=63977bbc33fe74147b1f72913cefee4a9ebec9c2 This one we didn’t even bother launching. Probably something internal in the team and also seemed too good to be true (because if this works, doesn’t that just make us ultra-rich fast?). I saw a similar tool making 10k MRR so I guess I was wrong. Lessons: This one was pretty much just us getting into our heads. I declared that without an audience it would be impossible to ship this product and we needed to start a YouTube channel. Lol, and we did. And we couldn’t even film for 1 minute. I made bold statements like “We will commit to this for at least 1 year no matter what”. Learnery Make courses about any subject https://preview.redd.it/1nw6z448tf5c1.jpg?width=1112&format=pjpg&auto=webp&s=f2c73e8af23b0a6c3747a81e785960d4004feb48 This is probably the most “successful” project we’ve made. It grew from a couple of dozen to a couple of hundred users. It has 11 buy events for $9.99 LTD (we couldn’t be bothered connecting Stripe because we thought no one would buy it anyway). However what got us discouraged from seriously pursuing it more is, that this has very low defensibility, “Why wouldn’t someone just use chatGPT?” and it’s B2C so it’s hard to monetize. I used it myself for a month or so but then stopped. I don’t think it’s the app, I think the act of learning a concept from scratch isn’t something you do constantly in the way Learnery delivers it (ie course). I saw a bunch of similar apps that look like Ass make like 10k MRR. Lessons: Don’t do B2C, or if you do, do it properly Don’t just Mixpanel the buy button, connect your Stripe otherwise, it doesn’t feel real and you won’t get momentum. I doubt anyone (even me) will make this mistake again. I live in my GPT bubble where I make assumptions that everyone uses GPT the same way and as much as I do. In reality, the argument that this has low defensibility against GPT is invalid. Platforms that deliver a differentiated UX from ChatGPT to audiences who are not tightly integrated into the habit of using ChatGPT (which is like - everyone except for SOME tech evangelists). CuriosityFM Make podcasts about any subject https://preview.redd.it/zmosrcp8tf5c1.jpg?width=638&format=pjpg&auto=webp&s=d04ddffabef9050050b0d87939273cc96a8637dc This was our attempt at making Learnery more unique and more differentiated from chatGPT. We never really launched it. The unit economics didn’t work out and it was actually pretty boring to listen to, I don’t think I even fully listened to one 15-minute episode. I think this wasn’t that bad, it taught us more about ElevenLabs and voice AI. It took us maybe only 2-3 days to build so I think building to learn a new groundbreaking technology is fine. SleepyTale Make children’s bedtime stories https://preview.redd.it/14ue9nm9tf5c1.jpg?width=807&format=pjpg&auto=webp&s=267e18ec6f9270e6d1d11564b38136fa524966a1 My 8-year-old sister gave me that idea. She was too scared of making tea and I was curious about how she’d react if she heard a bedtime story about that exact scenario with the moral that I wanted her to absorb (which is that you shouldn’t be scared to try new things ie stop asking me to make your tea and do it yourself, it’s not that hard. You could say I went full Goebbels on her). Zane messaged a bunch of parents on Facebook but no one really cared. We showed this to one Lady at the place we worked from at Uni and she was impressed and wanted to show it to her kids but we already turned off our ElevenLabs subscription. Lessons: However, the truth behind this is beyond just “you need to be able to distribute”. It’s that you have to care about the audience. I don’t particularly want to build products for kids and parents. I am far away from that audience because I am neither a kid anymore nor going to be a parent anytime soon, and my sister still asked me to make her tea so the story didn’t work. I think it’s important to ask yourself whether you care about the audience. The way you answer that even when you are in full bias mode is, do you engage with them? Are you interested in what’s happening in their communities? Are you friends with them? Etc. User Survey Analyzer Big User Survey → GPT → Insights Report Me and my coworker were chatting about AI when he asked me to help him analyze a massive survey for him. I thought that was some pretty decent validation. Someone in an actual company asking for help. Lessons Market research is important but moving fast is also important. Ie building momentum. Also don’t revolve around 1 user. This has been a problem in multiple projects. Finding as many users as possible in the beginning to talk to is key. Otherwise, you are just waiting for 1 person to get back to you. AutoI18N Automated Internationalization of the codebase for webapps This one I might still do. It’s hard to find a solid distribution strategy. However, the idea came from me having to do it at my day job. It seems a solid problem. I’d say it’s validated and has some good players already. The key will be differentiation via the simplicity of UX and distribution (which means a slightly different audience). In the backlog for now because I don’t care about the problem or the audience that much. Documate - Part 1 Converts complex PDFs into Excel https://preview.redd.it/8b45k9katf5c1.jpg?width=1344&format=pjpg&auto=webp&s=57324b8720eb22782e28794d2db674b073193995 My mom needed to convert a catalog of furniture into an inventory which took her 3 full days of data entry. I automated it for her and thought this could have a big impact but there was no distribution because there was no ICP. We tried to find the ideal customers by talking to a bunch of different demographics but I flew to Kazakhstan for a holiday and so this kind of fizzled out. I am not writing this blog post linearity, this is my 2nd hour and I am tired and don’t want to finish this later so I don’t even know what lessons I learned. Figmatic Marketplace of high-quality Figma mockups of real apps https://preview.redd.it/h13yv45btf5c1.jpg?width=873&format=pjpg&auto=webp&s=aaa2896aeac2f22e9b7d9eed98c28bb8a2d2cdf1 This was a collab between me and my friend Alex. It was the classic Clarito where we both thought we had this problem and would pay to fix it. In reality, this is a vitamin. Neither I, nor I doubt Alex have thought of this as soon as we bought the domain. We posted it on Gumroad, sent it to a bunch of forums, and called it a day. Same issue as almost all the other ones. No distribution strategy. However, apps like Mobin show us that this concept is indeed profitable but it takes time. It needs SEO. It needs a community. None of those things, me and Alex had or was interested in. However shortly after HTML → Figma came out and it’s the best plugin. Maybe that should’ve been the idea. Podcast → Course Turns Podcaster’s episodes into a course This one I got baited by Jason :P I described to him the idea of repurposing his content for a course. He told me this was epic and he would pay. Then after I sent him the demo, he never checked it out. Anyhow during the development, we realized that doesn’t actually work because A podcast doesn’t have the correct format for the course, the most you can extract are concepts and ideas, seldom explanations. Most creators want video-based courses to be hosted on Kajabi or Udemy Another lesson is that when you pitch something to a user, what you articulate is a platform or a process, they imagine an outcome. However, the end result of your platform can be a very different outcome to what they had in mind and there is even a chance that what they want is not possible. You need to understand really well what the outcome looks like before you design the process. This is a classic problem where we thought of the solution before the problem. Yes, the problem exists. Podcasters want to make courses. However, if you really understand what they want, you can see how repurposing a podcast isn’t the best way to get there. However I only really spoke to 1-2 podcasters about this so making conclusions is dangerous for this can just be another asking ace mistake with the Redditor. Documate Part 2 Same concept as before but now I want to run some ads. We’ll see what happens. https://preview.redd.it/xb3npj0ctf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3cd4884a29fd11d870d010a2677b585551c49193 In conclusion https://preview.redd.it/2zrldc9dtf5c1.jpg?width=1840&format=pjpg&auto=webp&s=2b3105073e752ad41c23f205dbd1ea046c1da7ff It doesn’t actually matter that much whether you choose to do a B2C, or a social network or focus on growing your audience. All of these can make you successful. What’s important is that you choose. If I had to summarize my 2023 in one word it’s indecision. Most of these projects succeeded for other people, nothing was as fundamentally wrong about them as I proclaimed. In reality that itself was an excuse. New ideas seduce, and it is a form of discipline to commit to a single project for a respectful amount of time. https://preview.redd.it/zy9a2vzdtf5c1.jpg?width=1456&format=pjpg&auto=webp&s=901c621227bba0feb4efdb39142f66ab2ebb86fe Distribution is not just posting on Indiehackers and Reddit. It’s an actual strategy and you should think of it as soon as you think of the idea, even before the Figma designs. I like how Denis Shatalin taught me. You have to build a pipeline. That means a reliable way to get leads, launch campaigns at them, close deals, learn from them, and optimize. Whenever I get an idea now I always try to ask myself “Where can I find 1000s leads in one day?” If there is no good answer, this is not a good project to do now. &#x200B; https://preview.redd.it/2boh3fpetf5c1.jpg?width=1456&format=pjpg&auto=webp&s=1c0d5d7b000716fcbbb00cbad495e8b61e25be66 Talk to users before doing anything. Jumping on designing and coding to make your idea a reality is a satisfying activity in the short term. Especially for me, I like to create for the sake of creation. However, it is so important to understand the market, understand the audience, understand the distribution. There are a lot of things to understand before coding. https://preview.redd.it/lv8tt96ftf5c1.jpg?width=1456&format=pjpg&auto=webp&s=6c8735aa6ad795f216ff9ddfa2341712e8277724 Get out of your own head. The real reason we dropped so many projects is that we got into our own heads. We let the negative thoughts creep in and kill all the optimism. I am really good at coming up with excuses to start a project. However, I am equally as good at coming up with reasons to kill a project. And so you have this yin and yang of starting and stopping. Building momentum and not burning out. I can say with certainty my team ran out of juice this year. We lost momentum so many times we got burnt out towards the end. Realizing that the project itself has momentum is important. User feedback and sales bring momentum. Building also creates momentum but unless it is matched with an equal force of impact, it can stomp the project down. That is why so many of our projects died quickly after we launched. The smarter approach is to do things that have a low investment of momentum (like talking to users) but result in high impact (sales or feedback). Yes, that means the project can get invalidated which makes it more short-lived than if we built it first, but it preserves team life energy. At the end of 2023 here is a single sentence I am making about how I think one becomes a successful indiehacker. One becomes a successful Indiehacker when one starts to solve pain-killer problems in the market they understand, for an audience they care about and consistently engage with for a long enough timeframe. Therefore an unsuccessful Indiehacker in a single sentence is An unsuccessful Indiehacker constantly enters new markets they don’t understand to build solutions for people whose problems they don’t care about, in a timeframe that is shorter than than the time they spent thinking about distribution. However, an important note to be made. Life is not just about indiehacking. It’s about learning and having fun. In the human world, the best journey isn’t the one that gets you the fastest to your goals but the one you enjoy the most. I enjoyed making those silly little projects and although I do not regret them, I will not repeat the same mistakes in 2024. But while it’s still 2023, I have 2 more projects I want to do :) EDIT: For Devs, frontend is always react with vite (ts) and backend is either node with express (ts) or python. For DB either Postgres or mongo (usually Prisma for ORM). For deployment all of it is on AWS (S3, EC2). In terms of libraries/APIs Whisper.cpp is best open source for transcription Obviously the gpt apis Eleven labs for voice related stuff And other random stuff here and there

How to get your first 10 customers with cold email
reddit
LLM Vibe Score0
Human Vibe Score0.905
LieIgnorant6304This week

How to get your first 10 customers with cold email

Cold email is an insane channel for growth, especially for bootstrapped startups as it's very low cost but completely scalable. Yet there's a huge difference between blind cold emailing and crafting personalized outreach for select individuals. The latter is a legit channel which makes many businesses scale in short amounts of time (i.e. see Alex Hormozi’s ‘$100 Million Dollar Offer’). My goal here is to help other founders do what I did but quicker. So you can learn faster. And then teach me something new too. These are the step-by-step lessons I've learnt as a bootstrapped founder, showing you how to use cold email to get your first customers: Find your leads Write engaging email copy Personalize your outreach Send emails Scale up Find your leads This is a key step. Once you figure out exactly who you want to target and where to find them, you'll be printing money. There's a few different ways to go about finding valuable leads. The secret? Keep testing different approaches until you strike gold. First, dedicate some time every day to find and organise leads. Then, keep an eye on your numbers and bounce rates. If something's not working, switch it up. Stick with what's bringing in results and ditch what's not. It's all about staying flexible and learning as you go. Apollo.io is a great starting point as an effective lead source. Their tool allows you to specify filters including job titles, location, company size, industry, keywords, technologies, and revenue. Get specific with your searches to find your ideal customers. Once you have some results you can save and export them, you'll get a list of contact information including name, email, company, LinkedIn, ready to be verified and used. LinkedIn Sales Navigator is another good source. You can either do manual searches or use a scraper to automate the process. The scrapers I'd recommend checking out are FindyMail and Evaboot. As with Apollo, it's best to get very specific with your targeting so you know the prospect will be interested in your offer. BuiltWith is more expensive but ideal if you're targeting competitors. With BuiltWith you can build lists based on what technologies companies are using. For example if you're selling a Shopify app, you'd want to know websites or stores using Shopify, and reach out to them. The best lead sources will always be those that haven't been contacted a lot in the past. If you are able to find places where your target audience uniquely hangs out, and you can get their company website domains, they have the potential to be scrapped, and you have a way to personalize like "I spotted your comment on XYZ website". Once you've got your leads, keep them organized. Set up folders for different niches, countries, company sizes, so you can review what works and what doesn't. One more thing – before you start firing off emails, make sure those addresses are verified. Always use an email verifier to clean up your list and avoid bounces that may affect your sending reputation, and land you in the spam folder. I use Neverbounce for this but there are other tools available. Write engaging email copy Writing a good copy that gets replies is difficult, it changes depending on your offer/audience and nobody knows what's going to work. The best approach is to keep testing different targeting and messaging until you find what works. However, there are some key rules to stick to that I've outlined. For the subject line, keep it short and personalized. Try to write something that sparks interest, and mention the recipients name: Thought you’d like this {{first name}} {{firstName}} - quick question For the email body it's best to use a framework of personalization, offer, then call to action. Personalization is an entire subject in its own right, which I've covered below. In short, a personalized email opener is the best way to grab their attention, and let them know the email is relevant to them and to keep reading. Take it from Alex Hormozi and his $100M Offers playbook – your offer is very important to get right. Make sure your offer hits the mark for your target audience, and get as specific as possible. For example: I built a SaaS shopify app for small ecommerce businesses selling apparel that doubles your revenue in 60-days or your money back. We developed a cold email personalization tool for lead generation agencies that saves hundreds of hours, and can 3x your reply rate. Lastly, the CTA. The goal here isn't to get sign-ups directly from your first email. It's better to ask a brief question about whether the prospect would be interested in learning more. Something very low friction, that warrants a response. Some examples might include: Would you be interested in learning more about this? Can we connect a bit more on this? Mind if I send over a loom I recorded for you? Never send any links in the first email. You've reached out to this person because you have good reason to believe they'd find real value in your offer, and you want to verify if that's the case. After you get one reply, this is a great positive signal and from there you can send a link, book a call, provide a free resource, whatever makes sense based on their response. Personalize your outreach Personalization is one of the most important parts of the process to get right. Your recipient probably receives a multitude of emails every day, how can you make yours stand out, letting them know you've done your research, and that your email is relevant to them? Personalizing each email ensures you get more positive replies, and avoid spam filters, as your email is unique and hasn't been copied and pasted a million times over. The goal is to spark the recipient's interest, and let them know that you're contacting them for good reason. You might mention a recent achievement, blog post or product release that led you to reach out to the prospect specifically. For example: Your post on "Doing Nothing" gave me a good chuckle. Savvy marketing on Cadbury's part. Saw that you've been at Google for just under a year now as a new VP of sales. Spotted that you've got over 7 years of experience in the digital marketing space. Ideally you'll mention something specifically about the prospect or their company that relates to your offer. The downside to personalization is that it's hard to get right, and very time consuming at scale, but totally worth it. Full disclosure, me and my partner Igor just launched our new startup ColdClicks which uses AI to generate hyper-personalized email openers at scale. We built the tool as we were sending hundreds of emails a day, and personalizing every individual email took hours out of our day. ColdClicks automates this process, saving you time and getting you 2-3x more replies. Send emails At this stage you've decided on who you're targeting, you've mined some leads, and written copy. Now it's time to get sending. You can do this manually by copy and pasting each message, but one of the reasons cold email is so powerful is that it's scalable. When you build a process that gets customers, you'll want to send as many emails as you can to your target market. To get started quickly, you can use a mail-merge gmail tool, the best I've used is Maileteor. With Maileteor you upload your lead data to Google sheets, set-up an email template and Mailmetor will send out emails every day automatically. In your template you can define variables including name, company, and personalization to ensure your email is unique for each recipient. Alternatively, you may opt for a more comprehensive tool such as Instantly. Instantly includes unlimited email sending and accounts. There's more initial setup involved as you'll need to set-up Google workspace, buy sending domains, and warm up your email accounts, but when you become familiar with the process you can build a powerful lead generation / customer acquisition machine. Some key points to note, it's very important to warm up any new email accounts you set up. Warmup is the process of gradually establishing a positive reputation with email service providers like Gmail or Yahoo. Make sure to set up DKIM and DMARC on those new email accounts too, to maximise your chances of landing in the inbox. Scale up Once you've found a process that works, good things happen, and it becomes a numbers game. As you get replies and start to see new users signing up, you'll want to scale the process and send more emails. It's straightforward to add new sending accounts in a sending tool like Instantly, and you'll want to broaden your targeting when mining to test new markets. Unfortunately, sending more emails usually comes with a drop in reply rate as you have less time to personalize your messaging for each recipient. This is where ColdClicks shines. The tool allows you to upload thousands of leads and generate perfectly relevant email personalizations for every lead in your list, then export to your favorite sending tool. The examples I listed above in the personalization section were all generated by ColdClicks. Wrapping it up Cold email is an amazing way to validate your product and get new customers. The channel gets a bad rap, but there's a huge difference between blind cold emailing and crafting personalized outreach for individuals who will find value in your product. It's perfect for bootstrapped founders due to its affordability and scalability, and it's the driver of growth for many SaaS businesses. Time to get your first 10 customers! As you start sending, make it a habit to regularly check for new leads. Always experiment with market/messaging, track every campaign so you can learn what's working and iterate, and when you do get positive responses, reply as soon as you can!

I built an app to find who’s interested in your app by monitoring social media
reddit
LLM Vibe Score0
Human Vibe Score0.857
lmcaraigThis week

I built an app to find who’s interested in your app by monitoring social media

Hi everyone! I hope you’re all doing great folks! I’d love to know your thoughts about what I’ve been working on recently! 🙏 If you’re busy or wanna see the app scroll to the bottom to see the video demo, otherwise, continue reading. Very brief presentation of myself first: I’m Marvin, and I live in Florence, Italy, 👋 This year I decided to go all-in on solopreneurship, I’ve been in tech as Software Engineer first, and then in Engineering Leadership for 10+ years, I’ve always worked in startups, except for last year, when I was the Director of Engineering at the Linux Foundation. Follow me on X or subscribe to my newsletter if you’re curious about this journey. The vision Most founders start building digital startups because they love crafting and being impactful by helping other people or companies. First-time founders then face reality when they realize that nailing distribution is key. All other founders already learned this, most likely the hard way. The outcome is the same: a great product will unlikely succeed without great distribution. Letting people know about your product should be easier and not an unfair advantage. The following meme is so true, but also quite sad. I wanna help this to change by easing the marketing and distribution part. https://preview.redd.it/g52pz46upqtd1.png?width=679&format=png&auto=webp&s=cf8398a3592f25c05c396bb2ff5d028331a36315 The story behind Distribution is a huge space: lead generation, demand generation, content marketing, social media marketing, cold outreach, etc. I cannot solve everything altogether. A few months ago I was checking the traffic to a job board I own (NextCommit). That's when I noticed that the “baseline” traffic increased by almost 10x. 🤯 I started investigating why. I realized that the monthly traffic from Reddit increased from 10-ish to 350+. Yeah, the job board doesn’t get much traffic in total, but this was an interesting finding. After digging more, it seems that all that increase came from a single Reddit comment: https://www.reddit.com/r/remotework/comments/1crwcei/comment/l5fb1yy/ This is the moment when I realized two things: It’s cool that someone quoted it! Engaging with people on Reddit, even just through comments, can be VERY powerful. And this was just one single comment! https://preview.redd.it/nhxcv4h2qqtd1.png?width=1192&format=png&auto=webp&s=d31905f56ae59426108ddbb61f2d6b668eedf27a Some weeks later I started noticing a few apps like ReplyGuy. These were automatically engaging with Reddit posts identified through keywords. I decided to sign up for the free plan of ReplyGuy to know more, but many things didn’t convince me: One of the keywords I used for my job board was “remote” and that caused a lot of false positives, The generated replies were good as a kickstart, but most of the time they needed to be tuned to sound more like me. The latter is expected. In the end, the platform doesn’t know me, doesn’t know my opinions, doesn’t know my story, etc.. The only valuable feature left for me was identifying the posts, but that also didn’t work well for me due to false positives. I ended up using it after only 15 minutes. I’m not saying they did a poor job, but it was not working well for me. In the end, the product got quite some traction, so it helped confirm there’s interest in that kind of tool. What bothered me was the combination of auto-replies that felt non-authentic. It’s not that I’m against bots, automation is becoming more common, and people are getting used to it. But in this context, I believe bots should act as an extension of ourselves, enhancing our interactions rather than just generating generic responses (like tools such as HeyGen, Synthesia, PhotoAI). I’m not there yet with my app, but a lot can be done. I'd love to reach the point where a user feels confident to automate the replies because they sound as written by themselves. I then decided to start from the same space, helping engage with Reddit posts, for these reasons: I experienced myself that it can be impactful, It aligns with my vision to ease distribution, Some competitors validated that there’s interest in this specific feature and I could use it as a starting point, I’m confident I can provide a better experience even with what I already have. The current state The product currently enables you to: Create multiple projects and assign keywords, Find the posts that are relevant for engagement using a fuzzy match of keywords and post-filtered using AI to avoid false positives, Provide an analysis of each post to assess the best way to engage, Generate a helpful reply that you’d need to review and post. So currently the product is more on the demand gen side, but this is just the beginning. I’m speaking with people from Marketing, Sales, RevOps, and Growth agencies to better understand their lives, struggles, and pain points. This will help me ensure that I build a product that enables them to help users find the products they need. I’m currently looking for up to 10 people to join the closed beta for free. If you’re interested in joining or to get notified once generally available you can do it here! https://tally.so/r/3XYbj4 After the closed beta, I will start onboarding people in batches. This will let me gather feedback, iterate, and provide a great experience to everyone aligned with my vision. I’m not going to add auto-reply unless the conditions I explained above are met or someone convinces me there’s a good reason for doing so. Each batch will probably get bigger with an increasing price until I’m confident about making it generally available. The next steps The next steps will depend on the feedback I get from the customers and the learnings from the discovery calls I’m having. I will talk about future developments in another update, but I have some ideas already. Check out the demo video below, and I'd love to hear your thoughts! ❤️ Oh and BTW, the app is called HaveYouHeard! https://reddit.com/link/1fzsnrd/video/34lat9snpqtd1/player This is the link to Loom in case the upload doesn't work: https://www.loom.com/share/460c4033b1f94e3bb5e1d081a05eedfd

Enhancing Time Management & Journaling with AI: A Hybrid Physical-Digital Approach
reddit
LLM Vibe Score0
Human Vibe Score1
Educational-Sand8635This week

Enhancing Time Management & Journaling with AI: A Hybrid Physical-Digital Approach

Hey everyone! I wanted to share my experience combining AI, physical journaling, and time tracking - and get your thoughts on taking this further. Background: My AI-Enhanced Productivity Journey I recently did an intensive experiment tracking my time down to the minute (as a software engineer juggling multiple projects, Kendo practice, and side hustles). I used Claude/ChatGPT to analyze my patterns and got some fascinating insights about my productivity and habits. The AIs helped me spot patterns I was blind to and asked surprisingly thoughtful questions that made me reflect deeper. What really struck me was how AI turned from just an analysis tool into something like a wise friend who remembers everything and asks the right questions at the right time. This got me thinking about creating a more structured approach. The Hybrid Model Concept I'm exploring an idea that combines: Physical journaling/tracking (for tactile experience and mindfulness) AI-powered digital companion (for insights and reflection) Flexible input methods (write in a notebook, take photos, type, or voice record) The key insight is: while AI can track digital activities, our lives happen both online and offline. Sometimes we're in meetings, reading books, or having coffee with friends. By combining human input with AI analysis, we get both accuracy and insight. How It Would Work: \- Write in your physical journal/planner as usual \- Optionally snap photos or type key points into the app \- AI companion provides: \- Smart comparisons (today vs last week/month/year) \- Pattern recognition ("I notice you're most creative after morning exercise...") \- Thoughtful reflection prompts ("How has your approach to \[recurring challenge\] evolved?") \- Connection-making between entries ("This reminds me of what you wrote about...") What Makes This Different Human Agency: You control what to track and share, maintaining mindfulness AI as Coach: Beyond just tracking, it asks meaningful questions based on your patterns Temporal Intelligence: Helps you see how your behaviors and thoughts evolve over time Flexibility: Works whether you prefer paper, digital, or both Early Insights from My Testing: \- Initial tracking caused some anxiety (couldn't sleep first two nights!) but became natural \- AI feedback varies by tool (Claude more encouraging, ChatGPT more direct) \- The combination of manual tracking + AI analysis led to better self-awareness \- Having AI ask unexpected questions led to deeper insights than solo journaling Questions for the Community: Have you tried combining AI with traditional productivity/journaling methods? What worked/didn't? What kinds of AI-generated insights/questions would be most valuable to you? How would you balance the convenience of automation with the benefits of manual tracking? What features would make this truly useful for your productivity practice? I believe there's something powerful in combining the mindfulness of manual tracking, the wisdom of AI, and the flexibility of modern tools. But I'd love to hear your thoughts and experiences! Looking forward to the discussion! 🤔✍️

I built an AI social monitoring that looks for relevant posts, not just keywords
reddit
LLM Vibe Score0
Human Vibe Score1
Chunky_CheezeThis week

I built an AI social monitoring that looks for relevant posts, not just keywords

Hey everyone! I've been working on a side project that I'm excited to share with you all—it's called BillyBuzz What is BillyBuzz? BillyBuzz is an AI-powered social monitoring tool that helps businesses spot and analyze relevant conversations on social media platforms, starting with Reddit. It surfaces the most promising leads directly to your Slack channels, email, or Discord, so you don't have to spend hours scrolling through threads. Why I Built It I was spending a ton of time searching for relevant posts in niche subreddits for another product I was working to get off the ground. It was not only time-consuming but also distracting (you know how easy it is to fall into a Reddit rabbit hole). I couldn't find any existing tool that did more than basic keyword searches—which wasn't enough, especially if your brand name has multiple meanings (like "Apple"). So, I decided to build BillyBuzz. It uses AI to understand your business, products, target audience, and value proposition, alongside specific keywords you might want to include. This way, it finds posts where you can genuinely contribute by introducing your product. I used BillyBuzz for a previous product launch and managed to grow it to over $80k/month in volume within about 3 months, purely through Reddit engagement. How It Works Add Information About Your Business: Input details about your business and products. Select Subreddits to Monitor: Choose the subreddits relevant to your niche. Receive Timely Alerts: Get notified via Slack, email, or Discord when relevant posts are identified. Features AI-Powered Relevancy Scoring: Goes beyond keywords by understanding the context to identify truly relevant opportunities. Subreddit Tracking: Monitor specific subreddits with AI-recommended keywords tailored to your company's needs. Real-Time Alerts: Checks for new relevant conversations every 15 minutes, so you can engage at the perfect time. Automated Categorization (Coming Soon): The AI will categorize conversations into topics like competitors, customer complaints, and more. Who It's For BillyBuzz is designed for startup founders, growth marketers, and small business owners who are tech-savvy and focused on scaling their operations. If you're looking to save time and engage more effectively with your target audience on social media, this might be up your alley. Looking for Feedback I'm sharing this here because I'd love to get your thoughts, feedback, or any suggestions you might have. If you're interested in checking it out, you can find more info here: https://billybuzz.com. Feel free to ask me anything or share your experiences with similar challenges!

How me and my team made 15+ apps and not made a single sale in 2023
reddit
LLM Vibe Score0
Human Vibe Score0.818
MichaelbetterecycleThis week

How me and my team made 15+ apps and not made a single sale in 2023

Hey, my name is Michael, I am in Auckland NZ. This year was the official beginning of my adult life. I graduated from university and started a full-time job. I’ve also really dug into indiehacking/bootstrapping and started 15 projects (and it will be at least 17 before the year ends). I think I’ve learned a lot but I consciously repeated mistakes. Upto (Nov) Discord Statuses + Your Location + Facebook Poke https://preview.redd.it/4nqt7tp2tf5c1.png?width=572&format=png&auto=webp&s=b0223484bc54b45b5c65e0b1afd0dc52f9c02ad1 This was the end of uni, I often messaged (and got messaged) requests of status and location to (and from my) friends. I thought, what if we make a social app that’s super basic and all it does is show you where your friends are? To differentiate from snap maps and others we wanted something with more privacy where you select the location. However, never finished the codebase or launched it. This is because I slowly started to realize that B2C (especially social networks) are way too hard to make into an actual business and the story with Fistbump would repeat itself. However, this decision not to launch it almost launched a curse on our team. From that point, we permitted ourselves to abandon projects even before launching. Lessons: Don’t do social networks if your goal is 10k MRR ASAP. If you build something to 90% competition ship it or you will think it’s okay to abandon projects Insight Bites (Nov) Youtube Summarizer Extension &#x200B; https://preview.redd.it/h6drqej4tf5c1.jpg?width=800&format=pjpg&auto=webp&s=0f211456c390ac06f4fcb54aa51f9d50b0826658 Right after Upto, we started ideating and conveniently the biggest revolution in the recent history of tech was released → GPT. We instantly began ideating. The first problem we chose to use AI for is to summarize YouTube videos. Comical. Nevertheless, I am convinced we have had the best UX because you could right-click on a video to get a slideshow of insights instead of how everyone else did it. We dropped it because there was too much competition and unit economics didn’t work out (and it was a B2C). PodPigeon (Dec) Podcast → Tweet Threads https://preview.redd.it/0ukge245tf5c1.png?width=2498&format=png&auto=webp&s=23303e1cab330578a3d25cd688fa67aa3b97fb60 Then we thought, to make unit economics work we need to make this worthwhile for podcasters. This is when I got into Twitter and started seeing people summarize podcasts. Then I thought, what if we make something that converts a podcast into tweets? This was probably one of the most important projects because it connected me with Jason and Jonaed, both of whom I regularly stay in contact with and are my go-to experts on ideas related to content creation. Jonaed was even willing to buy Podpigeon and was using it on his own time. However, the unit economics still didn’t work out (and we got excited about other things). Furthermore, we got scared of the competition because I found 1 - 2 other people who did similar things poorly. This was probably the biggest mistake we’ve made. Very similar projects made 10k MRR and more, launching later than we did. We didn’t have a coherent product vision, we didn’t understand the customer well enough, and we had a bad outlook on competition and a myriad of other things. Lessons: I already made another post about the importance of outlook on competition. Do not quit just because there are competitors or just because you can’t be 10x better. Indiehackers and Bootstrappers (or even startups) need to differentiate in the market, which can be via product (UX/UI), distribution, or both. Asking Ace Intro.co + Crowdsharing &#x200B; https://preview.redd.it/0hu2tt16tf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3d397568ef2331e78198d64fafc1a701a3e75999 As I got into Twitter, I wanted to chat with some people I saw there. However, they were really expensive. I thought, what if we made some kind of crowdfunding service for other entrepreneurs to get a private lecture from their idols? It seemed to make a lot of sense on paper. It was solving a problem (validated via the fact that Intro.co is a thing and making things cheaper and accessible is a solid ground to stand on), we understood the market (or so we thought), and it could monetize relatively quickly. However, after 1-2 posts on Reddit and Indiehackers, we quickly learned three things. Firstly, no one cares. Secondly, even if they do, they think they can get the same information for free online. Thirdly, the reasons before are bad because for the first point → we barely talked to people, and for the second people → we barely talked to the wrong people. However, at least we didn’t code anything this time and tried to validate via a landing page. Lessons Don’t give up after 1 Redditor says “I don’t need this” Don’t be scared to choose successful people as your audience. Clarito Journaling with AI analyzer https://preview.redd.it/8ria2wq6tf5c1.jpg?width=1108&format=pjpg&auto=webp&s=586ec28ae75003d9f71b4af2520b748d53dd2854 Clarito is a classic problem all amateur entrepreneurs have. It’s where you lie to yourself that you have a real problem and therefore is validated but when your team asks you how much you would pay you say I guess you will pay, maybe, like 5 bucks a month…? Turns out, you’d have to pay me to use our own product lol. We sent it off to a few friends and posted on some forums, but never really got anything tangible and decided to move away. Honestly, a lot of it is us in our own heads. We say the market is too saturated, it’ll be hard to monetize, it’s B2C, etc. Lessons: You use the Mom Test on other people. You have to do it yourself as well. However, recognizing that the Mom Test requires a lot of creativity in its investigation because knowing what questions to ask can determine the outcome of the validation. I asked myself “Do I journal” but I didn’t ask myself “How often do I want GPT to chyme in on my reflections”. Which was practically never. That being said I think with the right audience and distribution, this product can work. I just don’t know (let alone care) about the audience that much (and I thought I was one of them)/ Horns & Claw Scrapes financial news texts you whether you should buy/sell the stock (news sentiment analysis) &#x200B; https://preview.redd.it/gvfxdgc7tf5c1.jpg?width=1287&format=pjpg&auto=webp&s=63977bbc33fe74147b1f72913cefee4a9ebec9c2 This one we didn’t even bother launching. Probably something internal in the team and also seemed too good to be true (because if this works, doesn’t that just make us ultra-rich fast?). I saw a similar tool making 10k MRR so I guess I was wrong. Lessons: This one was pretty much just us getting into our heads. I declared that without an audience it would be impossible to ship this product and we needed to start a YouTube channel. Lol, and we did. And we couldn’t even film for 1 minute. I made bold statements like “We will commit to this for at least 1 year no matter what”. Learnery Make courses about any subject https://preview.redd.it/1nw6z448tf5c1.jpg?width=1112&format=pjpg&auto=webp&s=f2c73e8af23b0a6c3747a81e785960d4004feb48 This is probably the most “successful” project we’ve made. It grew from a couple of dozen to a couple of hundred users. It has 11 buy events for $9.99 LTD (we couldn’t be bothered connecting Stripe because we thought no one would buy it anyway). However what got us discouraged from seriously pursuing it more is, that this has very low defensibility, “Why wouldn’t someone just use chatGPT?” and it’s B2C so it’s hard to monetize. I used it myself for a month or so but then stopped. I don’t think it’s the app, I think the act of learning a concept from scratch isn’t something you do constantly in the way Learnery delivers it (ie course). I saw a bunch of similar apps that look like Ass make like 10k MRR. Lessons: Don’t do B2C, or if you do, do it properly Don’t just Mixpanel the buy button, connect your Stripe otherwise, it doesn’t feel real and you won’t get momentum. I doubt anyone (even me) will make this mistake again. I live in my GPT bubble where I make assumptions that everyone uses GPT the same way and as much as I do. In reality, the argument that this has low defensibility against GPT is invalid. Platforms that deliver a differentiated UX from ChatGPT to audiences who are not tightly integrated into the habit of using ChatGPT (which is like - everyone except for SOME tech evangelists). CuriosityFM Make podcasts about any subject https://preview.redd.it/zmosrcp8tf5c1.jpg?width=638&format=pjpg&auto=webp&s=d04ddffabef9050050b0d87939273cc96a8637dc This was our attempt at making Learnery more unique and more differentiated from chatGPT. We never really launched it. The unit economics didn’t work out and it was actually pretty boring to listen to, I don’t think I even fully listened to one 15-minute episode. I think this wasn’t that bad, it taught us more about ElevenLabs and voice AI. It took us maybe only 2-3 days to build so I think building to learn a new groundbreaking technology is fine. SleepyTale Make children’s bedtime stories https://preview.redd.it/14ue9nm9tf5c1.jpg?width=807&format=pjpg&auto=webp&s=267e18ec6f9270e6d1d11564b38136fa524966a1 My 8-year-old sister gave me that idea. She was too scared of making tea and I was curious about how she’d react if she heard a bedtime story about that exact scenario with the moral that I wanted her to absorb (which is that you shouldn’t be scared to try new things ie stop asking me to make your tea and do it yourself, it’s not that hard. You could say I went full Goebbels on her). Zane messaged a bunch of parents on Facebook but no one really cared. We showed this to one Lady at the place we worked from at Uni and she was impressed and wanted to show it to her kids but we already turned off our ElevenLabs subscription. Lessons: However, the truth behind this is beyond just “you need to be able to distribute”. It’s that you have to care about the audience. I don’t particularly want to build products for kids and parents. I am far away from that audience because I am neither a kid anymore nor going to be a parent anytime soon, and my sister still asked me to make her tea so the story didn’t work. I think it’s important to ask yourself whether you care about the audience. The way you answer that even when you are in full bias mode is, do you engage with them? Are you interested in what’s happening in their communities? Are you friends with them? Etc. User Survey Analyzer Big User Survey → GPT → Insights Report Me and my coworker were chatting about AI when he asked me to help him analyze a massive survey for him. I thought that was some pretty decent validation. Someone in an actual company asking for help. Lessons Market research is important but moving fast is also important. Ie building momentum. Also don’t revolve around 1 user. This has been a problem in multiple projects. Finding as many users as possible in the beginning to talk to is key. Otherwise, you are just waiting for 1 person to get back to you. AutoI18N Automated Internationalization of the codebase for webapps This one I might still do. It’s hard to find a solid distribution strategy. However, the idea came from me having to do it at my day job. It seems a solid problem. I’d say it’s validated and has some good players already. The key will be differentiation via the simplicity of UX and distribution (which means a slightly different audience). In the backlog for now because I don’t care about the problem or the audience that much. Documate - Part 1 Converts complex PDFs into Excel https://preview.redd.it/8b45k9katf5c1.jpg?width=1344&format=pjpg&auto=webp&s=57324b8720eb22782e28794d2db674b073193995 My mom needed to convert a catalog of furniture into an inventory which took her 3 full days of data entry. I automated it for her and thought this could have a big impact but there was no distribution because there was no ICP. We tried to find the ideal customers by talking to a bunch of different demographics but I flew to Kazakhstan for a holiday and so this kind of fizzled out. I am not writing this blog post linearity, this is my 2nd hour and I am tired and don’t want to finish this later so I don’t even know what lessons I learned. Figmatic Marketplace of high-quality Figma mockups of real apps https://preview.redd.it/h13yv45btf5c1.jpg?width=873&format=pjpg&auto=webp&s=aaa2896aeac2f22e9b7d9eed98c28bb8a2d2cdf1 This was a collab between me and my friend Alex. It was the classic Clarito where we both thought we had this problem and would pay to fix it. In reality, this is a vitamin. Neither I, nor I doubt Alex have thought of this as soon as we bought the domain. We posted it on Gumroad, sent it to a bunch of forums, and called it a day. Same issue as almost all the other ones. No distribution strategy. However, apps like Mobin show us that this concept is indeed profitable but it takes time. It needs SEO. It needs a community. None of those things, me and Alex had or was interested in. However shortly after HTML → Figma came out and it’s the best plugin. Maybe that should’ve been the idea. Podcast → Course Turns Podcaster’s episodes into a course This one I got baited by Jason :P I described to him the idea of repurposing his content for a course. He told me this was epic and he would pay. Then after I sent him the demo, he never checked it out. Anyhow during the development, we realized that doesn’t actually work because A podcast doesn’t have the correct format for the course, the most you can extract are concepts and ideas, seldom explanations. Most creators want video-based courses to be hosted on Kajabi or Udemy Another lesson is that when you pitch something to a user, what you articulate is a platform or a process, they imagine an outcome. However, the end result of your platform can be a very different outcome to what they had in mind and there is even a chance that what they want is not possible. You need to understand really well what the outcome looks like before you design the process. This is a classic problem where we thought of the solution before the problem. Yes, the problem exists. Podcasters want to make courses. However, if you really understand what they want, you can see how repurposing a podcast isn’t the best way to get there. However I only really spoke to 1-2 podcasters about this so making conclusions is dangerous for this can just be another asking ace mistake with the Redditor. Documate Part 2 Same concept as before but now I want to run some ads. We’ll see what happens. https://preview.redd.it/xb3npj0ctf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3cd4884a29fd11d870d010a2677b585551c49193 In conclusion https://preview.redd.it/2zrldc9dtf5c1.jpg?width=1840&format=pjpg&auto=webp&s=2b3105073e752ad41c23f205dbd1ea046c1da7ff It doesn’t actually matter that much whether you choose to do a B2C, or a social network or focus on growing your audience. All of these can make you successful. What’s important is that you choose. If I had to summarize my 2023 in one word it’s indecision. Most of these projects succeeded for other people, nothing was as fundamentally wrong about them as I proclaimed. In reality that itself was an excuse. New ideas seduce, and it is a form of discipline to commit to a single project for a respectful amount of time. https://preview.redd.it/zy9a2vzdtf5c1.jpg?width=1456&format=pjpg&auto=webp&s=901c621227bba0feb4efdb39142f66ab2ebb86fe Distribution is not just posting on Indiehackers and Reddit. It’s an actual strategy and you should think of it as soon as you think of the idea, even before the Figma designs. I like how Denis Shatalin taught me. You have to build a pipeline. That means a reliable way to get leads, launch campaigns at them, close deals, learn from them, and optimize. Whenever I get an idea now I always try to ask myself “Where can I find 1000s leads in one day?” If there is no good answer, this is not a good project to do now. &#x200B; https://preview.redd.it/2boh3fpetf5c1.jpg?width=1456&format=pjpg&auto=webp&s=1c0d5d7b000716fcbbb00cbad495e8b61e25be66 Talk to users before doing anything. Jumping on designing and coding to make your idea a reality is a satisfying activity in the short term. Especially for me, I like to create for the sake of creation. However, it is so important to understand the market, understand the audience, understand the distribution. There are a lot of things to understand before coding. https://preview.redd.it/lv8tt96ftf5c1.jpg?width=1456&format=pjpg&auto=webp&s=6c8735aa6ad795f216ff9ddfa2341712e8277724 Get out of your own head. The real reason we dropped so many projects is that we got into our own heads. We let the negative thoughts creep in and kill all the optimism. I am really good at coming up with excuses to start a project. However, I am equally as good at coming up with reasons to kill a project. And so you have this yin and yang of starting and stopping. Building momentum and not burning out. I can say with certainty my team ran out of juice this year. We lost momentum so many times we got burnt out towards the end. Realizing that the project itself has momentum is important. User feedback and sales bring momentum. Building also creates momentum but unless it is matched with an equal force of impact, it can stomp the project down. That is why so many of our projects died quickly after we launched. The smarter approach is to do things that have a low investment of momentum (like talking to users) but result in high impact (sales or feedback). Yes, that means the project can get invalidated which makes it more short-lived than if we built it first, but it preserves team life energy. At the end of 2023 here is a single sentence I am making about how I think one becomes a successful indiehacker. One becomes a successful Indiehacker when one starts to solve pain-killer problems in the market they understand, for an audience they care about and consistently engage with for a long enough timeframe. Therefore an unsuccessful Indiehacker in a single sentence is An unsuccessful Indiehacker constantly enters new markets they don’t understand to build solutions for people whose problems they don’t care about, in a timeframe that is shorter than than the time they spent thinking about distribution. However, an important note to be made. Life is not just about indiehacking. It’s about learning and having fun. In the human world, the best journey isn’t the one that gets you the fastest to your goals but the one you enjoy the most. I enjoyed making those silly little projects and although I do not regret them, I will not repeat the same mistakes in 2024. But while it’s still 2023, I have 2 more projects I want to do :) EDIT: For Devs, frontend is always react with vite (ts) and backend is either node with express (ts) or python. For DB either Postgres or mongo (usually Prisma for ORM). For deployment all of it is on AWS (S3, EC2). In terms of libraries/APIs Whisper.cpp is best open source for transcription Obviously the gpt apis Eleven labs for voice related stuff And other random stuff here and there

I searched for unexplored AI business opportunities for 2024 and found 8 promising ideas
reddit
LLM Vibe Score0
Human Vibe Score0
yuki_taylorThis week

I searched for unexplored AI business opportunities for 2024 and found 8 promising ideas

https://solansync.beehiiv.com/p/8-innovative-ai-business-opportunities-2024-evaluation-resources Entering 2024, the AI landscape presents numerous uncharted business opportunities. Solan Sync, on February 06, 2024, shared an insightful exploration into nine innovative AI business prospects that stand out for their potential market impact and implementation feasibility. Here's a brief overview of each: No-Code AI Chatbot Development Platforms: These platforms enable businesses to create efficient chatbots without coding knowledge, catering to a variety of communication needs and boasting a significant market potential projected at $19.8 billion by 2027. AI-Powered Document Management Systems: Offering a solution to automate data extraction and management, this opportunity targets sectors overwhelmed by paperwork, with a market growth expected to reach $4.4 billion by 2026. Automated AI Customer Support Platforms: AI-driven platforms are transforming customer support by handling inquiries with advanced conversational agents, aiming for a part of the $15.3 billion market by 2027. AI-Driven Content Generation Platforms: Utilizing advanced language models for content creation, this area addresses the high demand for engaging content across digital platforms, with the market projected to hit $12 billion by 2025. AI-Powered Recommendation System APIs: Tailored product recommendations can significantly enhance user experience, tapping into a market anticipated to grow to $6.3 billion by 2027. AI-Enhanced Digital Media Buying Solutions: These platforms optimize advertising strategies using AI, targeting the native advertising market expected to reach $59 billion by 2025. Enterprise-grade Voice-activated AI Assistants: Improving workplace efficiency with voice commands, this segment has a potential market of $1.1 billion by 2026. AI-Enhanced Supply Chain Management Solutions: By applying AI for real-time optimization, this opportunity aims at improving efficiency within the vast data-rich environments of modern supply chains. Each idea is detailed with its overview, target customer segments, key AI functionalities, profitability evaluations, and examples of current pioneers. This exploration not only highlights the vast potential within AI-driven business models but also encourages entrepreneurs and corporations to delve into these promising sectors. The rapid advancement of AI technology and its practical applications suggest these ideas represent just the beginning of what the future holds. Now is the time to leverage AI's capabilities to innovate and enhance products, services, and operations across industries.

Introducing Stratify: Your Ultimate AI Strategy Builder for Business Success
reddit
LLM Vibe Score0
Human Vibe Score0
vsengarThis week

Introducing Stratify: Your Ultimate AI Strategy Builder for Business Success

Hello, I’m thrilled to announce the launch of my new startup, Stratify! 🔍 What is Stratify? Stratify is an AI Strategy Builder designed to help businesses of all sizes develop, implement, and optimize their strategic plans using cutting-edge artificial intelligence. Whether you're a startup looking to scale or an established company aiming to innovate, Stratify provides the tools and insights you need to stay ahead in today's competitive landscape. 🌟 Key Features: Automated Strategy Development: Leverage AI to analyze market trends, competitor data, and internal metrics to create comprehensive strategic plans tailored to your business goals. Real-Time Analytics & Insights: Monitor your strategy's performance with real-time data dashboards, enabling you to make informed decisions quickly. Scenario Planning: Use AI-driven simulations to forecast different business scenarios and understand potential outcomes, helping you prepare for uncertainties. Collaborative Tools: Facilitate team collaboration with integrated communication features, ensuring everyone is aligned and contributing to the strategy development process. Customizable Templates: Access a library of industry-specific strategy templates that can be customized to fit your unique business needs. 💡 Why Stratify? In today's fast-paced business environment, creating and adapting effective strategies can be challenging. Many companies struggle with data overload, lack of actionable insights, and inefficient planning processes. Stratify addresses these pain points by harnessing the power of AI to streamline strategy building, making it more efficient, data-driven, and adaptable. 🚀 Our Journey So Far: Founded: August 2024 Milestones Achieved: Developed and tested our MVP with a select group of beta users What's Next: Launching our public beta in Q4 2024 Expanding our feature set based on user feedback Growing our team with experts in AI, business strategy, and customer success 🤝 How You Can Help: We’re eager to connect with early adopters, business strategists, and industry experts who can benefit from or contribute to Stratify. Here’s how you can get involved: Join Our Beta Program: Be among the first to experience Stratify and provide valuable feedback. Share Your Insights: Help us refine our features by sharing your business strategy challenges and needs. Spread the Word: If you know someone who could benefit from an AI-driven strategy builder, please share our mission and be an affiliate to earn rewards! 🌐 Learn More: Visit our website at AI-Powered Brand Strategy & Content Creation | Stratify (brandprovoke.com) and follow us for the latest updates and insights. 🙏 Thank You! A heartfelt thank you to the Reddit community for your support and encouragement. We’re excited to embark on this journey and look forward to your feedback and suggestions! Looking forward to your thoughts and questions!

Looking for Innovators to Join my Stealth-Mode AI and Automation Startup
reddit
LLM Vibe Score0
Human Vibe Score1
Content-Shopping8791This week

Looking for Innovators to Join my Stealth-Mode AI and Automation Startup

Hi everyone, I’m currently working on building a stealth-mode startup that focuses on AI, automation, management consulting, and streamlining business processes. Right now, it’s just me working on this, and I’m looking for passionate, creative people to join me and help shape the future of the company. A bit about me: I’m from the UK and have a Business Management degree and an IT diploma, so I’ve got a good mix of business and tech knowledge to push this forward. I’m currently using tools like UiPath, Python, Make.com, Automation Anywhere, and others to create innovative solutions, but I’m not tied to these. I’m open to using any tools or technologies that fit the business and help us succeed. This is unpaid for now, but once we hit revenue targets, the plan is to transition into paid positions. If you’re excited about startups, innovation, and building something meaningful, this might be for you. I’m building AI-powered tools that solve real business problems, workflows to automate processes, and management consulting services to help businesses streamline and work smarter. It’s about combining tech innovation with business strategy to deliver something that really works. I’d love to work with people who have skills in things like Python, TensorFlow, UiPath, Automation Anywhere, web development (frontend, backend, or full-stack), or just a talent for improving workflows. If you’re great at problem-solving, strategy, or even just brainstorming new ideas, there’s a place for you. What’s in it for you? First off, you’ll get real-world experience in AI, automation, and consulting. You’ll also get the chance to help shape the company as part of the founding team and grow with it. Once the startup hits revenue goals, paid roles will follow. It’s flexible too, work remotely and set your own schedule. If this sounds interesting to you, just comment or send me a DM with a bit about your experience, any projects you’ve worked on, and how you think you could contribute to the startup. I’ll be running interviews soon to chat with people and see how we can work together. If you’re excited about joining a startup from the ground up, let’s connect. I’d love to hear from you.

I recreated an AI phone calling agent that increased booked calls by 30% for a plumbing business in 30 days
reddit
LLM Vibe Score0
Human Vibe Score1
Will_feverThis week

I recreated an AI phone calling agent that increased booked calls by 30% for a plumbing business in 30 days

AI has always intrigued me, especially when it comes to automating repetitive tasks and streamlining business operations. Recently, I found a compelling case study about a voice agent that significantly enhanced customer service and lead capture for a plumbing company. Motivated by the potential of this technology, I decided to build a similar system to see how it could be adapted for other industries. I’ve added the case study below along with a number to the demo voice agent I created to see if this is something people would really be interested in. AI technology is advancing rapidly, and I’m excited to dive deeper into this space. Case Study A family-owned plumbing business was facing challenges managing a high volume of customer calls. They were missing potential leads, particularly during after-hours and weekends, which meant lost revenue opportunities. Hiring a dedicated call support team was considered but deemed too expensive and hard to scale. Solution To solve these issues, the company deployed an AI-powered voice agent capable of handling calls autonomously. The system collected essential customer information, identified service needs, and sent real-time alerts to service technicians via SMS. It also had the ability to transfer calls to human agents if necessary, ensuring a seamless experience for customers. Impact The AI voice agent quickly proved its worth by streamlining call management and improving response times. With the AI handling routine inquiries and initial call filtering, the plumbing business saw a noticeable improvement in how quickly they could respond to customer needs. Details The AI-powered voice agent included several advanced features designed to optimize customer service: Answer Calls Anytime: Ensured every call received a friendly and professional response, regardless of the time of day. Spot Emergencies Fast: Quickly identified high-priority issues that required urgent attention. Collect Important Info: Accurately recorded critical customer details to facilitate seamless follow-ups and service scheduling. Send Alerts Right Away: Immediately notified service technicians about emergencies, enabling faster response times. Live Transfers: Live call transfer options when human assistance was needed. Results The AI-powered voice agent delivered measurable improvements across key performance metrics: 100% Call Answer Rate: No missed calls ensured that every customer inquiry was addressed promptly. 5-minute Emergency Response Time: The average response time for urgent calls was reduced significantly. 30% Increase in Lead Capture: The business saw more qualified leads, improving their chances of conversion. 25% Improvement in Resource Efficiency: Better allocation of resources allowed the team to focus on high-priority tasks. By implementing the AI-powered voice agent, the plumbing business enhanced its ability to capture more leads and provide better service to its customers. The improved call handling efficiency helped reduce missed opportunities and boosted overall customer satisfaction. Here’s the number to the demo agent I created: +1 (210) 405-0982 I’d love to hear some honest thoughts on it and which industries you think could benefit the most from this technology.

Acquired our first 10 customer for Trustty Reporter - an AI first Business Intelligence Platform.
reddit
LLM Vibe Score0
Human Vibe Score1
Longjumping-Buddy501This week

Acquired our first 10 customer for Trustty Reporter - an AI first Business Intelligence Platform.

Hi All, My co founder and I have built Trustty Reporter (www.trusttyreporter.com).  We spent the last couple of months working on launch our AI powered BI platform and gain our first 10 users. We wanted to reach out the community to get your feedback on the platform and how we can take it to the next level. Below is a brief introduction of the platform: Trustty Reporter – your AI-first business intelligence partner that transforms data into actionable insights in minutes! Imagine turning complex data and documents into easy-to-understand reports with clear recommendations, all at the click of a button. No more BI complexities—Trustty Reporter makes business insights accessible to everyone, from business owners to CXOs. Here’s Why You’ll Love Trustty Reporter: Instant Insight Generation – Convert raw data into insights in just 5-15 minutes. No expertise needed! Easy Reporting Access – Persistent reports that let you track, compare, and build strategies over time. Tailored Solutions for Business Problems – Just describe your challenge, and Trustty Reporter delivers custom insights. Interactive Reports – Dive deeper with a chat interface that offers further clarification and recommendations. By now you would have realized that this aces any traditional BI tools. That aside, it’s better than the likes of ChatGPT and Claude since you don’t have to supply multiple prompts to get context specific insights catering to your business! File Requirements: For Excel files with multiple sheets/tabs: Please save each sheet as a separate file Upload them as individual files for processing File Format: The first row must contain your column headers Remove any empty rows above the headers https://preview.redd.it/olmk6lfmwuzd1.png?width=3024&format=png&auto=webp&s=aa2bbc8edb4a299dbeee67b692cd4acf1704c2be

I recreated an AI phone calling agent that increased booked calls by 30% for a plumbing business in 30 days
reddit
LLM Vibe Score0
Human Vibe Score1
Will_feverThis week

I recreated an AI phone calling agent that increased booked calls by 30% for a plumbing business in 30 days

AI has always intrigued me, especially when it comes to automating repetitive tasks and streamlining business operations. Recently, I found a compelling case study about a voice agent that significantly enhanced customer service and lead capture for a plumbing company. Motivated by the potential of this technology, I decided to build a similar system to see how it could be adapted for other industries. I’ve added the case study below along with a number to the demo voice agent I created to see if this is something people would really be interested in. AI technology is advancing rapidly, and I’m excited to dive deeper into this space. Case Study A family-owned plumbing business was facing challenges managing a high volume of customer calls. They were missing potential leads, particularly during after-hours and weekends, which meant lost revenue opportunities. Hiring a dedicated call support team was considered but deemed too expensive and hard to scale. Solution To solve these issues, the company deployed an AI-powered voice agent capable of handling calls autonomously. The system collected essential customer information, identified service needs, and sent real-time alerts to service technicians via SMS. It also had the ability to transfer calls to human agents if necessary, ensuring a seamless experience for customers. Impact The AI voice agent quickly proved its worth by streamlining call management and improving response times. With the AI handling routine inquiries and initial call filtering, the plumbing business saw a noticeable improvement in how quickly they could respond to customer needs. Details The AI-powered voice agent included several advanced features designed to optimize customer service: Answer Calls Anytime: Ensured every call received a friendly and professional response, regardless of the time of day. Spot Emergencies Fast: Quickly identified high-priority issues that required urgent attention. Collect Important Info: Accurately recorded critical customer details to facilitate seamless follow-ups and service scheduling. Send Alerts Right Away: Immediately notified service technicians about emergencies, enabling faster response times. Live Transfers: Live call transfer options when human assistance was needed. Results The AI-powered voice agent delivered measurable improvements across key performance metrics: 100% Call Answer Rate: No missed calls ensured that every customer inquiry was addressed promptly. 5-minute Emergency Response Time: The average response time for urgent calls was reduced significantly. 30% Increase in Lead Capture: The business saw more qualified leads, improving their chances of conversion. 25% Improvement in Resource Efficiency: Better allocation of resources allowed the team to focus on high-priority tasks. By implementing the AI-powered voice agent, the plumbing business enhanced its ability to capture more leads and provide better service to its customers. The improved call handling efficiency helped reduce missed opportunities and boosted overall customer satisfaction. Here’s the number to the demo agent I created: +1 (210) 405-0982 I’d love to hear some honest thoughts on it and which industries you think could benefit the most from this technology.

I acquired a SaaS for ~5 figures to solve my content problem
reddit
LLM Vibe Score0
Human Vibe Score1
Either_Discussion635This week

I acquired a SaaS for ~5 figures to solve my content problem

In 2023 I bought a SaaS called Cuppa AI. I actually found the product on twitter, run by a very talented engineer in the UK.  I’ve spent tens of thousands of dollars on content for various media companies. In one consumer health company, it cost us around $200-$500 for each SEO optimized article. This adds up pretty quickly. Not forgetting the 20 hours of edits! This isn’t just an isolated problem for a single company. It’s industry wide and affects small business + agency owners alike. I spent over a decade in media, and have seen many agency founders complain about long lead times and high costs for low output.  This is an issue. Large swathes of would-be customers that prefer to consume content before buying are being ignored - either because it takes too long or costs too much for founders to scale this channel.   I eventually became tired of the media content game in 2022 and looked into using SaaS to solve my previous life’s challenges. I started building, acquiring and scaling a portfolio of products that I found useful in my day to day. But the content issue was still there.  So I started to look for ways to reduce the time + cost content burden for my own portfolio.   I initially discovered Cuppa using it for my own personal pains of content research, editing, publishing, and scaling. But then I saw potential. I wanted to turn it into an end to end solution for the content gap that myself and other business owners weren’t taking advantage of because of time, cost, or other priorities.  I sent a DM. Then a few calls later, I acquired it in June 2023.  I chose cuppa vs other competing products for a few reasons:  The founder gave excellent support during and post acquisition  It already had a large, loyal existing user base I’d personally used it and solved a pain with it. I saw the potential to solve many others for more people like me  The founder has put a ton of quality and care into it. There wasn’t a risk of picking up a patchy product, plus it already had great social distribution  It naturally fits my expertise from the ‘other side’. I was the original customer of it, so I knew I could evolve it with features that could create content at scale without losing the human touch  Since then we’ve added a lot of new stuff: Chat with articles Image generation for articles API keys to reduce cost Brand / persona voice custom prompts  Month on month iterative content improvement  Full stack content team that blends AI and human editors for agencies I’m still in full build mode with the team. I want to take it to a place where agencies and SMB owners can trust the AI + human content model enough to see this product as a no-brainer for their biz. I don’t believe in AI slop - there’s enough of that out there - I DO believe in using AI to do the grunt work, but to always have that human element a machine can’t quite mimic.  We have a lot more to get through, but I’m very excited about it. View of the done for you content workflow

What I learn from my $200 MRR App I built 4 months ago
reddit
LLM Vibe Score0
Human Vibe Score0.857
ricky0603This week

What I learn from my $200 MRR App I built 4 months ago

4 month ago, I am just a 10-years experienced product manager without any software development experience. I have an $3K/month job, but I am so tired, I don’t like my life, don’t like my boss, don’t like my daily work, that make me feeling I already died however I am still living. I yearn for freedom and want to live each day the way I want to. So I quit my job, and become a Indie developer to build my own business, my own app, even my own life. I am so grateful for this time and experience, now my app reach $200 MRR, still very little compared to my previous salary, but I never regret. I have learned lots of things from this time and experience, more than I had in last 10 years. Here is the time-line of my App: &#x200B; Sep 2023: Launch first version to iOS App store Oct 2023: Release in-app-purchase features and have first subscriber, the revenue in October is $154 Nov 2023: Change from subscription to pay per use, and I did lots of marketing jobs in November, however, the revenue reduced to only $40. Dec 2023: Change back to subscription, and stop some invalid marketing jobs, only keep the ones that actually work. I almost did nothing in December, and the revenue come to $243. During this process, I have learned lots of things, there are some of them that I think could help you as well. Web or App My App is an iOS app that only can running on Apple’s device such like iPhone/iPad or Mac with Apple silicon. Many people ask me why my product is an iOS app not a website, because they don’t have any Apple device. It's true that promoting an app is much harder than promoting a website. However I am now very glad I made an App and not a website! If I make a website, I don't think it's possible to make $100 in the first month. My App is about keyword research, to help people find some ideas from search keyword, because every keyword people searched in Google are representing a real need of them, also can be used in SEO field. However there are a lot of website tools about keyword research, some of them are famous like Ahrefs, SEMrush… I have no intention of competing with them. Actually I don’t have any chance. While in app store, there are little apps about keyword research, each of them have terrible data and user experience, that means if my app has better data and experience that could be my chance. In fact, the App store brings me 20 organic installs a day that Google would never have been able to bring me if I had a website, at least for the first few months. Furthermore, Apple nearly did everything for developer, I don’t need to care about user login, payment and so on, Apple did everything, I just need to call their API, that save lots of time, if I build a website, I need to implement login and payment by myself, that would add some extra work. Not to mention I'd need to buy servers and domains, that would cost me a lot of money. Although Apple will take 30% of the revenue, I can live with that in the early stages because the most important thing for me is to get the product to market as soon as possible. Actually thought Apple’s SMB program, the take rate is 15% now. So Web or App is not important in the early stage, time is important, if people need my product, it's easy to make a website one. More Users or More Valuable Users In November, I notice some users would like use my app, and they were meet paywall, but they never subscribe. I provided 7 day free trail, but it seem that they don’t like it. So I decide to change subscription to pay per use. Because as a user, I don’t like subscription as well, pay per use seem like more friendly. So I change from subscription to pay per use. People can afford $9.99 to subscribe monthly for unlimited use or pay $1.99 for each data they want(First purchase is $0.99 then $1.99). I was expecting more user to pay, but it was the complete opposite! Some users who would have paid a higher subscription fee are switching to a lower priced single payment. Users are encountering paywalls more often, and each time they need to make a decision about whether or not to pay, which increases the probability that they will abandon payment. This resulted in a 75% decrease in revenue in November. In fact, the mostly of my revenue comes from a handful of long-cycle subscribers, such as annual subscription. Few bring in most of the revenue, that is the most important thing I learned. You don't need a lot of customers, you just need more valuable ones. That's why it's only right to design a mechanism to filter out high-value customers and focus on them, all the things you want do is just let more people into the filter, and from that point of view, subscription with free trial period is the best way, even if most people don't like it. The rule of 20/80 will always be there. The most important thing is always focus on the 20 percent things and people. Effort does not always guarantee rewards. Unless one engages in deep thinking, or most efforts are invalid. I have been working very hard to promote my product for a period of time. It’s about in November. I did a lot of job, such as write script to send message to my potential clients on Fiverr, post and write comments on others post on Reddit, find related questions and answer them on Quora, post and comments on Twitte, etc. During that period, I was exhausted every day, but the outcome did not meet my expectations. There is only little growth on App installation, even less revenue than before. That make me frustrated. I finally realized that If I need to put in a tremendous amount of effort just to make a little progress, there is must something wrong. So I stop 80% of promote work I have ever did, only keep app store search ad, which will bring a installation with less than $0.5 cost. Then I dive into long time and deeply thinking, I spent more time on reading books, investigate other product with great MRR, watch interviews with people who are already living the kind of life I aspire to live, for example, u/levelsio. These things have given me great inspiration, and my life has become easier. It seems that the life I anticipated when I resigned is getting closer. I also have a clearer understanding of my app. Meanwhile, MRR has been growing. This experience let me learn that effort does not always guarantee results. Many times, our efforts are just wishful thinking, they are invalid, do the right thing after deeply thinking is more important. What Next? My goal is reach $3K MRR, as same as my job payment, I will never stop to building things, and I will keep my currently lifestyle. I still don't know how to get more people to use my app, but levelsio's interviews give me some inspiration that I can verified something by manually instead of build a software. I plan to launch a trend analysis product based on the keyword data provided by my current app. I have always wanted to combine AI to build such a product, but I didn't know how to do it. Now I intend to manually complete it first and start software development once there are paying users. If you are interested to my App, you could try it. Gotrends

My experience trying to scrape google maps with no code
reddit
LLM Vibe Score0
Human Vibe Score1
youngkilogThis week

My experience trying to scrape google maps with no code

A few months back I was working on a project to help founders that sell to SMBs get better quality leads (Current solutions like Zoominfo and Apollo don’t do very well for the SMB market). Of course, I wanted to do this as quickly as possible with as little code as possible.  We found that people were manually going through Google Maps to find SMBs. They would use the search and manually type in the businesses they were looking for. For example, they would type “restaurants” and manually call/email them. What we decided to do was gather the Google Maps data autonomously and surface that to our customers so they could take all of it. The problem was that we would need a bunch of data from Google Maps to pull it off. We would need to grab all the SMBs across the United States which is a huge undertaking.  Initially, I tried no-code AI web scraping solutions and they worked horribly. For some reason, I couldn’t even get them to scroll down on the page. I was also able to reverse engineer their open-source code and discover that they were taking the entire web page and passing it into GPT to extract data. That just burned my Openai bill.  I then tried the semi-code approach (sorry no-code subreddit) where I would use something like Apify or Google Places API to scrape the businesses. This worked better but still, there was an issue of price at the scale we wanted. Eventually, we ended up writing our scraper for the task.  This experience was so horrible I ended up creating potarix.com. Firstly, we provide scraping as a service in conjunction with AI. We all know AI is shit and keeping this human in the loop allows the AI to do 90% of the work and then for us to tweak the script to 100% completion. Also since we use AI to create the scraper instead of using AI to scrape, we can run it for large scale tasks at a low cost.

How I'm automating all my SEO research & writing with AI by building an open source software
reddit
LLM Vibe Score0
Human Vibe Score1
frazrasThis week

How I'm automating all my SEO research & writing with AI by building an open source software

I make most of my current recurring income from writing articles for a few blogs. Over the years I have developed strategies and writing techniques that increase my chances of landing at the top of Google search results. I’m a writer, but I also write code. With the advent of AI I have been itching to codify many of my previous activities. I tried writing content with the general LLMs like ChatGPT and Claude but the results were terrible, especially for niches with technical information. I didn’t want to lose hope in AI because I realised with A LOT of hand-holding, it got better results. THEN IT HIT ME!  What if I could create a Human-Guided AI for Better AI-Written Articles: enter Building ContentScribe After months of coding with AI tools and trying different approaches, I’m excited to share that ContentScribe is finally taking shape. The journey to this point has been challenging but incredibly rewarding. Over the past six months, we’ve been using ContentScribe ourselves to automate blog content creation. We found other tools in the AI article generation space such as Koala AI and Cuppa that left us wanting more. They basically took a topic from you and let the AI loose. We consider this to be a better Koala AI and Cuppa alternative. I wanted to have more control and freedom from the expense of the credit system most of them use. Even after generation, every article required significant human input to make it truly SEO-friendly, and existing tools couldn’t handle the specific strategies we needed for our niche. So, we decided to build something new: an AI-powered, open-source tool that doesn’t just spit out generic articles, but actually allows users to shape how the content is written. ContentScribe is designed to integrate the SEO techniques that we’ve developed over years of building profitable blogs. It codifies our best practices and turns them into a process that anyone can use to create researched, optimized content, every time. The product works, and it’s live! We’ve been populating our latest blog with human-guided AI-written articles, and the results are already impressive. The coolest part? This project scratches our own itch and addresses the pain points we faced when using other tools. Plus there is nothing to lose because it’s free and open source, you can run it locally or in the cloud. It’s still early days, but I’m excited to share more as we keep building in public. We’re working on tutorials, and adding more features. The feedback we’ve gotten so far from our in-house team has been invaluable, and I’m looking forward to sharing this with more content creators out there. For anyone struggling to get their ideas off the ground: keep experimenting, keep building. ContentScribe is proof that when you combine persistence with innovation, the results can be something you’re genuinely proud of. This is just the beginning!

Free AI Tool Directory 🔥
reddit
LLM Vibe Score0
Human Vibe Score1
Feeling_Run_2556This week

Free AI Tool Directory 🔥

Hey! I built an AI Website Tools Directory – a collection of free AI-powered tools designed to help website owners, marketers, designers, coders, and creators automate repetitive tasks. 💡 Why I Built This: I wanted a place where I can find a comprehensive list AI tools for free. Its a single hub of free AI utilities for website optimization, content creation, productivity and much more. 🚀 What’s Inside? (Use cases) 🔹 For Content Creators & Writers AI Blog Post Generator – Instantly create article ideas & drafts. AI Meta Tag Generator – SEO-optimized titles & descriptions. AI Content Improver – Rewrite & enhance your text. 🔹 For Coders & Developers AI Code Snippet Generator – Get working code snippets for web projects. AI Regex Generator – Easily create regex patterns. AI JSON Formatter – Auto-format & clean up JSON data. 🔹 For Designers & Video Creators AI UX Improvement Tool – Detects website usability & design issues. AI Image Optimizer – Compress & enhance images for better performance. AI Video Title Generator – Get engaging titles for YouTube & Reels. 🔹 For Audio & Music Creators AI Podcast Name Generator – Unique & catchy names for your show. AI Music Genre Classifier – Analyze & tag your music automatically. 🔹 For Productivity & SEO AI Keyword Research Tool – Find trending search terms. AI Headline Analyzer – Optimize titles for engagement. AI Email Subject Line Generator – Boost email open rates. 🔍 Looking for Feedback: What other AI-powered tools would be useful for your work? Any specific features you’d like to see added? 👉 Try it here: https://www.aiwebsitetools.com/ Would love your thoughts! 🚀

Made $3.5k Automating Social Media Posts with AI
reddit
LLM Vibe Score0
Human Vibe Score1
pakshal-codesThis week

Made $3.5k Automating Social Media Posts with AI

"Marketers & creators were spending hours crafting LinkedIn posts & X threads. Built an AI tool that automates the process—here’s how." Backstory A growing startup was struggling to maintain a consistent LinkedIn & X presence. Their team wasted hours every week: Manually drafting posts from raw ideas and reports Figuring out platform-specific formats (hooks, CTAs, structure) Scheduling posts across multiple accounts What I Built in 48 Hours ✅ AI-Powered Post Generator → Open-source LLM (Mistral) formats ideas into optimized LinkedIn/X posts ✅ Engagement Booster → Custom NLP ensures every post follows best practices (hooks, CTA, readability) ✅ Automated Scheduling → FastAPI + React dashboard lets users auto-post across platforms Tech Stack Content Processing: Open-source LLMs (Mistral, Phi-3) + Custom NLP Data Handling: FastAPI backend + PostgreSQL Frontend: React + Tailwind CSS Automation: CRON + Third-party APIs (LinkedIn, X) Results 💡 10x faster content creation (2 hours → 5 minutes per post) 💡 Increased engagement by 3x with AI-optimized copy 💡 $1.5k payout + ongoing $300/month maintenance 💬 "This tool writes better LinkedIn posts than I do—on autopilot!" Biggest Lesson "Most creators don’t lack ideas—they lack execution speed. Simple AI workflows + automation solve 90% of the problem." PSA to Developers Look for boring, repetitive tasks in niche domains like: Personal branding automation Sales outreach personalization E-commerce product descriptions A weekend project could turn into a $5k/month SaaS. What’s the most time-consuming task you’ve automated with AI? 🚀

[P] Building an Reinforcement Learning Agent to play The Legend of Zelda
reddit
LLM Vibe Score0
Human Vibe Score1
DarkAutumnThis week

[P] Building an Reinforcement Learning Agent to play The Legend of Zelda

A year go I started trying to use PPO to play the original Legend of Zelda, and I was able to train a model to beat the first boss after a few months of work. I wanted to share the project just for show and tell. I'd love to hear feedback and suggestions as this is just a hobby project. I don't do this for a living. The code for that lives in the original-design branch of my Triforce repo. I'm currently tinkering with new designs so the main branch is much less stable. Here's a video of the agent beating the first dungeon, which was trained with 5,000,000+ steps. At 38 seconds, you can see it learned that it's invulnerable at the screen edge, and it exploits that to avoid damage from a projectile. At 53 seconds it steps up to avoid damage from an unblockable projectile, even though it takes a -0.06 penalty for moving the wrong way (taking damage would be a larger penalty.) At 55 seconds it walks towards the rock projectile to block it. And so on, lots of little things the model does is easy to miss if you don't know the game inside and out. As a TLDR, here's an early version of my new (single) model. This doesn't make it quite as far, but if you watch closely it's combat is already far better, and is only trained on 320,000 steps (~6% of the steps the first model was trained on). This is pretty far along from my very first model. Original Design I got the original project working using stable-baselines's PPO and default neural network (Shared NatureCNN, I believe). SB was great to get started but ultimately stifling. In the new version of the project I've implemented PPO from scratch with torch with my own simple neural network similar to stable-baseline's default. I'm playing with all kinds of changes and designs now that I have more flexibility and control. Here is my rough original design: Overall Strategy My first pass through this project was basically "imagine playing Zelda with your older sibling telling you where to go and what to do". I give the model an objective vector which points to where I want it to go on the screen (as a bird flies, the agent still had to learn path finding to avoid damage and navigate around the map). This includes either point at the nearest enemy I want it to kill or a NSEW vector if it's supposed to move to the next room. Due a few limitations with stable-baselines (especially around action masking), I ended up training unique models for traversing the overworld vs the dungeon (since they have entirely different tilesets). I also trained a different model for when we have sword beams vs not. In the video above you can see what model is being used onscreen. In my current project I've removed this objective vector as it felt too much like cheating. Instead I give it a one-hot encoded objective (move north to the next room, pickup items, kill enemies, etc). So far it's working quite well without that crutch. The new project also does a much better job of combat even without multiple models to handle beams vs not. Observation/Action Space Image - The standard neural network had a really tough time being fed the entire screen. No amount of training seemed to help. I solved this by creating a viewport around Link that keeps him centered. This REALLY helped the model learn. I also had absolutely zero success with stacking frames to give Link a way to see enemy/projectile movement. The model simply never trained with stable-baselines when I implemented frame stacking and I never figured out why. I just added it to my current neural network and it seems to be working... Though my early experiments show that giving it 3 frames (skipping two in between, so frames curr, curr-3, curr-6) doesn't really give us that much better performance. It might if I took away some of the vectors. We'll see. Vectors - Since the model cannot see beyond its little viewport, I gave the model a vector to the closest item, enemy, and projectile onscreen. This made it so the model can shoot enemies across the room outside of its viewport. My new model gives it multiple enemies/items/projectiles and I plan to try to use an attention mechanism as part of the network to see if I can just feed it all of that data. Information - It also gets a couple of one-off datapoints like whether it currently has sword beams. The new model also gives it a "source" room (to help better understand dungeons where we have to backtrack), and a one-hot encoded objective. Action Space My original project just has a few actions, 4 for moving in the cardinal directions and 4 for attacking in each direction (I also added bombs but never spent any time training it). I had an idea to use masking to help speed up training. I.E. if link bumps into a wall, don't let him move in that direction again until he moves elsewhere, as the model would often spend an entire memory buffer running headlong straight into a wall before an update...better to do it once and get a huge negative penalty which is essentially the same result but faster. Unfortunately SB made it really annoying architecturally to pass that info down to the policy layer. I could have hacked it together, but eventually I just reimplemented PPO and my own neural network so I could properly mask actions in the new version. For example, when we start training a fresh model, it cannot attack when there aren't enemies on screen and I can disallow it from leaving certain areas. The new model actually understands splitting swinging the sword short range vs firing sword beams as two different actions, though I haven't yet had a chance to fully train with the split yet. Frameskip/Cooldowns - In the game I don't use a fixed frame skip for actions. Instead I use the internal ram state of game to know when Link is animation locked or not and only allow the agent to take actions when it's actually possible to give meaningful input to the game. This greatly sped up training. We also force movement to be between tiles on the game map. This means that when the agent decides to move it loses control for longer than a player would...a player can make more split second decisions. This made it easier to implement movement rewards though and might be something to clean up in the future. Other interesting details Pathfinding - To facilitate rewards, the original version of this project used A* to pathfind from link to what he should be doing. Here's a video of it in action. This information wasn't giving to the model directly but instead the agent would only be given the rewards if it exactly followed that path or the transposed version of it. It would also pathfind around enemies and not walk through them. This was a nightmare though. The corner cases were significant, and pushing Link towards enemies but not into them was really tricky. The new verison just uses a wavefront algorithm. I calculate a wave from the tiles we want to get to outwards, then make sure we are following the gradient. Also calculating the A* around enemies every frame (even with caching) was super slow. Wavefront was faster, especially because I give the new model no special rewards for walking around enemies...faster to compute and it has to learn from taking damage or not. Either way, the both the old and new models successfully learned how to pathfind around danger and obstacles, with or without the cheaty objective vector. Rewards - I programmed very dense rewards in both the old and new model. At basically every step, the model is getting rewarded or punished for something. I actually have some ideas I can't wait to try out to make the rewards more sparse. Or maybe we start with dense rewards for the first training, then fine-tune the model with sparser rewards. We'll see. Predicting the Future - Speaking of rewards. One interesting wrinkle is that the agent can do a lot of things that will eventually deal damage but not on that frame. For example, when Link sets a bomb it takes several seconds before it explodes, killing things. This can be a massive reward or penalty since he spent an extremely valuable resource, but may have done massive damage. PPO and other RL propagates rewards backwards, of course, but that spike in reward could land on a weird frame where we took damage or moved in the wrong direction. I probably could have just not solved that problem and let it shake out over time, but instead I used the fact that we are in an emulator to just see what the outcome of every decision is. When planting a bomb, shooting sword beams, etc, we let the game run forward until impact, then rewind time and reward the agent appropriately, continuing on from when we first paused. This greatly speeds up training, even if it's expensive to do this savestate, play forward, restore state. Neural Networks - When I first started this project (knowing very little about ML and RL), I thought most of my time would be tuning the shape of the neural network that we are using. In reality, the default provided by stable-baselines and my eventual reimplemnentation has been enough to make massive progress. Now that I have a solid codebase though, I really want to revisit this. I'd like to see if trying CoordConvs and similar networks might make the viewport unncessary. Less interesting details/thoughts Hyperparameters - Setting the entropy coefficinet way lower helped a TON in training stable models. My new PPO implementation is way less stable than stable-baselines (ha, imagine that), but still converges most of the time. Infinite Rewards - As with all reinforcement learning, if you give some way for the model to get infinite rewards, it will do just that and nothing else. I spent days, or maybe weeks tweaking reward functions to just get it to train and not find a spot on the wall it could hump for infinite rewards. Even just neutral rewards, like +0.5 moving forward and -0.5 for moving backwards, would often result in a model that just stepped left, then right infinitely. There has to be a real reward or punishment (non-neutral) for forward progress. Debugging Rewards - In fact, building a rewards debugger was the only way I made progress in this project. If you are tackling something this big, do that very early. Stable-Retro is pretty great - Couldn't be happier with the clean design for implementing emulation for AI. Torch is Awesome - My early versions heavily used numpy and relied on stable-baselines, with its multiproc parallelization support. It worked great. Moving the project over to torch was night and day though. It gave me so much more flexibility, instant multithreading for matrix operations. I have a pretty beefy computer and I'm almost at the same steps per second as 20 proc stable-retro/numpy. Future Ideas This has already gone on too long. I have some ideas for future projects, but maybe I'll just make them another post when I actually do them. Special Thanks A special thanks to Brad Flaugher for help with the early version of this, Fiskbit from the Zelda1 speedrunning community for help pulling apart the raw assembly to build this thing, and MatPoliquin for maintaining Stable-Retro. Happy to answer any questions, really I just love nerding out about this stuff.

[D] Why I'm Lukewarm on Graph Neural Networks
reddit
LLM Vibe Score0
Human Vibe Score0.6
VodkaHazeThis week

[D] Why I'm Lukewarm on Graph Neural Networks

TL;DR: GNNs can provide wins over simpler embedding methods, but we're at a point where other research directions matter more I also posted it on my blog here, has footnotes, a nicer layout with inlined images, etc. I'm only lukewarm on Graph Neural Networks (GNNs). There, I said it. It might sound crazy GNNs are one of the hottest fields in machine learning right now. [There][1] were at least [four][2] [review][3] [papers][4] just in the last few months. I think some progress can come of this research, but we're also focusing on some incorrect places. But first, let's take a step back and go over the basics. Models are about compression We say graphs are a "non-euclidean" data type, but that's not really true. A regular graph is just another way to think about a particular flavor of square matrix called the [adjacency matrix][5], like this. It's weird, we look at run-of-the-mill matrix full of real numbers and decide to call it "non-euclidean". This is for practical reasons. Most graphs are fairly sparse, so the matrix is full of zeros. At this point, where the non-zero numbers are matters most, which makes the problem closer to (computationally hard) discrete math rather than (easy) continuous, gradient-friendly math. If you had the full matrix, life would be easy If we step out of the pesky realm of physics for a minute, and assume carrying the full adjacency matrix around isn't a problem, we solve a bunch of problems. First, network node embeddings aren't a thing anymore. A node is a just row in the matrix, so it's already a vector of numbers. Second, all network prediction problems are solved. A powerful enough and well-tuned model will simply extract all information between the network and whichever target variable we're attaching to nodes. NLP is also just fancy matrix compression Let's take a tangent away from graphs to NLP. Most NLP we do can be [thought of in terms of graphs][6] as we'll see, so it's not a big digression. First, note that Ye Olde word embedding models like [Word2Vec][7] and [GloVe][8] are [just matrix factorization][9]. The GloVe algorithm works on a variation of the old [bag of words][10] matrix. It goes through the sentences and creates a (implicit) [co-occurence][11] graph where nodes are words and the edges are weighed by how often the words appear together in a sentence. Glove then does matrix factorization on the matrix representation of that co-occurence graph, Word2Vec is mathematically equivalent. You can read more on this in my [post on embeddings][12] and the one (with code) on [word embeddings][13]. Even language models are also just matrix compression Language models are all the rage. They dominate most of the [state of the art][14] in NLP. Let's take BERT as our main example. BERT predicts a word given the context of the rest of the sentence. This grows the matrix we're factoring from flat co-occurences on pairs of words to co-occurences conditional on the sentence's context, like this We're growing the "ideal matrix" we're factoring combinatorially. As noted by [Hanh & Futrell][15]: [...] human language—and language modelling—has infinite statistical complexity but that it can be approximated well at lower levels. This observation has two implications: 1) We can obtain good results with comparatively small models; and 2) there is a lot of potential for scaling up our models. Language models tackle such a large problem space that they probably approximate a compression of the entire language in the [Kolmogorov Complexity][16] sense. It's also possible that huge language models just [memorize a lot of it][17] rather than compress the information, for what it's worth. Can we upsample any graph like language models do? We're already doing it. Let's call a first-order embedding of a graph a method that works by directly factoring the graph's adjacency matrix or [Laplacian matrix][18]. If you embed a graph using [Laplacian Eigenmaps][19] or by taking the [principal components][20] of the Laplacian, that's first order. Similarly, GloVe is a first-order method on the graph of word co-occurences. One of my favorites first order methods for graphs is [ProNE][21], which works as well as most methods while being two orders of magnitude faster. A higher-order method embeds the original matrix plus connections of neighbours-of-neighbours (2nd degree) and deeper k-step connections. [GraRep][22], shows you can always generate higher-order representations from first order methods by augmenting the graph matrix. Higher order method are the "upsampling" we do on graphs. GNNs that sample on large neighborhoods and random-walk based methods like node2vec are doing higher-order embeddings. Where are the performance gain? Most GNN papers in the last 5 years present empirical numbers that are useless for practitioners to decide on what to use. As noted in the [OpenGraphsBenchmark][4] (OGB) paper, GNN papers do their empirical section on a handful of tiny graphs (Cora, CiteSeer, PubMed) with 2000-20,000 nodes. These datasets can't seriously differentiate between methods. Recent efforts are directly fixing this, but the reasons why researchers focused on tiny, useless datasets for so long are worth discussing. Performance matters by task One fact that surprises a lot of people is that even though language models have the best performance in a lot of NLP tasks, if all you're doing is cram sentence embeddings into a downstream model, there [isn't much gained][23] from language models embeddings over simple methods like summing the individual Word2Vec word embeddings (This makes sense, because the full context of the sentence is captured in the sentence co-occurence matrix that is generating the Word2Vec embeddings). Similarly, [I find][24] that for many graphs simple first-order methods perform just as well on graph clustering and node label prediction tasks than higher-order embedding methods. In fact higher-order methods are massively computationally wasteful for these usecases. Recommended first order embedding methods are ProNE and my [GGVec with order=1][25]. Higher order methods normally perform better on the link prediction tasks. I'm not the only one to find this. In the BioNEV paper, they find: "A large GraRep order value for link prediction tasks (e.g. 3, 4);a small value for node classification tasks (e.g.1, 2)" (p.9). Interestingly, the gap in link prediction performance is inexistant for artificially created graphs. This suggests higher order methods do learn some of the structure intrinsic to [real world graphs][26]. For visualization, first order methods are better. Visualizations of higher order methods tend to have artifacts of their sampling. For instance, Node2Vec visualizations tend to have elongated/filament-like structures which come from the embeddings coming from long single strand random walks. See the following visualizations by [Owen Cornec][27] created by first embedding the graph to 32-300 dimensions using a node embedding algorithm, then mapping this to 2d or 3d with the excellent UMAP algorithm, like this Lastly, sometimes simple methods soundly beat higher order methods (there's an instance of it in the OGB paper). The problem here is that we don't know when any method is better than another and we definitely don't know the reason. There's definitely a reason different graph types respond better/worse to being represented by various methods. This is currently an open question. A big part of why is that the research space is inundated under useless new algorithms because... Academic incentives work against progress Here's the cynic's view of how machine learning papers are made: Take an existing algorithm Add some new layer/hyperparameter, make a cute mathematical story for why it matters Gridsearch your hyperparameters until you beat baselines from the original paper you aped Absolutely don't gridsearch stuff you're comparing against in your results section Make a cute ACRONYM for your new method, put impossible to use python 2 code on github (Or no code at all!) and bask in the citations I'm [not][28] the [only one][29] with these views on the state reproducible research. At least it's gotten slightly better in the last 2 years. Sidebar: I hate Node2Vec A side project of mine is a [node embedding library][25] and the most popular method in it is by far Node2Vec. Don't use Node2Vec. [Node2Vec][30] with p=1; q=1 is the [Deepwalk][31] algorithm. Deepwalk is an actual innovation. The Node2Vec authors closely followed the steps 1-5 including bonus points on step 5 by getting word2vec name recognition. This is not academic fraud -- the hyperparameters [do help a tiny bit][32] if you gridsearch really hard. But it's the presentable-to-your-parents sister of where you make the ML community worse off to progress your academic career. And certainly Node2Vec doesn't deserve 7500 citations. Progress is all about practical issues We've known how to train neural networks for well over 40 years. Yet they only exploded in popularity with [AlexNet][33] in 2012. This is because implementations and hardware came to a point where deep learning was practical. Similarly, we've known about factoring word co-occurence matrices into Word embeddings for at least 20 years. But word embeddings only exploded in 2013 with Word2Vec. The breakthrough here was that the minibatch-based methods let you train a Wikipedia-scale embedding model on commodity hardware. It's hard for methods in a field to make progress if training on a small amount of data takes days or weeks. You're disincentivized to explore new methods. If you want progress, your stuff has to run in reasonable time on commodity hardware. Even Google's original search algorithm [initially ran on commodity hardware][34]. Efficiency is paramount to progress The reason deep learning research took off the way it did is because of improvements in [efficiency][35] as well as much better libraries and hardware support. Academic code is terrible Any amount of time you spend gridsearching Node2Vec on p and q is all put to better use gridsearching Deepwalk itself (on number of walks, length of walks, or word2vec hyperparameters). The problem is that people don't gridsearch over deepwalk because implementations are all terrible. I wrote the [Nodevectors library][36] to have a fast deepwalk implementation because it took 32 hours to embed a graph with a measly 150,000 nodes using the reference Node2Vec implementation (the same takes 3min with Nodevectors). It's no wonder people don't gridsearch on Deepwalk a gridsearch would take weeks with the terrible reference implementations. To give an example, in the original paper of [GraphSAGE][37] they their algorithm to DeepWalk with walk lengths of 5, which is horrid if you've ever hyperparameter tuned a deepwalk algorithm. From their paper: We did observe DeepWalk’s performance could improve with further training, and in some cases it could become competitive with the unsupervised GraphSAGE approaches (but not the supervised approaches) if we let it run for >1000× longer than the other approaches (in terms of wall clock time for prediction on the test set) I don't even think the GraphSAGE authors had bad intent -- deepwalk implementations are simply so awful that they're turned away from using it properly. It's like trying to do deep learning with 2002 deep learning libraries and hardware. Your architectures don't really matter One of the more important papers this year was [OpenAI's "Scaling laws"][38] paper, where the raw number of parameters in your model is the most predictive feature of overall performance. This was noted even in the original BERT paper and drives 2020's increase in absolutely massive language models. This is really just [Sutton' Bitter Lesson][39] in action: General methods that leverage computation are ultimately the most effective, and by a large margin Transformers might be [replacing convolution][40], too. As [Yannic Kilcher said][41], transformers are ruining everything. [They work on graphs][6], in fact it's one of the [recent approaches][42], and seems to be one of the more succesful [when benchmarked][1] Researchers seem to be putting so much effort into architecture, but it doesn't matter much in the end because you can approximate anything by stacking more layers. Efficiency wins are great -- but neural net architectures are just one way to achieve that, and by tremendously over-researching this area we're leaving a lot of huge gains elsewhere on the table. Current Graph Data Structure Implementations suck NetworkX is a bad library. I mean, it's good if you're working on tiny graphs for babies, but for anything serious it chokes and forces you to rewrite everything in... what library, really? At this point most people working on large graphs end up hand-rolling some data structure. This is tough because your computer's memory is a 1-dimensional array of 1's and 0's and a graph has no obvious 1-d mapping. This is even harder when we take updating the graph (adding/removing some nodes/edges) into account. Here's a few options: Disconnected networks of pointers NetworkX is the best example. Here, every node is an object with a list of pointers to other nodes (the node's edges). This layout is like a linked list. Linked lists are the [root of all performance evil][43]. Linked lists go completely against how modern computers are designed. Fetching things from memory is slow, and operating on memory is fast (by two orders of magnitude). Whenever you do anything in this layout, you make a roundtrip to RAM. It's slow by design, you can write this in Ruby or C or assembly and it'll be slow regardless, because memory fetches are slow in hardware. The main advantage of this layout is that adding a new node is O(1). So if you're maintaining a massive graph where adding and removing nodes happens as often as reading from the graph, it makes sense. Another advantage of this layout is that it "scales". Because everything is decoupled from each other you can put this data structure on a cluster. However, you're really creating a complex solution for a problem you created for yourself. Sparse Adjacency Matrix This layout great for read-only graphs. I use it as the backend in my [nodevectors][25] library, and many other library writers use the [Scipy CSR Matrix][44], you can see graph algorithms implemented on it [here][45]. The most popular layout for this use is the [CSR Format][46] where you have 3 arrays holding the graph. One for edge destinations, one for edge weights and an "index pointer" which says which edges come from which node. Because the CSR layout is simply 3 arrays, it scales on a single computer: a CSR matrix can be laid out on a disk instead of in-memory. You simply [memory map][47] the 3 arrays and use them on-disk from there. With modern NVMe drives random seeks aren't slow anymore, much faster than distributed network calls like you do when scaling the linked list-based graph. I haven't seen anyone actually implement this yet, but it's in the roadmap for my implementation at least. The problem with this representation is that adding a node or edge means rebuilding the whole data structure. Edgelist representations This representation is three arrays: one for the edge sources, one for the edge destinations, and one for edge weights. [DGL][48] uses this representation internally. This is a simple and compact layout which can be good for analysis. The problem compared to CSR Graphs is some seek operations are slower. Say you want all the edges for node #4243. You can't jump there without maintaining an index pointer array. So either you maintain sorted order and binary search your way there (O(log2n)) or unsorted order and linear search (O(n)). This data structure can also work on memory mapped disk array, and node append is fast on unsorted versions (it's slow in the sorted version). Global methods are a dead end Methods that work on the entire graph at once can't leverage computation, because they run out of RAM at a certain scale. So any method that want a chance of being the new standard need to be able to update piecemeal on parts of the graph. Sampling-based methods Sampling Efficiency will matter more in the future Edgewise local methods. The only algorithms I know of that do this are GloVe and GGVec, which they pass through an edge list and update embedding weights on each step. The problem with this approach is that it's hard to use them for higher-order methods. The advantage is that they easily scale even on one computer. Also, incrementally adding a new node is as simple as taking the existing embeddings, adding a new one, and doing another epoch over the data Random Walk sampling. This is used by deepwalk and its descendants, usually for node embeddings rather than GNN methods. This can be computationally expensive and make it hard to add new nodes. But this does scale, for instance [Instagram][49] use it to feed their recommendation system models Neighbourhood sampling. This is currently the most common one in GNNs, and can be low or higher order depending on the neighborhood size. It also scales well, though implementing efficiently can be challenging. It's currently used by [Pinterest][50]'s recommendation algorithms. Conclusion Here are a few interesting questions: What is the relation between graph types and methods? Consolidated benchmarking like OGB We're throwing random models at random benchmarks without understanding why or when they do better More fundamental research. Heree's one I'm curious about: can other representation types like [Poincarre Embeddings][51] effectively encode directed relationships? On the other hand, we should stop focusing on adding spicy new layers to test on the same tiny datasets. No one cares. [1]: https://arxiv.org/pdf/2003.00982.pdf [2]: https://arxiv.org/pdf/2002.11867.pdf [3]: https://arxiv.org/pdf/1812.08434.pdf [4]: https://arxiv.org/pdf/2005.00687.pdf [5]: https://en.wikipedia.org/wiki/Adjacency_matrix [6]: https://thegradient.pub/transformers-are-graph-neural-networks/ [7]: https://en.wikipedia.org/wiki/Word2vec [8]: https://nlp.stanford.edu/pubs/glove.pdf [9]: https://papers.nips.cc/paper/2014/file/feab05aa91085b7a8012516bc3533958-Paper.pdf [10]: https://en.wikipedia.org/wiki/Bag-of-words_model [11]: https://en.wikipedia.org/wiki/Co-occurrence [12]: https://www.singlelunch.com/2020/02/16/embeddings-from-the-ground-up/ [13]: https://www.singlelunch.com/2019/01/27/word-embeddings-from-the-ground-up/ [14]: https://nlpprogress.com/ [15]: http://socsci.uci.edu/~rfutrell/papers/hahn2019estimating.pdf [16]: https://en.wikipedia.org/wiki/Kolmogorov_complexity [17]: https://bair.berkeley.edu/blog/2020/12/20/lmmem/ [18]: https://en.wikipedia.org/wiki/Laplacian_matrix [19]: http://citeseerx.ist.psu.edu/viewdoc/download;jsessionid=1F03130B02DC485C78BF364266B6F0CA?doi=10.1.1.19.8100&rep=rep1&type=pdf [20]: https://en.wikipedia.org/wiki/Principalcomponentanalysis [21]: https://www.ijcai.org/Proceedings/2019/0594.pdf [22]: https://dl.acm.org/doi/10.1145/2806416.2806512 [23]: https://openreview.net/pdf?id=SyK00v5xx [24]: https://github.com/VHRanger/nodevectors/blob/master/examples/link%20prediction.ipynb [25]: https://github.com/VHRanger/nodevectors [26]: https://arxiv.org/pdf/1310.2636.pdf [27]: http://byowen.com/ [28]: https://arxiv.org/pdf/1807.03341.pdf [29]: https://www.youtube.com/watch?v=Kee4ch3miVA [30]: https://cs.stanford.edu/~jure/pubs/node2vec-kdd16.pdf [31]: https://arxiv.org/pdf/1403.6652.pdf [32]: https://arxiv.org/pdf/1911.11726.pdf [33]: https://en.wikipedia.org/wiki/AlexNet [34]: https://en.wikipedia.org/wiki/Googledatacenters#Original_hardware [35]: https://openai.com/blog/ai-and-efficiency/ [36]: https://www.singlelunch.com/2019/08/01/700x-faster-node2vec-models-fastest-random-walks-on-a-graph/ [37]: https://arxiv.org/pdf/1706.02216.pdf [38]: https://arxiv.org/pdf/2001.08361.pdf [39]: http://incompleteideas.net/IncIdeas/BitterLesson.html [40]: https://arxiv.org/abs/2010.11929 [41]: https://www.youtube.com/watch?v=TrdevFK_am4 [42]: https://arxiv.org/pdf/1710.10903.pdf [43]: https://www.youtube.com/watch?v=fHNmRkzxHWs [44]: https://docs.scipy.org/doc/scipy/reference/generated/scipy.sparse.csr_matrix.html [45]: https://docs.scipy.org/doc/scipy/reference/sparse.csgraph.html [46]: https://en.wikipedia.org/wiki/Sparsematrix#Compressedsparserow(CSR,CRSorYaleformat) [47]: https://en.wikipedia.org/wiki/Mmap [48]: https://github.com/dmlc/dgl [49]: https://ai.facebook.com/blog/powered-by-ai-instagrams-explore-recommender-system/ [50]: https://medium.com/pinterest-engineering/pinsage-a-new-graph-convolutional-neural-network-for-web-scale-recommender-systems-88795a107f48 [51]: https://arxiv.org/pdf/1705.08039.pdf

[D] Misuse of Deep Learning in Nature Journal’s Earthquake Aftershock Paper
reddit
LLM Vibe Score0
Human Vibe Score0.333
milaworldThis week

[D] Misuse of Deep Learning in Nature Journal’s Earthquake Aftershock Paper

Recently, I saw a post by Rajiv Shah, Chicago-based data-scientist, regarding an article published in Nature last year called Deep learning of aftershock patterns following large earthquakes, written by scientists at Harvard in collaboration with Google. Below is the article: Stand Up for Best Practices: Misuse of Deep Learning in Nature’s Earthquake Aftershock Paper The Dangers of Machine Learning Hype Practitioners of AI, machine learning, predictive modeling, and data science have grown enormously over the last few years. What was once a niche field defined by its blend of knowledge is becoming a rapidly growing profession. As the excitement around AI continues to grow, the new wave of ML augmentation, automation, and GUI tools will lead to even more growth in the number of people trying to build predictive models. But here’s the rub: While it becomes easier to use the tools of predictive modeling, predictive modeling knowledge is not yet a widespread commodity. Errors can be counterintuitive and subtle, and they can easily lead you to the wrong conclusions if you’re not careful. I’m a data scientist who works with dozens of expert data science teams for a living. In my day job, I see these teams striving to build high-quality models. The best teams work together to review their models to detect problems. There are many hard-to-detect-ways that lead to problematic models (say, by allowing target leakage into their training data). Identifying issues is not fun. This requires admitting that exciting results are “too good to be true” or that their methods were not the right approach. In other words, it’s less about the sexy data science hype that gets headlines and more about a rigorous scientific discipline. Bad Methods Create Bad Results Almost a year ago, I read an article in Nature that claimed unprecedented accuracy in predicting earthquake aftershocks by using deep learning. Reading the article, my internal radar became deeply suspicious of their results. Their methods simply didn’t carry many of the hallmarks of careful predicting modeling. I started to dig deeper. In the meantime, this article blew up and became widely recognized! It was even included in the release notes for Tensorflow as an example of what deep learning could do. However, in my digging, I found major flaws in the paper. Namely, data leakage which leads to unrealistic accuracy scores and a lack of attention to model selection (you don’t build a 6 layer neural network when a simpler model provides the same level of accuracy). To my earlier point: these are subtle, but incredibly basic predictive modeling errors that can invalidate the entire results of an experiment. Data scientists are trained to recognize and avoid these issues in their work. I assumed that this was simply overlooked by the author, so I contacted her and let her know so that she could improve her analysis. Although we had previously communicated, she did not respond to my email over concerns with the paper. Falling On Deaf Ears So, what was I to do? My coworkers told me to just tweet it and let it go, but I wanted to stand up for good modeling practices. I thought reason and best practices would prevail, so I started a 6-month process of writing up my results and shared them with Nature. Upon sharing my results, I received a note from Nature in January 2019 that despite serious concerns about data leakage and model selection that invalidate their experiment, they saw no need to correct the errors, because “Devries et al. are concerned primarily with using machine learning as [a] tool to extract insight into the natural world, and not with details of the algorithm design.” The authors provided a much harsher response. You can read the entire exchange on my github. It’s not enough to say that I was disappointed. This was a major paper (it’s Nature!) that bought into AI hype and published a paper despite it using flawed methods. Then, just this week, I ran across articles by Arnaud Mignan and Marco Broccardo on shortcomings that they found in the aftershocks article. Here are two more data scientists with expertise in earthquake analysis who also noticed flaws in the paper. I also have placed my analysis and reproducible code on github. Standing Up For Predictive Modeling Methods I want to make it clear: my goal is not to villainize the authors of the aftershocks paper. I don’t believe that they were malicious, and I think that they would argue their goal was to just show how machine learning could be applied to aftershocks. Devries is an accomplished earthquake scientist who wanted to use the latest methods for her field of study and found exciting results from it. But here’s the problem: their insights and results were based on fundamentally flawed methods. It’s not enough to say, “This isn’t a machine learning paper, it’s an earthquake paper.” If you use predictive modeling, then the quality of your results are determined by the quality of your modeling. Your work becomes data science work, and you are on the hook for your scientific rigor. There is a huge appetite for papers that use the latest technologies and approaches. It becomes very difficult to push back on these papers. But if we allow papers or projects with fundamental issues to advance, it hurts all of us. It undermines the field of predictive modeling. Please push back on bad data science. Report bad findings to papers. And if they don’t take action, go to twitter, post about it, share your results and make noise. This type of collective action worked to raise awareness of p-values and combat the epidemic of p-hacking. We need good machine learning practices if we want our field to continue to grow and maintain credibility. Link to Rajiv's Article Original Nature Publication (note: paywalled) GitHub repo contains an attempt to reproduce Nature's paper Confrontational correspondence with authors

[P] How I found & fixed 4 bugs in Microsoft's Phi-4 model
reddit
LLM Vibe Score0
Human Vibe Score1
danielhanchenThis week

[P] How I found & fixed 4 bugs in Microsoft's Phi-4 model

Hey r/MachineLearning! Last week, Microsoft released Phi-4, a 14B open-source model that rivals OpenAI's GPT-4-o-mini. I managed to find & fix 4 bugs impacting its output quality. You might remember me previously from fixing 8 bugs in Google's Gemma model! :) I'm going to walk you through how I found & fixed the bugs. Phi-4's benchmarks were amazing, however many users reported weird or just wrong outputs. Since I maintain the open-source project called 'Unsloth' (fine-tuning LLMs 2x faster with 70% less VRAM) with my brother, I firstly tested Phi-4 for inference and found many errors. Our GitHub repo: https://github.com/unslothai/unsloth This time, the model had no implementation issues (unlike Gemma 2) but did have problems in the model card. For my first inference run, I randomly found an extra token which is obviously incorrect (2 eos tokens is never a good idea). Also during more runs, I found there was an extra assistant prompt which is once again incorrect. And, lastly, from past experience with Unsloth's bug fixes, I already knew fine-tuning was wrong when I read the code. These bugs caused Phi-4 to have some drop in accuracy and also broke fine-tuning runs. Our fixes are now under review by Microsoft to be officially added to Hugging Face. We uploaded the fixed versions to https://huggingface.co/unsloth/phi-4-GGUF Here’s a breakdown of the bugs and their fixes: Tokenizer bug fixes The Phi-4 tokenizer interestingly uses as the BOS (beginning of sentence), EOS (end of sentence) and PAD (padding) tokens. The main issue is the EOS token is wrong - it should be . Otherwise, you will get in generations. Fine-tuning bug fixes The padding token should be a designated pad token like in Llama () or we can use an untrained token - for example we use , fixing infinite generations and outputs. Chat template issues The Phi-4 tokenizer always adds an assistant prompt - it should only do this if prompted by add\generation\prompt. Most LLM serving libraries expect non auto assistant additions, and this might cause issues during serving. We dive deeper into the bugs in our blog: https://unsloth.ai/blog/phi4 Do our Fixes Work? Yes! Our fixed Phi-4 uploads show clear performance gains, with even better scores than Microsoft's original uploads on the Open LLM Leaderboard. https://preview.redd.it/d8hew26e06ce1.png?width=2366&format=png&auto=webp&s=173c23feacc625566271470839fe7a5e25eb860e Some redditors even tested our fixes to show greatly improved results in: Example 1: Multiple-choice tasks https://preview.redd.it/qx50pkq706ce1.png?width=1579&format=png&auto=webp&s=437da2cabdbf98ef5a8b8cbdc5592907a20e2316 Example 2: ASCII art generation https://preview.redd.it/sw1o3a3yt4de1.png?width=2326&format=png&auto=webp&s=fc6bfc45d14134d45f332ba58bbd1de049f5776b We also made a Colab notebook fine-tune Phi-4 completely for free using Google's free Tesla T4 (16GB) GPUs: https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Phi\4-Conversational.ipynb Thank you for reading this long post and hope you all found this insightful! If you have any questions, please feel free to ask! :) How I found the bugs: I first downloaded the original Phi-4 from https://huggingface.co/microsoft/phi-4, and tested inference out. Weirdly I found assistant to be appended at the even with addgenerationprompt = False in Hugging Face, so I theorized there was a chat template problem. Adding assistant prompts by default can break serving libraries. And yes, https://huggingface.co/microsoft/phi-4/blob/f957856cd926f9d681b14153374d755dd97e45ed/tokenizer\config.json#L774 had by default added the assistant prompt - I first fixed this! I then found ` to be used for the BOS, EOS and PAD tokens, which is a common issue amongst models - I ignored the BOS, since Phi-4 did not have one anyways, but changed the PAD token to `. You can select any of the tokens since they're empty and not trained. This counteracts issues of infinite generations during finetuning. For Llama-fication, I used torch.allclose to confirm all tensors are in fact equivalent. I also used some fake random data to check all activations are also mostly similar bitwise. I also uploaded the model to the HF Open LLM Leaderboard to confirm if the original Phi-4 arch and the new Llama-fied models are equivalent. Finally I verified all finetuning runs with Unsloth in a Colab Notebook to confirm all runs were correct.

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.
reddit
LLM Vibe Score0
Human Vibe Score0.6
AlexSnakeKingThis week

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.

TD;LR: At Company A, Team X does advanced analytics using on-prem ERP tools and older programming languages. Their tools work very well and are designed based on very deep business and domain expertise. Team Y is a new and ambitious Data Science team that thinks they can replace Team X's tools with a bunch of R scripts and a custom built ML platform. Their models are simplistic, but more "fashionable" compared to the econometric models used by Team X, and team Y benefits from the ML/DS moniker so leadership is allowing Team Y to start a large scale overhaul of the analytics platform in question. Team Y doesn't have the experience for such a larger scale transformation, and is refusing to collaborate with team X. This project is very likely going to fail, and cause serious harm to the company as a whole financially and from a people perspective. I argue that this is not just because of bad leadership, but also because of various trends and mindsets in the DS community at large. Update (Jump to below the line for the original story): Several people in the comments are pointing out that this just a management failure, not something due to ML/DS, and that you can replace DS with any buzz tech and the story will still be relevant. My response: Of course, any failure at an organization level is ultimately a management failure one way or the other. Moreover, it is also the case that ML/DS when done correctly, will always improve a company's bottom line. There is no scenario where the proper ML solution, delivered at a reasonable cost and in a timely fashion, will somehow hurt the company's bottom line. My point is that in this case management is failing because of certain trends and practices that are specific to the ML/DS community, namely: The idea that DS teams should operate independently of tech and business orgs -- too much autonomy for DS teams The disregard for domain knowledge that seems prevalent nowadays thanks to the ML hype, that DS can be generalists and someone with good enough ML chops can solve any business problem. That wasn't the case when I first left academia for the industry in 2009 (back then nobody would even bother with a phone screen if you didn't have the right domain knowledge). Over reliance on resources who check all the ML hype related boxes (knows Python, R, Tensorflow, Shiny, etc..., has the right Coursera certifications, has blogged on the topic, etc...), but are lacking in depth of experience. DS interviews nowadays all seem to be: Can you tell me what a p-value is? What is elastic net regression? Show me how to fit a model in sklearn? How do you impute NAs in an R dataframe? Any smart person can look those up on Stackoverflow or Cross-Validated,.....Instead teams should be asking stuff like: why does portfolio optimization use QP not LP? How does a forecast influence a customer service level? When should a recommendation engine be content based and when should it use collaborative filtering? etc... (This is a true story, happening to the company I currently work for. Names, domains, algorithms, and roles have been shuffled around to protect my anonymity)  Company A has been around for several decades. It is not the biggest name in its domain, but it is a well respected one. Risk analysis and portfolio optimization have been a core of Company A's business since the 90s. They have a large team of 30 or so analysts who perform those tasks on a daily basis. These analysts use ERP solutions implemented for them by one the big ERP companies (SAP, Teradata, Oracle, JD Edwards,...) or one of the major tech consulting companies (Deloitte, Accenture, PWC, Capgemini, etc...) in collaboration with their own in house engineering team. The tools used are embarrassingly old school: Classic RDBMS running on on-prem servers or maybe even on mainframes, code written in COBOL, Fortran, weird proprietary stuff like ABAP or SPSS.....you get the picture. But the models and analytic functions were pretty sophisticated, and surprisingly cutting edge compared to the published academic literature. Most of all, they fit well with the company's enterprise ecosystem, and were honed based on years of deep domain knowledge.  They have a tech team of several engineers (poached from the aforementioned software and consulting companies) and product managers (who came from the experienced pools of analysts and managers who use the software, or poached from business rivals) maintaining and running this software. Their technology might be old school, but collectively, they know the domain and the company's overall architecture very, very well. They've guided the company through several large scale upgrades and migrations and they have a track record of delivering on time, without too much overhead. The few times they've stumbled, they knew how to pick themselves up very quickly. In fact within their industry niche, they have a reputation for their expertise, and have very good relations with the various vendors they've had to deal with. They were the launching pad of several successful ERP consulting careers.  Interestingly, despite dealing on a daily basis with statistical modeling and optimization algorithms, none of the analysts, engineers, or product managers involved describe themselves as data scientists or machine learning experts. It is mostly a cultural thing: Their expertise predates the Data Science/ML hype that started circa 2010, and they got most of their chops using proprietary enterprise tools instead of the open source tools popular nowadays. A few of them have formal statistical training, but most of them came from engineering or domain backgrounds and learned stats on the fly while doing their job. Call this team "Team X".  Sometime around the mid 2010s, Company A started having some serious anxiety issues: Although still doing very well for a company its size, overall economic and demographic trends were shrinking its customer base, and a couple of so called disruptors came up with a new app and business model that started seriously eating into their revenue. A suitable reaction to appease shareholders and Wall Street was necessary. The company already had a decent website and a pretty snazzy app, what more could be done? Leadership decided that it was high time that AI and ML become a core part of the company's business. An ambitious Manager, with no science or engineering background, but who had very briefly toyed with a recommender system a couple of years back, was chosen to build a data science team, call it team "Y" (he had a bachelor's in history from the local state college and worked for several years in the company's marketing org). Team "Y" consists mostly of internal hires who decided they wanted to be data scientists and completed a Coursera certification or a Galvanize boot camp, before being brought on to the team, along with a few of fresh Ph.D or M.Sc holders who didn't like academia and wanted to try their hand at an industry role. All of them were very bright people, they could write great Medium blog posts and give inspiring TED talks, but collectively they had very little real world industry experience. As is the fashion nowadays, this group was made part of a data science org that reported directly to the CEO and Board, bypassing the CIO and any tech or business VPs, since Company A wanted to claim the monikers "data driven" and "AI powered" in their upcoming shareholder meetings. In 3 or 4 years of existence, team Y produced a few Python and R scripts. Their architectural experience  consisted almost entirely in connecting Flask to S3 buckets or Redshift tables, with a couple of the more resourceful ones learning how to plug their models into Tableau or how to spin up a Kuberneties pod.  But they needn't worry: The aforementioned manager, who was now a director (and was also doing an online Masters to make up for his qualifications gap and bolster his chances of becoming VP soon - at least he now understands what L1 regularization is), was a master at playing corporate politics and self-promotion. No matter how few actionable insights team Y produced or how little code they deployed to production, he always had their back and made sure they had ample funding. In fact he now had grandiose plans for setting up an all-purpose machine learning platform that can be used to solve all of the company's data problems.  A couple of sharp minded members of team Y, upon googling their industry name along with the word "data science", realized that risk analysis was a prime candidate for being solved with Bayesian models, and there was already a nifty R package for doing just that, whose tutorial they went through on R-Bloggers.com. One of them had even submitted a Bayesian classifier Kernel for a competition on Kaggle (he was 203rd on the leaderboard), and was eager to put his new-found expertise to use on a real world problem. They pitched the idea to their director, who saw a perfect use case for his upcoming ML platform. They started work on it immediately, without bothering to check whether anybody at Company A was already doing risk analysis. Since their org was independent, they didn't really need to check with anybody else before they got funding for their initiative. Although it was basically a Naive Bayes classifier, the term ML was added to the project tile, to impress the board.  As they progressed with their work however, tensions started to build. They had asked the data warehousing and CA analytics teams to build pipelines for them, and word eventually got out to team X about their project. Team X was initially thrilled: They offered to collaborate whole heartedly, and would have loved to add an ML based feather to their already impressive cap. The product owners and analysts were totally onboard as well: They saw a chance to get in on the whole Data Science hype that they kept hearing about. But through some weird mix of arrogance and insecurity, team Y refused to collaborate with them or share any of their long term goals with them, even as they went to other parts of the company giving brown bag presentations and tutorials on the new model they created.  Team X got resentful: from what they saw of team Y's model, their approach was hopelessly naive and had little chances of scaling or being sustainable in production, and they knew exactly how to help with that. Deploying the model to production would have taken them a few days, given how comfortable they were with DevOps and continuous delivery (team Y had taken several months to figure out how to deploy a simple R script to production). And despite how old school their own tech was, team X were crafty enough to be able to plug it in to their existing architecture. Moreover, the output of the model was such that it didn't take into account how the business will consume it or how it was going to be fed to downstream systems, and the product owners could have gone a long way in making the model more amenable to adoption by the business stakeholders. But team Y wouldn't listen, and their leads brushed off any attempts at communication, let alone collaboration. The vibe that team Y was giving off was "We are the cutting edge ML team, you guys are the legacy server grunts. We don't need your opinion.", and they seemed to have a complete disregard for domain knowledge, or worse, they thought that all that domain knowledge consisted of was being able to grasp the definitions of a few business metrics.  Team X got frustrated and tried to express their concerns to leadership. But despite owning a vital link in Company A's business process, they were only \~50 people in a large 1000 strong technology and operations org, and they were several layers removed from the C-suite, so it was impossible for them to get their voices heard.  Meanwhile, the unstoppable director was doing what he did best: Playing corporate politics. Despite how little his team had actually delivered, he had convinced the board that all analysis and optimization tasks should now be migrated to his yet to be delivered ML platform. Since most leaders now knew that there was overlap between team Y and team X's objectives, his pitch was no longer that team Y was going to create a new insight, but that they were going to replace (or modernize) the legacy statistics based on-prem tools with more accurate cloud based ML tools. Never mind that there was no support in the academic literature for the idea that Naive Bayes works better than the Econometric approaches used by team X, let alone the additional wacky idea that Bayesian Optimization would definitely outperform the QP solvers that were running in production.  Unbeknownst to team X, the original Bayesian risk analysis project has now grown into a multimillion dollar major overhaul initiative, which included the eventual replacement of all of the tools and functions supported by team X along with the necessary migration to the cloud. The CIO and a couple of business VPs are on now board, and tech leadership is treating it as a done deal. An outside vendor, a startup who nobody had heard of, was contracted to help build the platform, since team Y has no engineering skills. The choice was deliberate, as calling on any of the established consulting or software companies would have eventually led leadership to the conclusion that team X was better suited for a transformation on this scale than team Y.  Team Y has no experience with any major ERP deployments, and no domain knowledge, yet they are being tasked with fundamentally changing the business process that is at the core of Company A's business. Their models actually perform worse than those deployed by team X, and their architecture is hopelessly simplistic, compared to what is necessary for running such a solution in production.  Ironically, using Bayesian thinking and based on all the evidence, the likelihood that team Y succeeds is close to 0%. At best, the project is going to end up being a write off of 50 million dollars or more. Once the !@#$!@hits the fan, a couple of executive heads are going to role, and dozens of people will get laid off. At worst, given how vital risk analysis and portfolio optimization is to Company A's revenue stream, the failure will eventually sink the whole company. It probably won't go bankrupt, but it will lose a significant portion of its business and work force. Failed ERP implementations can and do sink large companies: Just see what happened to National Grid US, SuperValu or Target Canada.  One might argue that this is more about corporate disfunction and bad leadership than about data science and AI. But I disagree. I think the core driver of this debacle is indeed the blind faith in Data Scientists, ML models and the promise of AI, and the overall culture of hype and self promotion that is very common among the ML crowd.  We haven't seen the end of this story: I sincerely hope that this ends well for the sake of my colleagues and all involved. Company A is a good company, and both its customers and its employees deserver better. But the chances of that happening are negligible given all the information available, and this failure will hit my company hard.

[N] Netflix and European Space Agency no longer working with Siraj Raval
reddit
LLM Vibe Score0
Human Vibe Score0
inarrearsThis week

[N] Netflix and European Space Agency no longer working with Siraj Raval

According to article in The Register: A Netflix spokesperson confirmed to The Register it wasn’t working with Raval, and the ESA has cancelled the whole workshop altogether. “The situation is as it is. The workshop is cancelled, and that’s all,” Guillaume Belanger, an astrophysicist and the INTEGRAL Science Operations Coordinator at the ESA, told The Register on Monday. Raval isn’t about to quit his work any time soon, however. He promised students who graduated from his course that they would be referred to recruiters at Nvidia, Intel, Google and Amazon for engineering positions, or matched with a startup co-founder or a consulting client. In an unlisted YouTube video recorded live for his students discussing week eight of his course, and seen by El Reg, he read out a question posed to him: “Will your referrals hold any value now?” “Um, yeah they’re going to hold value. I don’t see why they wouldn’t. I mean, yes, some people on Twitter were angry but that has nothing to do with… I mean… I’ve also had tons of support, you know. I’ve had tons of support from people, who, uh, you know, support me, who work at these companies. He continues to justify his actions: “Public figures called me in private to remind me that this happens. You know, people make mistakes. You just have to keep going. They’re basically just telling me to not to stop. Of course, you make mistakes but you just keep going,” he claimed. When The Register asked Raval for comment, he responded: I've hardly taken any time off to relax since I first started my YouTube channel almost four years ago. And despite the enormous amount of work it takes to release two high quality videos a week for my audience, I progressively started to take on multiple other projects simultaneously by myself – a book, a docu-series, podcasts, YouTube videos, the course, the school of AI. Basically, these past few weeks, I've been experiencing a burnout unlike anything I've felt before. As a result, all of my output has been subpar. I made the [neural qubits] video and paper in one week. I remember wishing I had three to six months to really dive into quantum machine-learning and make something awesome, but telling myself I couldn't take that long as it would hinder my other projects. I plagiarized large chunks of the paper to meet my self-imposed one-week deadline. The associated video with animations took a lot more work to make. I didn't expect the paper to be cited as serious research, I considered it an additional reading resource for people who enjoyed the associated video to learn more about quantum machine learning. If I had a second chance, I'd definitely take way more time to write the paper, and in my own words. I've given refunds to every student who's asked so far, and the majority of students are still enrolled in the course. There are many happy students, they're just not as vocal on social media. We're on week 8 of 10 of my course, fully committed to student success. “And, no, I haven't plagiarized research for any other paper,” he added. https://www.theregister.co.uk/2019/10/14/ravelaiyoutube/

[R] Analysis of 400+ ML competitions in 2024
reddit
LLM Vibe Score0
Human Vibe Score1
hcarlensThis week

[R] Analysis of 400+ ML competitions in 2024

I run mlcontests.com, a website that lists ML competitions from across multiple platforms - Kaggle, DrivenData, AIcrowd, Zindi, etc… I’ve just spent a few months looking through all the info I could find on last year’s competitions, as well as winning solutions.  I found over 400 competitions that happened last year, plus info on the #1 winning solution for 70 of those.  Some highlights: Kaggle is still the biggest platform by total prize money, and also has a much bigger user base than the other platforms - though there are well over a dozen other platforms worth keeping track of, with regular interesting competitions and meaningful prize money. An increase in competitions with $1m+ prize pools (ARC Prize, AI Mathematical Olympiad, Vesuvius Challenge, AI Cyber Challenge) compared to previous years. Python continues to be the language of choice among competition winners, with almost everyone using Python as their main language. One winner used Rust, two used R.  Convolutional neural nets continue to do well in computer vision competitions, and are still more common among competition winners than transformer-based vision models.  PyTorch is still used a lot more than TensorFlow, roughly 9:1. Didn’t find any competition winners implementing neural nets in JAX or other libraries.  There were a few competition winners using AutoML packages, which seem to be getting increasingly useful. Any claims of generalist autonomous grandmaster-level agents seem premature though.  In language/text/sequence-related competitions, quantisation was key for making use of limited resources effectively. Usually 4-, 5-, or 8-bit. LoRA/QLoRA was also used quite often, though not always.  Gradient-boosted decision trees continue to win a lot of tabular/time-series competitions. They’re often ensembled with deep learning models. No tabular/time-series pre-trained foundation models were used by winners in 2024, as far as I can tell.  Starting to see more uptake of Polars for dataframes, with 7 winners using Polars in 2024 (up from 3 in 2023) vs 58 using Pandas. All those who used Polars also still used Pandas in some parts of their code.  In terms of hardware, competition winners almost entirely used NVIDIA GPUs to train their models. Some trained on CPU-only, or used a TPU through Colab. No AMD GPUs. The NVIDIA A100 was the most commonly used GPU among winners. Two of the $1m+ prize pool competitions were won by teams using 8xH100 nodes for training. A lot of other GPUs too though: T4/P100 (through Kaggle Notebooks), or consumer GPUs like RTX 3090/4090/3080/3060. Some spent hundreds of dollars on cloud compute to train their solutions.  An emerging pattern: using generative models to create additional synthetic training data to augment the training data provided.  There’s way more detail in the full report, which you can read here (no paywall): https://mlcontests.com/state-of-machine-learning-competitions-2024?ref=mlcr Processing img xmm4ywg9h9le1... The full report also features: A deep dive into the ARC Prize and the AI Mathematical Olympiad An overview of winning solutions to NLP/sequence competitions A breakdown of Python packages used in winning solutions (e.g. relative popularity of various gradient-boosted tree libraries) If you’d like to support this research, I’d really appreciate it if you could share it with anyone else who might find it interesting. You can also check out my newly-launched online magazine, Jolt ML \- featuring news from top ML conferences as well as long-read articles (just one so far, more to come!).  Thanks to the competition winners who shared info on their solutions, and also to the competition platforms who shared high-level data on their competitions.

[D] Overwhelmed by fast advances in recent weeks
reddit
LLM Vibe Score0
Human Vibe Score1
iamx9000againThis week

[D] Overwhelmed by fast advances in recent weeks

I was watching the GTC keynote and became entirely overwhelmed by the amount of progress achieved from last year. I'm wondering how everyone else feels. &#x200B; Firstly, the entire ChatGPT, GPT-3/GPT-4 chaos has been going on for a few weeks, with everyone scrambling left and right to integrate chatbots into their apps, products, websites. Twitter is flooded with new product ideas, how to speed up the process from idea to product, countless promp engineering blogs, tips, tricks, paid courses. &#x200B; Not only was ChatGPT disruptive, but a few days later, Microsoft and Google also released their models and integrated them into their search engines. Microsoft also integrated its LLM into its Office suite. It all happenned overnight. I understand that they've started integrating them along the way, but still, it seems like it hapenned way too fast. This tweet encompases the past few weeks perfectly https://twitter.com/AlphaSignalAI/status/1638235815137386508 , on a random Tuesday countless products are released that seem revolutionary. &#x200B; In addition to the language models, there are also the generative art models that have been slowly rising in mainstream recognition. Now Midjourney AI is known by a lot of people who are not even remotely connected to the AI space. &#x200B; For the past few weeks, reading Twitter, I've felt completely overwhelmed, as if the entire AI space is moving beyond at lightning speed, whilst around me we're just slowly training models, adding some data, and not seeing much improvement, being stuck on coming up with "new ideas, that set us apart". &#x200B; Watching the GTC keynote from NVIDIA I was again, completely overwhelmed by how much is being developed throughout all the different domains. The ASML EUV (microchip making system) was incredible, I have no idea how it does lithography and to me it still seems like magic. The Grace CPU with 2 dies (although I think Apple was the first to do it?) and 100 GB RAM, all in a small form factor. There were a lot more different hardware servers that I just blanked out at some point. The omniverse sim engine looks incredible, almost real life (I wonder how much of a domain shift there is between real and sim considering how real the sim looks). Beyond it being cool and usable to train on synthetic data, the car manufacturers use it to optimize their pipelines. This change in perspective, of using these tools for other goals than those they were designed for I find the most interesting. &#x200B; The hardware part may be old news, as I don't really follow it, however the software part is just as incredible. NVIDIA AI foundations (language, image, biology models), just packaging everything together like a sandwich. Getty, Shutterstock and Adobe will use the generative models to create images. Again, already these huge juggernauts are already integrated. &#x200B; I can't believe the point where we're at. We can use AI to write code, create art, create audiobooks using Britney Spear's voice, create an interactive chatbot to converse with books, create 3D real-time avatars, generate new proteins (?i'm lost on this one), create an anime and countless other scenarios. Sure, they're not perfect, but the fact that we can do all that in the first place is amazing. &#x200B; As Huang said in his keynote, companies want to develop "disruptive products and business models". I feel like this is what I've seen lately. Everyone wants to be the one that does something first, just throwing anything and everything at the wall and seeing what sticks. &#x200B; In conclusion, I'm feeling like the world is moving so fast around me whilst I'm standing still. I want to not read anything anymore and just wait until everything dies down abit, just so I can get my bearings. However, I think this is unfeasible. I fear we'll keep going in a frenzy until we just burn ourselves at some point. &#x200B; How are you all fairing? How do you feel about this frenzy in the AI space? What are you the most excited about?

I am Jürgen Schmidhuber, AMA!
reddit
LLM Vibe Score0
Human Vibe Score1
JuergenSchmidhuberThis week

I am Jürgen Schmidhuber, AMA!

Hello /r/machinelearning, I am Jürgen Schmidhuber (pronounce: You_again Shmidhoobuh) and I will be here to answer your questions on 4th March 2015, 10 AM EST. You can post questions in this thread in the meantime. Below you can find a short introduction about me from my website (you can read more about my lab’s work at people.idsia.ch/~juergen/). Edits since 9th March: Still working on the long tail of more recent questions hidden further down in this thread ... Edit of 6th March: I'll keep answering questions today and in the next few days - please bear with my sluggish responses. Edit of 5th March 4pm (= 10pm Swiss time): Enough for today - I'll be back tomorrow. Edit of 5th March 4am: Thank you for great questions - I am online again, to answer more of them! Since age 15 or so, Jürgen Schmidhuber's main scientific ambition has been to build an optimal scientist through self-improving Artificial Intelligence (AI), then retire. He has pioneered self-improving general problem solvers since 1987, and Deep Learning Neural Networks (NNs) since 1991. The recurrent NNs (RNNs) developed by his research groups at the Swiss AI Lab IDSIA (USI & SUPSI) & TU Munich were the first RNNs to win official international contests. They recently helped to improve connected handwriting recognition, speech recognition, machine translation, optical character recognition, image caption generation, and are now in use at Google, Microsoft, IBM, Baidu, and many other companies. IDSIA's Deep Learners were also the first to win object detection and image segmentation contests, and achieved the world's first superhuman visual classification results, winning nine international competitions in machine learning & pattern recognition (more than any other team). They also were the first to learn control policies directly from high-dimensional sensory input using reinforcement learning. His research group also established the field of mathematically rigorous universal AI and optimal universal problem solvers. His formal theory of creativity & curiosity & fun explains art, science, music, and humor. He also generalized algorithmic information theory and the many-worlds theory of physics, and introduced the concept of Low-Complexity Art, the information age's extreme form of minimal art. Since 2009 he has been member of the European Academy of Sciences and Arts. He has published 333 peer-reviewed papers, earned seven best paper/best video awards, and is recipient of the 2013 Helmholtz Award of the International Neural Networks Society.

[D] Advanced courses update
reddit
LLM Vibe Score0
Human Vibe Score1
actbshThis week

[D] Advanced courses update

EDIT Jan 2021 : I am still updating the list as of Jan, 2021 and will most probably continue to do so for foreseeable future. So, please feel free to message me any courses you find interesting that fit here. - - We have a PhD level or Advanced courses thread in the sidebar but it's three year old now. There were two other 7-8 month old threads (1, 2) but they don't have many quality responses either. So, can we have a new one here? To reiterate - CS231n, CS229, ones from Udemy etc are not advanced. Advanced ML/DL/RL, attempts at building theory of DL, optimization theory, advanced applications etc are some examples of what I believe should belong here, much like the original sidebar post. You can also suggest (new) categories for the courses you share. :) - - Here are some courses we've found so far. ML >> Learning Discrete Latent Structure - sta4273/csc2547 Spring'18 Learning to Search - csc2547 Fall'19 Scalable and Flexible Models of Uncertainty - csc2541 Fundamentals of Machine Learning Over Networks - ep3260 Machine Learning on Graphs - cs224w, videos Mining Massive Data Sets - cs246 Interactive Learning - cse599 Machine Learning for Sequential Decision Making Under Uncertainty - ee290s/cs194 Probabilistic Graphical Methods - 10-708 Introduction to Causal Inference ML >> Theory Statistical Machine Learning - 10-702/36-702 with videos, 2016 videos Statistical Learning Theory - cs229T/stats231 Stanford Autumn'18-19 Statistical Learning Theory - cs281b /stat241b UC Berkeley, Spring'14 Statistical Learning Theory - csc2532 Uni of Toronto, Spring'20 ML >> Bayesian Bayesian Data Analysis Bayesian Methods Research Group, Moscow, Bayesian Methods in ML - spring2020, fall2020 Deep Learning and Bayesian Methods - summer school, videos available for 2019 version ML >> Systems and Operations Stanford MLSys Seminar Series Visual Computing Systems- cs348v - Another systems course that discusses hardware from a persepective of visual computing but is relevant to ML as well Advanced Machine Learning Systems - cs6787 - lecture 9 and onwards discuss hardware side of things Machine Learning Systems Design - cs329S Topics in Deployable ML - 6.S979 Machine Learning in Production / AI Engineering (17-445/17-645/17-745/11-695) AutoML - Automated Machine Learning DL >> Deep Unsupervised Learning - cs294 Deep Multi-task and Meta learning - cs330 Topics in Deep Learning - stat991 UPenn/Wharton most chapters start with introductory topics and dig into advanced ones towards the end. Deep Generative Models - cs236 Deep Geometric Learning of Big Data and Applications Deep Implicit Layers - NeurIPS 2020 tutorial DL >> Theory Topics course on Mathematics of Deep Learning - CSCI-GA 3033 Topics Course on Deep Learning - stat212b Analyses of Deep Learning - stats385, videos from 2017 version Mathematics of Deep Learning Geometry of Deep Learning RL >> Meta-Learning - ICML 2019 Tutorial , Metalearning: Applications to Data Mining - google books link Deep Multi-Task and Meta Learning - cs330, videos Deep Reinforcement Learning - cs285 Advanced robotics - cs287 Reinforcement Learning - cs234, videos for 2019 run Reinforcement Learning Summer School 2019: Bandits, RL & Deep RL Optimization >> Convex Optimization I - ee364a, has quite recent videos too. Convex Optimization II - ee364b, 2008 videos Convex Optimization and Approximation - ee227c Convex Optimization - ee227bt Variational Methods for Computer Vision Advanced Optimization and Randomized Algorithms - 10-801, videos Optimization Methods for Machine Learning and Engineering - Karlsruhe Institute of Technology Applications >> Computer Vision Computational Video Manipulation - cs448v Advanced Topics in ML: Modeling and Segmentation of Multivariate Mixed Data TUM AI Guest lecture series - many influential researchers in DL, vision, graphics talk about latest advances and their latest works. Advanced Deep Learning for Computer Vision - TUM ADL4CV Detection, Segmentation and Tracking - TUM CV3DST Guest lectures at TUM Dynamic Vision and Learning group Vision Seminar at MIT Autonomous Vision Group, Talk@Tübingen Seminar Applications >> Natural Language Processing Natural Language Processing with Deep Learning - cs224n ( not sure if it belongs here, people working in NLP can help me out) Neural networks for NLP - cs11-747 Natural Language Understanding - cs224u, video Applications >> 3D Graphics Non-Euclidean Methods in Machine Learning - cs468, 2020 Machine Learning for 3D Data - cs468, spring 2017 Data-Driven Shape Analysis - cs468, 2014 Geometric Deep Learning - Not a course but the website links a few tutorials on Geometric DL Deep Learning for Computer Graphics - SIGGRAPH 2019 Machine Learning for Machine Vision as Inverse Graphics - csc2547 Winter'20 Machine Learning Meets Geometry, winter 2020; Machine Learning for 3D Data, winter 2018 Edit: Upon suggestion, categorized the courses. There might be some misclassifications as I'm not trained on this task ;). Added some good ones from older (linked above) discussions.

[P] The Big Sleep: Text-to-image generation using BigGAN and OpenAI's CLIP via a Google Colab notebook from Twitter user Adverb
reddit
LLM Vibe Score0
Human Vibe Score0.333
WiskkeyThis week

[P] The Big Sleep: Text-to-image generation using BigGAN and OpenAI's CLIP via a Google Colab notebook from Twitter user Adverb

From https://twitter.com/advadnoun/status/1351038053033406468: The Big Sleep Here's the notebook for generating images by using CLIP to guide BigGAN. It's very much unstable and a prototype, but it's also a fair place to start. I'll likely update it as time goes on. colab.research.google.com/drive/1NCceX2mbiKOSlAd\o7IU7nA9UskKN5WR?usp=sharing I am not the developer of The Big Sleep. This is the developer's Twitter account; this is the developer's Reddit account. Steps to follow to generate the first image in a given Google Colab session: Optionally, if this is your first time using Google Colab, view this Colab introduction and/or this Colab FAQ. Click this link. Sign into your Google account if you're not already signed in. Click the "S" button in the upper right to do this. Note: Being signed into a Google account has privacy ramifications, such as your Google search history being recorded in your Google account. In the Table of Contents, click "Parameters". Find the line that reads "tx = clip.tokenize('''a cityscape in the style of Van Gogh''')" and change the text inside of the single quote marks to your desired text; example: "tx = clip.tokenize('''a photo of New York City''')". The developer recommends that you keep the three single quote marks on both ends of your desired text so that mult-line text can be used An alternative is to remove two of the single quotes on each end of your desired text; example: "tx = clip.tokenize('a photo of New York City')". In the Table of Contents, click "Restart the kernel...". Position the pointer over the first cell in the notebook, which starts with text "import subprocess". Click the play button (the triangle) to run the cell. Wait until the cell completes execution. Click menu item "Runtime->Restart and run all". In the Table of Contents, click "Diagnostics". The output appears near the end of the Train cell that immediately precedes the Diagnostics cell, so scroll up a bit. Every few minutes (or perhaps 10 minutes if Google assigned you relatively slow hardware for this session), a new image will appear in the Train cell that is a refinement of the previous image. This process can go on for as long as you want until Google ends your Google Colab session, which is a total of up to 12 hours for the free version of Google Colab. Steps to follow if you want to start a different run using the same Google Colab session: Click menu item "Runtime->Interrupt execution". Save any images that you want to keep by right-clicking on them and using the appropriate context menu command. Optionally, change the desired text. Different runs using the same desired text almost always results in different outputs. Click menu item "Runtime->Restart and run all". Steps to follow when you're done with your Google Colab session: Click menu item "Runtime->Manage sessions". Click "Terminate" to end the session. Optionally, log out of your Google account due to the privacy ramifications of being logged into a Google account. The first output image in the Train cell (using the notebook's default of seeing every 100th image generated) usually is a very poor match to the desired text, but the second output image often is a decent match to the desired text. To change the default of seeing every 100th image generated, change the number 100 in line "if itt % 100 == 0:" in the Train cell to the desired number. For free-tier Google Colab users, I recommend changing 100 to a small integer such as 5. Tips for the text descriptions that you supply: In Section 3.1.4 of OpenAI's CLIP paper (pdf), the authors recommend using a text description of the form "A photo of a {label}." or "A photo of a {label}, a type of {type}." for images that are photographs. A Reddit user gives these tips. The Big Sleep should generate these 1,000 types of things better on average than other types of things. Here is an article containing a high-level description of how The Big Sleep works. The Big Sleep uses a modified version of BigGAN as its image generator component. The Big Sleep uses the ViT-B/32 CLIP model to rate how well a given image matches your desired text. The best CLIP model according to the CLIP paper authors is the (as of this writing) unreleased ViT-L/14-336px model; see Table 10 on page 40 of the CLIP paper (pdf) for a comparison. There are many other sites/programs/projects that use CLIP to steer image/video creation to match a text description. Some relevant subreddits: r/bigsleep (subreddit for images/videos generated from text-to-image machine learning algorithms). r/deepdream (subreddit for images/videos generated from machine learning algorithms). r/mediasynthesis (subreddit for media generation/manipulation techniques that use artificial intelligence; this subreddit shouldn't be used to post images/videos unless new techniques are demonstrated, or the images/videos are of high quality relative to other posts). Example using text 'a black cat sleeping on top of a red clock': https://preview.redd.it/7xq58v7022c61.png?width=512&format=png&auto=webp&s=a229ae9add555cd1caba31c42b60d907ffe67773 Example using text 'the word ''hot'' covered in ice': https://preview.redd.it/6kxdp8u3k2c61.png?width=512&format=png&auto=webp&s=5bd078b0111575f5d88a1dc53b0aeb933f3b0da6 Example using text 'a monkey holding a green lightsaber': https://preview.redd.it/rdsybsoaz2c61.png?width=512&format=png&auto=webp&s=2769d4c6c883c1c35ae0b1c629bebe9bc1d41393 Example using text 'The White House in Washington D.C. at night with green and red spotlights shining on it': https://preview.redd.it/w4mg90xsf5c61.png?width=512&format=png&auto=webp&s=5f18318de2f77bcd8a86e71e87048fadd30383d1 Example using text '''A photo of the Golden Gate Bridge at night, illuminated by spotlights in a tribute to Prince''': https://preview.redd.it/cn4ecuafhic61.png?width=512&format=png&auto=webp&s=397c838fdc49f13c5f17110b92c78b95bf0dcac0 Example using text '''a Rembrandt-style painting titled "Robert Plant decides whether to take the stairway to heaven or the ladder to heaven"''': https://preview.redd.it/h7rb3y6j5jc61.png?width=512&format=png&auto=webp&s=537bfe8210af185647b00e7585c948aa2c4e0ffb Example using text '''A photo of the Empire State Building being shot at with the laser cannons of a TIE fighter.''': https://preview.redd.it/cwi7i639c5d61.png?width=512&format=png&auto=webp&s=0510c8b93adb40eee4d3f41607f1c215d41e55ff Example using text '''A cartoon of a new mascot for the Reddit subreddit DeepDream that has a mouse-like face and wears a cape''': https://preview.redd.it/wtxbduevcbd61.png?width=512&format=png&auto=webp&s=c5d266258922bc62f25c80a08cd9cabc07d9cb1c Example using text '''Bugs Bunny meets the Eye of Sauron, drawn in the Looney Tunes cartoon style''': https://preview.redd.it/gmljaeekuid61.png?width=512&format=png&auto=webp&s=9ea578de165e12afc3a62bf6886bc1ae9dc19bec Example using text '''Photo of a blue and red neon-colored frog at night.''': https://preview.redd.it/nzlypte6wzd61.png?width=512&format=png&auto=webp&s=7e10b06f22cfc57c64b6d05738c7486b895083df Example using text '''Hell begins to freeze over''': https://preview.redd.it/vn99we9ngmf61.png?width=512&format=png&auto=webp&s=2408efd607f0ab40a08db6ee67448791aa813993 Example using text '''A scene with vibrant colors''': https://preview.redd.it/4z133mvrgmf61.png?width=512&format=png&auto=webp&s=b78e7a8e3f736769655056093a9904ff09a355a1 Example using text '''The Great Pyramids were turned into prisms by a wizard''': https://preview.redd.it/zxt6op7vgmf61.png?width=512&format=png&auto=webp&s=53e578cfde14b28afe27957e95e610b89afadd44

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[D] I don't really trust papers out of "Top Labs" anymore
reddit
LLM Vibe Score0
Human Vibe Score0.333
MrAcuriteThis week

[D] I don't really trust papers out of "Top Labs" anymore

I mean, I trust that the numbers they got are accurate and that they really did the work and got the results. I believe those. It's just that, take the recent "An Evolutionary Approach to Dynamic Introduction of Tasks in Large-scale Multitask Learning Systems" paper. It's 18 pages of talking through this pretty convoluted evolutionary and multitask learning algorithm, it's pretty interesting, solves a bunch of problems. But two notes. One, the big number they cite as the success metric is 99.43 on CIFAR-10, against a SotA of 99.40, so woop-de-fucking-doo in the grand scheme of things. Two, there's a chart towards the end of the paper that details how many TPU core-hours were used for just the training regimens that results in the final results. The sum total is 17,810 core-hours. Let's assume that for someone who doesn't work at Google, you'd have to use on-demand pricing of $3.22/hr. This means that these trained models cost $57,348. Strictly speaking, throwing enough compute at a general enough genetic algorithm will eventually produce arbitrarily good performance, so while you can absolutely read this paper and collect interesting ideas about how to use genetic algorithms to accomplish multitask learning by having each new task leverage learned weights from previous tasks by defining modifications to a subset of components of a pre-existing model, there's a meta-textual level on which this paper is just "Jeff Dean spent enough money to feed a family of four for half a decade to get a 0.03% improvement on CIFAR-10." OpenAI is far and away the worst offender here, but it seems like everyone's doing it. You throw a fuckton of compute and a light ganache of new ideas at an existing problem with existing data and existing benchmarks, and then if your numbers are infinitesimally higher than their numbers, you get to put a lil' sticker on your CV. Why should I trust that your ideas are even any good? I can't check them, I can't apply them to my own projects. Is this really what we're comfortable with as a community? A handful of corporations and the occasional university waving their dicks at everyone because they've got the compute to burn and we don't? There's a level at which I think there should be a new journal, exclusively for papers in which you can replicate their experimental results in under eight hours on a single consumer GPU.

[N] OpenAI's new language model gpt-3.5-turbo-instruct can defeat chess engine Fairy-Stockfish 14 at level 5
reddit
LLM Vibe Score0
Human Vibe Score1
WiskkeyThis week

[N] OpenAI's new language model gpt-3.5-turbo-instruct can defeat chess engine Fairy-Stockfish 14 at level 5

This Twitter thread (Nitter alternative for those who aren't logged into Twitter and want to see the full thread) claims that OpenAI's new language model gpt-3.5-turbo-instruct can "readily" beat Lichess Stockfish level 4 (Lichess Stockfish level and its rating) and has a chess rating of "around 1800 Elo." This tweet shows the style of prompts that are being used to get these results with the new language model. I used website parrotchess\[dot\]com (discovered here) (EDIT: parrotchess doesn't exist anymore, as of March 7, 2024) to play multiple games of chess purportedly pitting this new language model vs. various levels at website Lichess, which supposedly uses Fairy-Stockfish 14 according to the Lichess user interface. My current results for all completed games: The language model is 5-0 vs. Fairy-Stockfish 14 level 5 (game 1, game 2, game 3, game 4, game 5), and 2-5 vs. Fairy-Stockfish 14 level 6 (game 1, game 2, game 3, game 4, game 5, game 6, game 7). Not included in the tally are games that I had to abort because the parrotchess user interface stalled (5 instances), because I accidentally copied a move incorrectly in the parrotchess user interface (numerous instances), or because the parrotchess user interface doesn't allow the promotion of a pawn to anything other than queen (1 instance). Update: There could have been up to 5 additional losses - the number of times the parrotchess user interface stalled - that would have been recorded in this tally if this language model resignation bug hadn't been present. Also, the quality of play of some online chess bots can perhaps vary depending on the speed of the user's hardware. The following is a screenshot from parrotchess showing the end state of the first game vs. Fairy-Stockfish 14 level 5: https://preview.redd.it/4ahi32xgjmpb1.jpg?width=432&format=pjpg&auto=webp&s=7fbb68371ca4257bed15ab2828fab58047f194a4 The game results in this paragraph are from using parrotchess after the forementioned resignation bug was fixed. The language model is 0-1 vs. Fairy-Stockfish level 7 (game 1), and 0-1 vs. Fairy-Stockfish 14 level 8 (game 1). There is one known scenario (Nitter alternative) in which the new language model purportedly generated an illegal move using language model sampling temperature of 0. Previous purported illegal moves that the parrotchess developer examined turned out (Nitter alternative) to be due to parrotchess bugs. There are several other ways to play chess against the new language model if you have access to the OpenAI API. The first way is to use the OpenAI Playground as shown in this video. The second way is chess web app gptchess\[dot\]vercel\[dot\]app (discovered in this Twitter thread / Nitter thread). Third, another person modified that chess web app to additionally allow various levels of the Stockfish chess engine to autoplay, resulting in chess web app chessgpt-stockfish\[dot\]vercel\[dot\]app (discovered in this tweet). Results from other people: a) Results from hundreds of games in blog post Debunking the Chessboard: Confronting GPTs Against Chess Engines to Estimate Elo Ratings and Assess Legal Move Abilities. b) Results from 150 games: GPT-3.5-instruct beats GPT-4 at chess and is a \~1800 ELO chess player. Results of 150 games of GPT-3.5 vs stockfish and 30 of GPT-3.5 vs GPT-4. Post #2. The developer later noted that due to bugs the legal move rate was actually above 99.9%. It should also be noted that these results didn't use a language model sampling temperature of 0, which I believe could have induced illegal moves. c) Chess bot gpt35-turbo-instruct at website Lichess. d) Chess bot konaz at website Lichess. From blog post Playing chess with large language models: Computers have been better than humans at chess for at least the last 25 years. And for the past five years, deep learning models have been better than the best humans. But until this week, in order to be good at chess, a machine learning model had to be explicitly designed to play games: it had to be told explicitly that there was an 8x8 board, that there were different pieces, how each of them moved, and what the goal of the game was. Then it had to be trained with reinforcement learning agaist itself. And then it would win. This all changed on Monday, when OpenAI released GPT-3.5-turbo-instruct, an instruction-tuned language model that was designed to just write English text, but that people on the internet quickly discovered can play chess at, roughly, the level of skilled human players. Post Chess as a case study in hidden capabilities in ChatGPT from last month covers a different prompting style used for the older chat-based GPT 3.5 Turbo language model. If I recall correctly from my tests with ChatGPT-3.5, using that prompt style with the older language model can defeat Stockfish level 2 at Lichess, but I haven't been successful in using it to beat Stockfish level 3. In my tests, both the quality of play and frequency of illegal attempted moves seems to be better with the new prompt style with the new language model compared to the older prompt style with the older language model. Related article: Large Language Model: world models or surface statistics? P.S. Since some people claim that language model gpt-3.5-turbo-instruct is always playing moves memorized from the training dataset, I searched for data on the uniqueness of chess positions. From this video, we see that for a certain game dataset there were 763,331,945 chess positions encountered in an unknown number of games without removing duplicate chess positions, 597,725,848 different chess positions reached, and 582,337,984 different chess positions that were reached only once. Therefore, for that game dataset the probability that a chess position in a game was reached only once is 582337984 / 763331945 = 76.3%. For the larger dataset cited in that video, there are approximately (506,000,000 - 200,000) games in the dataset (per this paper), and 21,553,382,902 different game positions encountered. Each game in the larger dataset added a mean of approximately 21,553,382,902 / (506,000,000 - 200,000) = 42.6 different chess positions to the dataset. For this different dataset of \~12 million games, \~390 million different chess positions were encountered. Each game in this different dataset added a mean of approximately (390 million / 12 million) = 32.5 different chess positions to the dataset. From the aforementioned numbers, we can conclude that a strategy of playing only moves memorized from a game dataset would fare poorly because there are not rarely new chess games that have chess positions that are not present in the game dataset.

[D] What is your honest experience with reinforcement learning?
reddit
LLM Vibe Score0
Human Vibe Score1
Starks-TechnologyThis week

[D] What is your honest experience with reinforcement learning?

In my personal experience, SOTA RL algorithms simply don't work. I've tried working with reinforcement learning for over 5 years. I remember when Alpha Go defeated the world famous Go player, Lee Sedol, and everybody thought RL would take the ML community by storm. Yet, outside of toy problems, I've personally never found a practical use-case of RL. What is your experience with it? Aside from Ad recommendation systems and RLHF, are there legitimate use-cases of RL? Or, was it all hype? Edit: I know a lot about AI. I built NexusTrade, an AI-Powered automated investing tool that lets non-technical users create, update, and deploy their trading strategies. I’m not an idiot nor a noob; RL is just ridiculously hard. Edit 2: Since my comments are being downvoted, here is a link to my article that better describes my position. It's not that I don't understand RL. I released my open-source code and wrote a paper on it. It's the fact that it's EXTREMELY difficult to understand. Other deep learning algorithms like CNNs (including ResNets), RNNs (including GRUs and LSTMs), Transformers, and GANs are not hard to understand. These algorithms work and have practical use-cases outside of the lab. Traditional SOTA RL algorithms like PPO, DDPG, and TD3 are just very hard. You need to do a bunch of research to even implement a toy problem. In contrast, the decision transformer is something anybody can implement, and it seems to match or surpass the SOTA. You don't need two networks battling each other. You don't have to go through hell to debug your network. It just naturally learns the best set of actions in an auto-regressive manner. I also didn't mean to come off as arrogant or imply that RL is not worth learning. I just haven't seen any real-world, practical use-cases of it. I simply wanted to start a discussion, not claim that I know everything. Edit 3: There's a shockingly number of people calling me an idiot for not fully understanding RL. You guys are wayyy too comfortable calling people you disagree with names. News-flash, not everybody has a PhD in ML. My undergraduate degree is in biology. I self-taught myself the high-level maths to understand ML. I'm very passionate about the field; I just have VERY disappointing experiences with RL. Funny enough, there are very few people refuting my actual points. To summarize: Lack of real-world applications Extremely complex and inaccessible to 99% of the population Much harder than traditional DL algorithms like CNNs, RNNs, and GANs Sample inefficiency and instability Difficult to debug Better alternatives, such as the Decision Transformer Are these not legitimate criticisms? Is the purpose of this sub not to have discussions related to Machine Learning? To the few commenters that aren't calling me an idiot...thank you! Remember, it costs you nothing to be nice! Edit 4: Lots of people seem to agree that RL is over-hyped. Unfortunately those comments are downvoted. To clear up some things: We've invested HEAVILY into reinforcement learning. All we got from this investment is a robot that can be super-human at (some) video games. AlphaFold did not use any reinforcement learning. SpaceX doesn't either. I concede that it can be useful for robotics, but still argue that it's use-cases outside the lab are extremely limited. If you're stumbling on this thread and curious about an RL alternative, check out the Decision Transformer. It can be used in any situation that a traditional RL algorithm can be used. Final Edit: To those who contributed more recently, thank you for the thoughtful discussion! From what I learned, model-based models like Dreamer and IRIS MIGHT have a future. But everybody who has actually used model-free models like DDPG unanimously agree that they suck and don’t work.

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[Discussion]: Mark Zuckerberg on Meta's Strategy on Open Source and AI during the earnings call
reddit
LLM Vibe Score0
Human Vibe Score1
noiseinvacuumThis week

[Discussion]: Mark Zuckerberg on Meta's Strategy on Open Source and AI during the earnings call

During the recent earnings call, Mark Zuckerberg answered a question from Eric Sheridan of Goldman Sachs on Meta's AI strategy, opportunities to integrate into products, and why they open source models and how it would benefit their business. I found the reasoning to be very sound and promising for the OSS and AI community. The biggest risk from AI, in my opinion, is not the doomsday scenarios that intuitively come to mind but rather that the most powerful AI systems will only be accessible to the most powerful and resourceful corporations. Quote copied from Ben Thompson's write up on Meta's earning in his Stratechery blog post which goes beyond AI. It's behind a paywall but I highly recommend it personally. Some noteworthy quotes that signal the thought process at Meta FAIR and more broadly We’re just playing a different game on the infrastructure than companies like Google or Microsoft or Amazon We would aspire to and hope to make even more open than that. So, we’ll need to figure out a way to do that. ...lead us to do more work in terms of open sourcing, some of the lower level models and tools Open sourcing low level tools make the way we run all this infrastructure more efficient over time. On PyTorch: It’s generally been very valuable for us to provide that because now all of the best developers across the industry are using tools that we’re also using internally. I would expect us to be pushing and helping to build out an open ecosystem. For all the negative that comes out of the popular discourse on Meta, I think their work to open source key tech tools over the last 10 years has been exceptional, here's hoping it continues into this decade of AI and pushes other tech giants to also realize the benefits of Open Source. Full Transcript: Right now most of the companies that are training large language models have business models that lead them to a closed approach to development. I think there’s an important opportunity to help create an open ecosystem. If we can help be a part of this, then much of the industry will standardize on using these open tools and help improve them further. So this will make it easier for other companies to integrate with our products and platforms as we enable more integrations, and that will help our products stay at the leading edge as well. Our approach to AI and our infrastructure has always been fairly open. We open source many of our state of the art models so people can experiment and build with them. This quarter we released our LLaMa LLM to researchers. It has 65 billion parameters but outperforms larger models and has proven quite popular. We’ve also open-sourced three other groundbreaking visual models along with their training data and model weights — Segment Anything, DinoV2, and our Animated Drawings tool — and we’ve gotten positive feedback on all of those as well. I think that there’s an important distinction between the products we offer and a lot of the technical infrastructure, especially the software that we write to support that. And historically, whether it’s the Open Compute project that we’ve done or just open sourcing a lot of the infrastructure that we’ve built, we’ve historically open sourced a lot of that infrastructure, even though we haven’t open sourced the code for our core products or anything like that. And the reason why I think why we do this is that unlike some of the other companies in the space, we’re not selling a cloud computing service where we try to keep the different software infrastructure that we’re building proprietary. For us, it’s way better if the industry standardizes on the basic tools that we’re using and therefore we can benefit from the improvements that others make and others’ use of those tools can, in some cases like Open Compute, drive down the costs of those things which make our business more efficient too. So I think to some degree we’re just playing a different game on the infrastructure than companies like Google or Microsoft or Amazon, and that creates different incentives for us. So overall, I think that that’s going to lead us to do more work in terms of open sourcing, some of the lower level models and tools. But of course, a lot of the product work itself is going to be specific and integrated with the things that we do. So it’s not that everything we do is going to be open. Obviously, a bunch of this needs to be developed in a way that creates unique value for our products, but I think in terms of the basic models, I would expect us to be pushing and helping to build out an open ecosystem here, which I think is something that’s going to be important. On the AI tools, and we have a bunch of history here, right? So if you if you look at what we’ve done with PyTorch, for example, which has generally become the standard in the industry as a tool that a lot of folks who are building AI models and different things in that space use, it’s generally been very valuable for us to provide that because now all of the best developers across the industry are using tools that we’re also using internally. So the tool chain is the same. So when they create some innovation, we can easily integrate it into the things that we’re doing. When we improve something, it improves other products too. Because it’s integrated with our technology stack, when there are opportunities to make integrations with products, it’s much easier to make sure that developers and other folks are compatible with the things that we need in the way that our systems work. So there are a lot of advantages, but I view this more as a kind of back end infrastructure advantage with potential integrations on the product side, but one that should hopefully enable us to stay at the leading edge and integrate more broadly with the community and also make the way we run all this infrastructure more efficient over time. There are a number of models. I just gave PyTorch as an example. Open Compute is another model that has worked really well for us in this way, both to incorporate both innovation and scale efficiency into our own infrastructure. So I think that there’s, our incentives I think are basically aligned towards moving in this direction. Now that said, there’s a lot to figure out, right? So when you asked if there are going to be other opportunities, I hope so. I can’t speak to what all those things might be now. This is all quite early in getting developed. The better we do at the foundational work, the more opportunities I think that will come and present themselves. So I think that that’s all stuff that we need to figure out. But at least at the base level, I think we’re generally incentivized to move in this direction. And we also need to figure out how to go in that direction over time. I mean, I mentioned LLaMA before and I also want to be clear that while I’m talking about helping contribute to an open ecosystem, LLaMA is a model that we only really made available to researchers and there’s a lot of really good stuff that’s happening there. But a lot of the work that we’re doing, I think, we would aspire to and hope to make even more open than that. So, we’ll need to figure out a way to do that.

[P] I Trained a Model to Generate Video Game Pages
reddit
LLM Vibe Score0
Human Vibe Score1
pcvisionThis week

[P] I Trained a Model to Generate Video Game Pages

These past two months I've been working on a project I've called THIS GAME DOES NOT EXIST. I've always wanted to try building something with generative A.I. so this project scratched that itch for me. Here's a video with a few of my favourites read by voice actors: https://www.youtube.com/watch?v=\mTWMLhpJoA &#x200B; THIS GAME DOES NOT EXIST is an experiment in generative artificial intelligence. This site contains 130 video game pages that were generated using an implementation of OpenAI's Generative Pre-trained Transformer 2 (GPT-2) to generate text and a simple implementation of generative adversarial networks (GAN) to generate header images and "screenshots". To generate the names, descriptions, publishers, and developers of the games I finetuned the HuggingFace implementation of GPT-2. I used the Steam Store Games (Clean dataset) from Kaggle with slight modifications and preprocessing.Here is what one training sample looks like: Half-LifeValve ValveNamed Game of the Year by over 50 publications, Valve's debut title blends action and adventure with award-winning technology to create a frighteningly realistic world where players must think to survive. Also includes an exciting multiplayer mode that allows you to play against friends and enemies around the world. The model uses the tokens (e.g. and ) to prompt each class of data while keeping context during the entire generation. Image generation was done by training a custom GAN very similar to the architecture seen in the PyTorch DCGAN Tutorial which was built to generate faces. I created two models for this site: one for generating the header images and one for generating multiple screenshots for each game.To assemble the dataset I wrote a script that downloads the images from the URLs in the Steam Store Games (Clean dataset) dataset. Due to my lack of resources and time to put into this project, the image generation is less than ideal. You may notice though, that the header image model will generate artifacts in images that look like the titles of games, and the screenshot image model with generate what looks like levels of a 2D platformer.

[P] Open-source Neural Search framework to implement semantic search & multimedia search. Just released 2.0, seeking your feedback.
reddit
LLM Vibe Score0
Human Vibe Score0
opensourcecolumbusThis week

[P] Open-source Neural Search framework to implement semantic search & multimedia search. Just released 2.0, seeking your feedback.

I heard your feedback on 1.0 release post on my project Jina, many people were keen to use Jina for multimedia search because that's where use of Neural Networks makes significant difference. So I focused on that part and I was able to transform it from 1.0 to 2.0 within 3 months. Last post on 1.0 release to give you some idea what this project is about Actually, I should say - "'we' made this", because there were more than 155 contributors who did it, not just me. The primary changes we made We saw MachineLearning beginners struggle in using Jina 1.0, so we separated the codebase where Machine Learning expertise is required(jina-hub) and the one which MachineLearning beginners can use(the jina core). Now ML beginners don't need to worry about jina-hub and can use jina hub packages directly to implement ML specific tasks without the need to understand advanced ML concepts. While advanced ML users can create their own jina-hub packages. We cut down a lots of abstractions to make it easy to use for beginners Made python APIs more intuitive to use Improved performance(3.6x faster on startup) Here's Jina 2.0 and here's Jina 1.0. I seek feedback from people who are looking at this project for the first time, as well as people who have tried their hands before but had some challenges in using it. Few questions, I'm seeking answers to Do you feel that we have reduced complexity by a lot of margin? How easy it is to use for a beginner now? What questions are still unanswered?

[D] Playing big league at home on a budget?
reddit
LLM Vibe Score0
Human Vibe Score0.778
ballerburg9005This week

[D] Playing big league at home on a budget?

I am a hobbyist and my Nvidia 660 is 10 years old and only has 2GB. Obviously that isn't going to cut it nowadays anymore. I am thinking about options here. I don't have thousands and thousands of dollars. And I highly doubt that spending close to a thousand dollars on a brand new card is still viable in 2020-2022. I wanted to use Wavenet today and then found out about Melnet. I mean, maybe I could run Wavenet but nobody in their right mind wants to after hearing Melnet results. On Github this one guy complained he couldn't get his implementation to work due to OOM with 2x 2080 RTX, which he bought solely for this purpose. Then on the other repo the guy casually mentioned that tier XY doesn't fit with some 10 year old lowfi dataset, even with batch size 1, on a 16GB Tesla P100. The wisdom for OOM has always been "decrease batch size". But as far as I can tell, for most of any of the interesting stuff in the last 8 years or so you simply can't decrease batch size. Either because batch sizes are already so tiny, or because the code is written in a way that would require you to somehow turn it inside out, probably involving extreme knowledge of higher mathematics. I am a hobbyist, not a researcher. I am happy if I crudely can grasp what is going on. Most of anything in the field suffers from exactly the same issue: It simply won't run without utterly absurd amounts of VRAM. So what about buying shitty cheapo AMD GPUs with lots of VRAM? This seems to be the sensible choice if you want to be able to run anything noteworthy at all that comes up in the next 2 years and maybe beyond. People say, don't but AMD its slow and it sucks, but those are apparently the same people that buy a 16GB Titan GPU for $1500 three times on Ebay without hesitation, when there are also 16GB AMD GPUs for $300. How much slower are AMD GPUs really? Let's say they are 5 times cheaper so they could be just 5 times slower. So I have to train my model over night instead of seeing the result in the afternoon. That would be totally awesome!; given that the alternative is to buy a $300 Nvidia GPU, which has maybe 4 or 6GB and simply can't run the code without running out of memory. And say $300 is not enough, let's buy a $700 RTX 3080. It still only has 10GB of VRAM not even 16GB. Then its just as useless! What's the point of buying a fast GPU if it can't even run the code? I don't know how much slower AMD GPUs really are. Maybe they are not 5x but 50x slower. Then of course training a model that was developed on some 64GB Tesla might take month and years. But maybe speed is not the issue, only memory. I have seen some stuff even being optimized for CPU, apparently because there weren't any big enough GPUs around. I don't really know how viable that can be (it seems rarely if ever it is), I have no experience. And what about renting AWS? Let's say, I am a beginner and I want to toy around for a week and probably max out 4 Teslas like 80% of the time without really getting anywhere. How expensive is that? $25, $50, $100, $500? (Found the answer: fucking $2000 https://aws.amazon.com/ec2/instance-types/p3/ ) Ok, so AWS is bullshit, here its 6x cheaper: https://vast.ai/console/create/ . They don't really have 4x 16GB V100 though, just one V100. $0.5 per hour 24 7 = $84 per month (there are more hidden cost like bandwidth, it doesn't seem to be huge but I never used this so don't take it at face value). On AWS the same is over $3 per hour. So a day is $12, this could be viable! (look at calculation below). There really isn't much info on the net about hardware requirements and performance for machine learning stuff. What bothers me the most is that people seem to be very ignorant of the VRAM issue. Either because they aren't looking ahead of what might come in 1-2 years. Or because they are simply so rich they have no issue spending thousands and thousands of dollars every year instead of just 500 every couple of years. Or maybe they are both. So, yeah, what are your thoughts? Here is what I found out just today: Until 2 years ago, tensorflow and pytorch wouldn't work with AMD cards, but this has changed. https://rocmdocs.amd.com/en/latest/Deep_learning/Deep-learning.html For older cards though, ROCm only works with certain CPUs: it needs PCIe 3.0 with atomics (see: https://github.com/RadeonOpenCompute/ROCm ). So you can't simply buy any 16GB card for $300 on Ebay like I suggested, even if it supports ROCm, because it will only work for "newer" PCs. The newer GFX9 AMD cards (like Radeon VII and Vega) don't suffer from this problem and work with PCIe 2.0 again... Although I have seen 16GB Vega cards for like $350 on Ebay, I think that is a pretty rare catch. However looking 1-2 years in the future, this is great because Radeon VII prices will be hugely inflated by Nvidia 3000 series hype (maybe down to $180 even) and maybe the next gen cards from AMD even have 24 or 32GB for $500-$1000 and can still run on old machines. According to this https://arxiv.org/pdf/1909.06842.pdf Radeon VII 16GB performs only half as good as Tesla V100 16GB, whereas V100 should be roughly along the lines of 11GB RTX 2080 Ti. So you could say that you get half the RAM, double the speed, double the price. I am not sure though if that holds. I think they were putting 16GB in those cards trying to push it for ML with ROCm, clearly addressing the problem of the time, but no one really jumped on the train and now Resnet shrinks RAM but needs more processing power. So they released 8GB cards again with slightly better performance, and I guess we are lucky if the next generation even has 16GB because games probably don't need it at all. Still though with Revnets and everything said in the comments, I think on a budget you are better on the safe side buying the card with the most amount of VRAM, rather than the most performance. Tomorrow some paper might come out that uses another method, then you can't trick-shrink your network anymore and then everyone needs to buy big ass cards again like it used to be and can do nothing but throw their fancy faster cards in the dumpster. Also the huge bulk of ML currently focuses on image processing, while sound has only been gaining real momentum recently and this will be followed by video processing and eventually human-alike thought processes that sit atop of all that and have not even been tackled yet. Its a rapidly evolving field, hard to predict what will come and stay. Running out of VRAM means total hardware failure, running slower just means waiting longer. If you just buy the newest card every year, its probably save to buy the fast card because things won't change that fast after all. If you buy a new card every 4 years or longer then just try to get as much VRAM as possible. Check this out: https://www.techspot.com/news/86811-gigabyte-accidentally-reveals-rtx-3070-16gb-rtx-3080.html There will be a 3070 16GB version! Let's compare renting one V100 at $12/day vs. buying a 3070 Ti 16GB: The 2080 Ti was 1.42x the price of the regular 2080 and released the next summer. So let's assume the same will be true to the 3070 Ti so it will cost $700. That is $30/month & $1.88/day for two years - $15/month & $0.94/day in four years (by which time you can probably rent some 32GB Tesla card for the same price and nothing recent runs on less anymore). If you max out your setup 24/7 all year, then power cost obviously becomes a huge factor to that figure. In my country running at 500W cost $4.21/day, or $1.60 / 9hrs overnight. If you live elsewhere it might be as much as a quarter of that price. Of course your PC may run 10h a day anyway, so its maybe just 300W plus, and an older graphics card is inefficient for games it eats more Watts to do the same things so you save some there as well. There is a lot to take into account if comparing. Anyway, factoring in power cost, to break even with buying the card vs. renting within two years, you would have to use it for at least 4 days a month, or almost 2 weeks every 3 month. If you use it less than that, you maybe have a nice new graphics card and less hassle with pushing stuff back and forth onto servers all the time. But it would have been more economic to rent. So renting isn't that bad after all. Overall if you are thinking about having this as your hobby, you could say that it will cost you at least $30 per month, if not $50 or more (when keeping up to date with cards every 2 instead of 4 years + using it more cost more power). I think that is quite hefty. Personally I am not even invested enough into this even if it wasn't over my finances. I want a new card of course and also play some new games, but I don't really need to. There are a lot of other (more) important things I am interested in, that are totally free.

[D] What is your honest experience with reinforcement learning?
reddit
LLM Vibe Score0
Human Vibe Score1
Starks-TechnologyThis week

[D] What is your honest experience with reinforcement learning?

In my personal experience, SOTA RL algorithms simply don't work. I've tried working with reinforcement learning for over 5 years. I remember when Alpha Go defeated the world famous Go player, Lee Sedol, and everybody thought RL would take the ML community by storm. Yet, outside of toy problems, I've personally never found a practical use-case of RL. What is your experience with it? Aside from Ad recommendation systems and RLHF, are there legitimate use-cases of RL? Or, was it all hype? Edit: I know a lot about AI. I built NexusTrade, an AI-Powered automated investing tool that lets non-technical users create, update, and deploy their trading strategies. I’m not an idiot nor a noob; RL is just ridiculously hard. Edit 2: Since my comments are being downvoted, here is a link to my article that better describes my position. It's not that I don't understand RL. I released my open-source code and wrote a paper on it. It's the fact that it's EXTREMELY difficult to understand. Other deep learning algorithms like CNNs (including ResNets), RNNs (including GRUs and LSTMs), Transformers, and GANs are not hard to understand. These algorithms work and have practical use-cases outside of the lab. Traditional SOTA RL algorithms like PPO, DDPG, and TD3 are just very hard. You need to do a bunch of research to even implement a toy problem. In contrast, the decision transformer is something anybody can implement, and it seems to match or surpass the SOTA. You don't need two networks battling each other. You don't have to go through hell to debug your network. It just naturally learns the best set of actions in an auto-regressive manner. I also didn't mean to come off as arrogant or imply that RL is not worth learning. I just haven't seen any real-world, practical use-cases of it. I simply wanted to start a discussion, not claim that I know everything. Edit 3: There's a shockingly number of people calling me an idiot for not fully understanding RL. You guys are wayyy too comfortable calling people you disagree with names. News-flash, not everybody has a PhD in ML. My undergraduate degree is in biology. I self-taught myself the high-level maths to understand ML. I'm very passionate about the field; I just have VERY disappointing experiences with RL. Funny enough, there are very few people refuting my actual points. To summarize: Lack of real-world applications Extremely complex and inaccessible to 99% of the population Much harder than traditional DL algorithms like CNNs, RNNs, and GANs Sample inefficiency and instability Difficult to debug Better alternatives, such as the Decision Transformer Are these not legitimate criticisms? Is the purpose of this sub not to have discussions related to Machine Learning? To the few commenters that aren't calling me an idiot...thank you! Remember, it costs you nothing to be nice! Edit 4: Lots of people seem to agree that RL is over-hyped. Unfortunately those comments are downvoted. To clear up some things: We've invested HEAVILY into reinforcement learning. All we got from this investment is a robot that can be super-human at (some) video games. AlphaFold did not use any reinforcement learning. SpaceX doesn't either. I concede that it can be useful for robotics, but still argue that it's use-cases outside the lab are extremely limited. If you're stumbling on this thread and curious about an RL alternative, check out the Decision Transformer. It can be used in any situation that a traditional RL algorithm can be used. Final Edit: To those who contributed more recently, thank you for the thoughtful discussion! From what I learned, model-based models like Dreamer and IRIS MIGHT have a future. But everybody who has actually used model-free models like DDPG unanimously agree that they suck and don’t work.

[D] The current and future state of AI/ML is shockingly demoralizing with little hope of redemption
reddit
LLM Vibe Score0
Human Vibe Score1
Flaky_Suit_8665This week

[D] The current and future state of AI/ML is shockingly demoralizing with little hope of redemption

I recently encountered the PaLM (Scaling Language Modeling with Pathways) paper from Google Research and it opened up a can of worms of ideas I’ve felt I’ve intuitively had for a while, but have been unable to express – and I know I can’t be the only one. Sometimes I wonder what the original pioneers of AI – Turing, Neumann, McCarthy, etc. – would think if they could see the state of AI that we’ve gotten ourselves into. 67 authors, 83 pages, 540B parameters in a model, the internals of which no one can say they comprehend with a straight face, 6144 TPUs in a commercial lab that no one has access to, on a rig that no one can afford, trained on a volume of data that a human couldn’t process in a lifetime, 1 page on ethics with the same ideas that have been rehashed over and over elsewhere with no attempt at a solution – bias, racism, malicious use, etc. – for purposes that who asked for? When I started my career as an AI/ML research engineer 2016, I was most interested in two types of tasks – 1.) those that most humans could do but that would universally be considered tedious and non-scalable. I’m talking image classification, sentiment analysis, even document summarization, etc. 2.) tasks that humans lack the capacity to perform as well as computers for various reasons – forecasting, risk analysis, game playing, and so forth. I still love my career, and I try to only work on projects in these areas, but it’s getting harder and harder. This is because, somewhere along the way, it became popular and unquestionably acceptable to push AI into domains that were originally uniquely human, those areas that sit at the top of Maslows’s hierarchy of needs in terms of self-actualization – art, music, writing, singing, programming, and so forth. These areas of endeavor have negative logarithmic ability curves – the vast majority of people cannot do them well at all, about 10% can do them decently, and 1% or less can do them extraordinarily. The little discussed problem with AI-generation is that, without extreme deterrence, we will sacrifice human achievement at the top percentile in the name of lowering the bar for a larger volume of people, until the AI ability range is the norm. This is because relative to humans, AI is cheap, fast, and infinite, to the extent that investments in human achievement will be watered down at the societal, educational, and individual level with each passing year. And unlike AI gameplay which superseded humans decades ago, we won’t be able to just disqualify the machines and continue to play as if they didn’t exist. Almost everywhere I go, even this forum, I encounter almost universal deference given to current SOTA AI generation systems like GPT-3, CODEX, DALL-E, etc., with almost no one extending their implications to its logical conclusion, which is long-term convergence to the mean, to mediocrity, in the fields they claim to address or even enhance. If you’re an artist or writer and you’re using DALL-E or GPT-3 to “enhance” your work, or if you’re a programmer saying, “GitHub Co-Pilot makes me a better programmer?”, then how could you possibly know? You’ve disrupted and bypassed your own creative process, which is thoughts -> (optionally words) -> actions -> feedback -> repeat, and instead seeded your canvas with ideas from a machine, the provenance of which you can’t understand, nor can the machine reliably explain. And the more you do this, the more you make your creative processes dependent on said machine, until you must question whether or not you could work at the same level without it. When I was a college student, I often dabbled with weed, LSD, and mushrooms, and for a while, I thought the ideas I was having while under the influence were revolutionary and groundbreaking – that is until took it upon myself to actually start writing down those ideas and then reviewing them while sober, when I realized they weren’t that special at all. What I eventually determined is that, under the influence, it was impossible for me to accurately evaluate the drug-induced ideas I was having because the influencing agent the generates the ideas themselves was disrupting the same frame of reference that is responsible evaluating said ideas. This is the same principle of – if you took a pill and it made you stupider, would even know it? I believe that, especially over the long-term timeframe that crosses generations, there’s significant risk that current AI-generation developments produces a similar effect on humanity, and we mostly won’t even realize it has happened, much like a frog in boiling water. If you have children like I do, how can you be aware of the the current SOTA in these areas, project that 20 to 30 years, and then and tell them with a straight face that it is worth them pursuing their talent in art, writing, or music? How can you be honest and still say that widespread implementation of auto-correction hasn’t made you and others worse and worse at spelling over the years (a task that even I believe most would agree is tedious and worth automating). Furthermore, I’ve yet to set anyone discuss the train – generate – train - generate feedback loop that long-term application of AI-generation systems imply. The first generations of these models were trained on wide swaths of web data generated by humans, but if these systems are permitted to continually spit out content without restriction or verification, especially to the extent that it reduces or eliminates development and investment in human talent over the long term, then what happens to the 4th or 5th generation of models? Eventually we encounter this situation where the AI is being trained almost exclusively on AI-generated content, and therefore with each generation, it settles more and more into the mean and mediocrity with no way out using current methods. By the time that happens, what will we have lost in terms of the creative capacity of people, and will we be able to get it back? By relentlessly pursuing this direction so enthusiastically, I’m convinced that we as AI/ML developers, companies, and nations are past the point of no return, and it mostly comes down the investments in time and money that we’ve made, as well as a prisoner’s dilemma with our competitors. As a society though, this direction we’ve chosen for short-term gains will almost certainly make humanity worse off, mostly for those who are powerless to do anything about it – our children, our grandchildren, and generations to come. If you’re an AI researcher or a data scientist like myself, how do you turn things back for yourself when you’ve spent years on years building your career in this direction? You’re likely making near or north of $200k annually TC and have a family to support, and so it’s too late, no matter how you feel about the direction the field has gone. If you’re a company, how do you standby and let your competitors aggressively push their AutoML solutions into more and more markets without putting out your own? Moreover, if you’re a manager or thought leader in this field like Jeff Dean how do you justify to your own boss and your shareholders your team’s billions of dollars in AI investment while simultaneously balancing ethical concerns? You can’t – the only answer is bigger and bigger models, more and more applications, more and more data, and more and more automation, and then automating that even further. If you’re a country like the US, how do responsibly develop AI while your competitors like China single-mindedly push full steam ahead without an iota of ethical concern to replace you in numerous areas in global power dynamics? Once again, failing to compete would be pre-emptively admitting defeat. Even assuming that none of what I’ve described here happens to such an extent, how are so few people not taking this seriously and discounting this possibility? If everything I’m saying is fear-mongering and non-sense, then I’d be interested in hearing what you think human-AI co-existence looks like in 20 to 30 years and why it isn’t as demoralizing as I’ve made it out to be. &#x200B; EDIT: Day after posting this -- this post took off way more than I expected. Even if I received 20 - 25 comments, I would have considered that a success, but this went much further. Thank you to each one of you that has read this post, even more so if you left a comment, and triply so for those who gave awards! I've read almost every comment that has come in (even the troll ones), and am truly grateful for each one, including those in sharp disagreement. I've learned much more from this discussion with the sub than I could have imagined on this topic, from so many perspectives. While I will try to reply as many comments as I can, the sheer comment volume combined with limited free time between work and family unfortunately means that there are many that I likely won't be able to get to. That will invariably include some that I would love respond to under the assumption of infinite time, but I will do my best, even if the latency stretches into days. Thank you all once again!

[D] The machine learning community has a toxicity problem
reddit
LLM Vibe Score0
Human Vibe Score1
yusuf-bengioThis week

[D] The machine learning community has a toxicity problem

It is omnipresent! First of all, the peer-review process is broken. Every fourth NeurIPS submission is put on arXiv. There are DeepMind researchers publicly going after reviewers who are criticizing their ICLR submission. On top of that, papers by well-known institutes that were put on arXiv are accepted at top conferences, despite the reviewers agreeing on rejection. In contrast, vice versa, some papers with a majority of accepts are overruled by the AC. (I don't want to call any names, just have a look the openreview page of this year's ICRL). Secondly, there is a reproducibility crisis. Tuning hyperparameters on the test set seem to be the standard practice nowadays. Papers that do not beat the current state-of-the-art method have a zero chance of getting accepted at a good conference. As a result, hyperparameters get tuned and subtle tricks implemented to observe a gain in performance where there isn't any. Thirdly, there is a worshiping problem. Every paper with a Stanford or DeepMind affiliation gets praised like a breakthrough. For instance, BERT has seven times more citations than ULMfit. The Google affiliation gives so much credibility and visibility to a paper. At every ICML conference, there is a crowd of people in front of every DeepMind poster, regardless of the content of the work. The same story happened with the Zoom meetings at the virtual ICLR 2020. Moreover, NeurIPS 2020 had twice as many submissions as ICML, even though both are top-tier ML conferences. Why? Why is the name "neural" praised so much? Next, Bengio, Hinton, and LeCun are truly deep learning pioneers but calling them the "godfathers" of AI is insane. It has reached the level of a cult. Fourthly, the way Yann LeCun talked about biases and fairness topics was insensitive. However, the toxicity and backlash that he received are beyond any reasonable quantity. Getting rid of LeCun and silencing people won't solve any issue. Fifthly, machine learning, and computer science in general, have a huge diversity problem. At our CS faculty, only 30% of undergrads and 15% of the professors are women. Going on parental leave during a PhD or post-doc usually means the end of an academic career. However, this lack of diversity is often abused as an excuse to shield certain people from any form of criticism. Reducing every negative comment in a scientific discussion to race and gender creates a toxic environment. People are becoming afraid to engage in fear of being called a racist or sexist, which in turn reinforces the diversity problem. Sixthly, moral and ethics are set arbitrarily. The U.S. domestic politics dominate every discussion. At this very moment, thousands of Uyghurs are put into concentration camps based on computer vision algorithms invented by this community, and nobody seems even remotely to care. Adding a "broader impact" section at the end of every people will not make this stop. There are huge shitstorms because a researcher wasn't mentioned in an article. Meanwhile, the 1-billion+ people continent of Africa is virtually excluded from any meaningful ML discussion (besides a few Indaba workshops). Seventhly, there is a cut-throat publish-or-perish mentality. If you don't publish 5+ NeurIPS/ICML papers per year, you are a looser. Research groups have become so large that the PI does not even know the name of every PhD student anymore. Certain people submit 50+ papers per year to NeurIPS. The sole purpose of writing a paper has become to having one more NeurIPS paper in your CV. Quality is secondary; passing the peer-preview stage has become the primary objective. Finally, discussions have become disrespectful. Schmidhuber calls Hinton a thief, Gebru calls LeCun a white supremacist, Anandkumar calls Marcus a sexist, everybody is under attack, but nothing is improved. Albert Einstein was opposing the theory of quantum mechanics. Can we please stop demonizing those who do not share our exact views. We are allowed to disagree without going for the jugular. The moment we start silencing people because of their opinion is the moment scientific and societal progress dies. Best intentions, Yusuf

[D]Stuck in AI Hell: What to do in post LLM world
reddit
LLM Vibe Score0
Human Vibe Score1
Educational_News_371This week

[D]Stuck in AI Hell: What to do in post LLM world

Hey Reddit, I’ve been in an AI/ML role for a few years now, and I’m starting to feel disconnected from the work. When I started, deep learning models were getting good, and I quickly fell in love with designing architectures, training models, and fine-tuning them for specific use cases. Seeing a loss curve finally converge, experimenting with layers, and debugging training runs—it all felt like a craft, a blend of science and creativity. I enjoyed implementing research papers to see how things worked under the hood. Backprop, gradients, optimization—it was a mental workout I loved. But these days, it feels like everything has shifted. LLMs dominate the scene, and instead of building and training models, the focus is on using pre-trained APIs, crafting prompt chains, and setting up integrations. Sure, there’s engineering involved, but it feels less like creating and more like assembling. I miss the hands-on nature of experimenting with architectures and solving math-heavy problems. It’s not just the creativity I miss. The economics of this new era also feel strange to me. Back when I started, compute was a luxury. We had limited GPUs, and a lot of the work was about being resourceful—quantizing models, distilling them, removing layers, and squeezing every bit of performance out of constrained setups. Now, it feels like no one cares about cost. We’re paying by tokens. Tokens! Who would’ve thought we’d get to a point where we’re not designing efficient models but feeding pre-trained giants like they’re vending machines? I get it—abstraction has always been part of the field. TensorFlow and PyTorch abstracted tensor operations, Python abstracts C. But deep learning still left room for creation. We weren’t just abstracting away math; we were solving it. We could experiment, fail, and tweak. Working with LLMs doesn’t feel the same. It’s like fitting pieces into a pre-defined puzzle instead of building the puzzle itself. I understand that LLMs are here to stay. They’re incredible tools, and I respect their potential to revolutionize industries. Building real-world products with them is still challenging, requiring a deep understanding of engineering, prompt design, and integrating them effectively into workflows. By no means is it an “easy” task. But the work doesn’t give me the same thrill. It’s not about solving math or optimization problems—it’s about gluing together APIs, tweaking outputs, and wrestling with opaque systems. It’s like we’ve traded craftsmanship for convenience. Which brings me to my questions: Is there still room for those of us who enjoy the deep work of model design and training? Or is this the inevitable evolution of the field, where everything converges on pre-trained systems? What use cases still need traditional ML expertise? Are there industries or problems that will always require specialized models instead of general-purpose LLMs? Am I missing the bigger picture here? LLMs feel like the “kernel” of a new computing paradigm, and we don’t fully understand their second- and third-order effects. Could this shift lead to new, exciting opportunities I’m just not seeing yet? How do you stay inspired when the focus shifts? I still love AI, but I miss the feeling of building something from scratch. Is this just a matter of adapting my mindset, or should I seek out niches where traditional ML still thrives? I’m not asking this to rant (though clearly, I needed to get some of this off my chest). I want to figure out where to go next from here. If you’ve been in AI/ML long enough to see major shifts—like the move from feature engineering to deep learning—how did you navigate them? What advice would you give someone in my position? And yeah, before anyone roasts me for using an LLM to structure this post (guilty!), I just wanted to get my thoughts out in a coherent way. Guess that’s a sign of where we’re headed, huh? Thanks for reading, and I’d love to hear your thoughts! TL;DR: I entered AI during the deep learning boom, fell in love with designing and training models, and thrived on creativity, math, and optimization. Now it feels like the field is all about tweaking prompts and orchestrating APIs for pre-trained LLMs. I miss the thrill of crafting something unique. Is there still room for people who enjoy traditional ML, or is this just the inevitable evolution of the field? How do you stay inspired amidst such shifts? Update: Wow, this blew up. Thanks everyone for your comments and suggestions. I really like some of those. This thing was on my mind for a long time, glad that I put it here. Thanks again!

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[D] Accessibility of Basic Models to Non-Technicals
reddit
LLM Vibe Score0
Human Vibe Score0
wildekansThis week

[D] Accessibility of Basic Models to Non-Technicals

Hello /r/machinelearning! I'm doing some research on easily generated models by non-technical/statistical people. It would be awesome if some of you could answer a quick questionnaire: If you're a machine learning developer/data scientist etc.: a) Has your manager/product lead etc. ever insist that you build a model on a correlation you felt wasn't there? b) Do you think if that people had a way to verify the lack of correlation through a naive model (random forest, svc, etc.) that it would have changed the situation? (Or, if you were able to show them the results) c) Would you want this technology for yourself, or wish that your company would have access to it? If you're a non-technical person (small business developer, student, non-tech entrepreneur, etc.): a) Have you ever not pursued a potential machine learning/data solution or feature because you weren't willing to invest the resources to see if it was viable? b) Would being able to verify correlations in your data (or lack thereof!) entice you to pursue possible machine learning solutions? c) Even if your previous answers were no, would you be interested in having this technology? Thanks in advance for all of the responses, I will personally read and respond to each one of you thoughtful enough to give me a response. Also, I hope this post will spark an interesting conversation about the barrier of entry to AI/machine learning.

[P] Jarvislabs.ai - An Affordable GPU Cloud with Fast launch, Pause and Resume. Scale GPUs post creation. A100/RTX6K/RTX5K
reddit
LLM Vibe Score0
Human Vibe Score1
vishnu_subramaniannThis week

[P] Jarvislabs.ai - An Affordable GPU Cloud with Fast launch, Pause and Resume. Scale GPUs post creation. A100/RTX6K/RTX5K

For the last few years, I have been learning and practicing Deep Learning. Participated in several Kaggle competitions and won few medals. During all these years, I tried several cloud platforms and on-premise systems. Some of them offered simplicity, flexibility, and affordability. But very few to none offered all of these in one platform. After struggling with different platforms, I know what I would need as a DL researcher. That gave birth to jarvislabs.ai with the aim of being simple and affordable. I along with my friends started working on this project a year back. Due to Covid, executing the project became more challenging. As first-time entrepreneurs, we underestimated the complexity of the problem at hand but with persistence, we were able to launch a beta version of the product in December 2020. With some of the amazing feedback from our early adopters, we have been able to make the product smoother. We would love to invite you all to come and try the platform. Features 1 click Jupyter Lab < \[30 seconds\] Pause the instance and Resume from where you left. SSH to the instance. Scale GPUs, storage and change GPU type on resume. Auto-Pause using jarviscloud.pause() in your code, so you can catch up some good night’s sleep while your model trains. Pay per usage – Minute Billing \[After first 15 minutes\] Competitive pricing \[Lowest to our Knowledge\]. &#x200B; Pricing |GPU Type|GPU RAM|Price -$/hr| |:-|:-|:-| |RTX 5000|16 GB|0.49| |RTX 6000|24 GB|0.99| |A100|40 GB|2.39| &#x200B; Talk to us We will be happy to assist you in spinning your first instance and many more. You can use one of these platforms to reach us. Chat option on cloud.jarvislabs.ai Email us - hello@jarvislabs.ai Comment here. We have come a long way, but we understand that a lot more has to be done. We have listed down all the upcoming product features here. Deep learning and AI are evolving and how we would use the cloud platforms could evolve in the coming years. Understanding this, we develop in the open by constantly keeping in touch with our users. Please help us in shaping Jarvislabs.ai with any valuable suggestions/feedback.

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.
reddit
LLM Vibe Score0
Human Vibe Score0.765
hardmaruThis week

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.

Schmidhuber interview expressing his views on the future of AI and AGI. Original source. I think the interview is of interest to r/MachineLearning, and presents an alternate view, compared to other influential leaders in AI. Juergen Schmidhuber, Renowned 'Father Of Modern AI,' Says His Life’s Work Won't Lead To Dystopia May 23, 2023. Contributed by Hessie Jones. Amid the growing concern about the impact of more advanced artificial intelligence (AI) technologies on society, there are many in the technology community who fear the implications of the advancements in Generative AI if they go unchecked. Dr. Juergen Schmidhuber, a renowned scientist, artificial intelligence researcher and widely regarded as one of the pioneers in the field, is more optimistic. He declares that many of those who suddenly warn against the dangers of AI are just seeking publicity, exploiting the media’s obsession with killer robots which has attracted more attention than “good AI” for healthcare etc. The potential to revolutionize various industries and improve our lives is clear, as are the equal dangers if bad actors leverage the technology for personal gain. Are we headed towards a dystopian future, or is there reason to be optimistic? I had a chance to sit down with Dr. Juergen Schmidhuber to understand his perspective on this seemingly fast-moving AI-train that will leap us into the future. As a teenager in the 1970s, Juergen Schmidhuber became fascinated with the idea of creating intelligent machines that could learn and improve on their own, becoming smarter than himself within his lifetime. This would ultimately lead to his groundbreaking work in the field of deep learning. In the 1980s, he studied computer science at the Technical University of Munich (TUM), where he earned his diploma in 1987. His thesis was on the ultimate self-improving machines that, not only, learn through some pre-wired human-designed learning algorithm, but also learn and improve the learning algorithm itself. Decades later, this became a hot topic. He also received his Ph.D. at TUM in 1991 for work that laid some of the foundations of modern AI. Schmidhuber is best known for his contributions to the development of recurrent neural networks (RNNs), the most powerful type of artificial neural network that can process sequential data such as speech and natural language. With his students Sepp Hochreiter, Felix Gers, Alex Graves, Daan Wierstra, and others, he published architectures and training algorithms for the long short-term memory (LSTM), a type of RNN that is widely used in natural language processing, speech recognition, video games, robotics, and other applications. LSTM has become the most cited neural network of the 20th century, and Business Week called it "arguably the most commercial AI achievement." Throughout his career, Schmidhuber has received various awards and accolades for his groundbreaking work. In 2013, he was awarded the Helmholtz Prize, which recognizes significant contributions to the field of machine learning. In 2016, he was awarded the IEEE Neural Network Pioneer Award for "pioneering contributions to deep learning and neural networks." The media have often called him the “father of modern AI,” because the most cited neural networks all build on his lab’s work. He is quick to point out, however, that AI history goes back centuries. Despite his many accomplishments, at the age of 60, he feels mounting time pressure towards building an Artificial General Intelligence within his lifetime and remains committed to pushing the boundaries of AI research and development. He is currently director of the KAUST AI Initiative, scientific director of the Swiss AI Lab IDSIA, and co-founder and chief scientist of AI company NNAISENSE, whose motto is "AI∀" which is a math-inspired way of saying "AI For All." He continues to work on cutting-edge AI technologies and applications to improve human health and extend human lives and make lives easier for everyone. The following interview has been edited for clarity. Jones: Thank you Juergen for joining me. You have signed letters warning about AI weapons. But you didn't sign the recent publication, "Pause Gigantic AI Experiments: An Open Letter"? Is there a reason? Schmidhuber: Thank you Hessie. Glad to speak with you. I have realized that many of those who warn in public against the dangers of AI are just seeking publicity. I don't think the latest letter will have any significant impact because many AI researchers, companies, and governments will ignore it completely. The proposal frequently uses the word "we" and refers to "us," the humans. But as I have pointed out many times in the past, there is no "we" that everyone can identify with. Ask 10 different people, and you will hear 10 different opinions about what is "good." Some of those opinions will be completely incompatible with each other. Don't forget the enormous amount of conflict between the many people. The letter also says, "If such a pause cannot be quickly put in place, governments should intervene and impose a moratorium." The problem is that different governments have ALSO different opinions about what is good for them and for others. Great Power A will say, if we don't do it, Great Power B will, perhaps secretly, and gain an advantage over us. The same is true for Great Powers C and D. Jones: Everyone acknowledges this fear surrounding current generative AI technology. Moreover, the existential threat of this technology has been publicly acknowledged by Sam Altman, CEO of OpenAI himself, calling for AI regulation. From your perspective, is there an existential threat? Schmidhuber: It is true that AI can be weaponized, and I have no doubt that there will be all kinds of AI arms races, but AI does not introduce a new quality of existential threat. The threat coming from AI weapons seems to pale in comparison to the much older threat from nuclear hydrogen bombs that don’t need AI at all. We should be much more afraid of half-century-old tech in the form of H-bomb rockets. The Tsar Bomba of 1961 had almost 15 times more destructive power than all weapons of WW-II combined. Despite the dramatic nuclear disarmament since the 1980s, there are still more than enough nuclear warheads to wipe out human civilization within two hours, without any AI I’m much more worried about that old existential threat than the rather harmless AI weapons. Jones: I realize that while you compare AI to the threat of nuclear bombs, there is a current danger that a current technology can be put in the hands of humans and enable them to “eventually” exact further harms to individuals of group in a very precise way, like targeted drone attacks. You are giving people a toolset that they've never had before, enabling bad actors, as some have pointed out, to be able to do a lot more than previously because they didn't have this technology. Schmidhuber: Now, all that sounds horrible in principle, but our existing laws are sufficient to deal with these new types of weapons enabled by AI. If you kill someone with a gun, you will go to jail. Same if you kill someone with one of these drones. Law enforcement will get better at understanding new threats and new weapons and will respond with better technology to combat these threats. Enabling drones to target persons from a distance in a way that requires some tracking and some intelligence to perform, which has traditionally been performed by skilled humans, to me, it seems is just an improved version of a traditional weapon, like a gun, which is, you know, a little bit smarter than the old guns. But, in principle, all of that is not a new development. For many centuries, we have had the evolution of better weaponry and deadlier poisons and so on, and law enforcement has evolved their policies to react to these threats over time. So, it's not that we suddenly have a new quality of existential threat and it's much more worrisome than what we have had for about six decades. A large nuclear warhead doesn’t need fancy face recognition to kill an individual. No, it simply wipes out an entire city with ten million inhabitants. Jones: The existential threat that’s implied is the extent to which humans have control over this technology. We see some early cases of opportunism which, as you say, tends to get more media attention than positive breakthroughs. But you’re implying that this will all balance out? Schmidhuber: Historically, we have a long tradition of technological breakthroughs that led to advancements in weapons for the purpose of defense but also for protection. From sticks, to rocks, to axes to gunpowder to cannons to rockets… and now to drones… this has had a drastic influence on human history but what has been consistent throughout history is that those who are using technology to achieve their own ends are themselves, facing the same technology because the opposing side is learning to use it against them. And that's what has been repeated in thousands of years of human history and it will continue. I don't see the new AI arms race as something that is remotely as existential a threat as the good old nuclear warheads. You said something important, in that some people prefer to talk about the downsides rather than the benefits of this technology, but that's misleading, because 95% of all AI research and AI development is about making people happier and advancing human life and health. Jones: Let’s touch on some of those beneficial advances in AI research that have been able to radically change present day methods and achieve breakthroughs. Schmidhuber: All right! For example, eleven years ago, our team with my postdoc Dan Ciresan was the first to win a medical imaging competition through deep learning. We analyzed female breast cells with the objective to determine harmless cells vs. those in the pre-cancer stage. Typically, a trained oncologist needs a long time to make these determinations. Our team, who knew nothing about cancer, were able to train an artificial neural network, which was totally dumb in the beginning, on lots of this kind of data. It was able to outperform all the other methods. Today, this is being used not only for breast cancer, but also for radiology and detecting plaque in arteries, and many other things. Some of the neural networks that we have developed in the last 3 decades are now prevalent across thousands of healthcare applications, detecting Diabetes and Covid-19 and what not. This will eventually permeate across all healthcare. The good consequences of this type of AI are much more important than the click-bait new ways of conducting crimes with AI. Jones: Adoption is a product of reinforced outcomes. The massive scale of adoption either leads us to believe that people have been led astray, or conversely, technology is having a positive effect on people’s lives. Schmidhuber: The latter is the likely case. There's intense commercial pressure towards good AI rather than bad AI because companies want to sell you something, and you are going to buy only stuff you think is going to be good for you. So already just through this simple, commercial pressure, you have a tremendous bias towards good AI rather than bad AI. However, doomsday scenarios like in Schwarzenegger movies grab more attention than documentaries on AI that improve people’s lives. Jones: I would argue that people are drawn to good stories – narratives that contain an adversary and struggle, but in the end, have happy endings. And this is consistent with your comment on human nature and how history, despite its tendency for violence and destruction of humanity, somehow tends to correct itself. Let’s take the example of a technology, which you are aware – GANs – General Adversarial Networks, which today has been used in applications for fake news and disinformation. In actuality, the purpose in the invention of GANs was far from what it is used for today. Schmidhuber: Yes, the name GANs was created in 2014 but we had the basic principle already in the early 1990s. More than 30 years ago, I called it artificial curiosity. It's a very simple way of injecting creativity into a little two network system. This creative AI is not just trying to slavishly imitate humans. Rather, it’s inventing its own goals. Let me explain: You have two networks. One network is producing outputs that could be anything, any action. Then the second network is looking at these actions and it’s trying to predict the consequences of these actions. An action could move a robot, then something happens, and the other network is just trying to predict what will happen. Now we can implement artificial curiosity by reducing the prediction error of the second network, which, at the same time, is the reward of the first network. The first network wants to maximize its reward and so it will invent actions that will lead to situations that will surprise the second network, which it has not yet learned to predict well. In the case where the outputs are fake images, the first network will try to generate images that are good enough to fool the second network, which will attempt to predict the reaction of the environment: fake or real image, and it will try to become better at it. The first network will continue to also improve at generating images whose type the second network will not be able to predict. So, they fight each other. The 2nd network will continue to reduce its prediction error, while the 1st network will attempt to maximize it. Through this zero-sum game the first network gets better and better at producing these convincing fake outputs which look almost realistic. So, once you have an interesting set of images by Vincent Van Gogh, you can generate new images that leverage his style, without the original artist having ever produced the artwork himself. Jones: I see how the Van Gogh example can be applied in an education setting and there are countless examples of artists mimicking styles from famous painters but image generation from this instance that can happen within seconds is quite another feat. And you know this is how GANs has been used. What’s more prevalent today is a socialized enablement of generating images or information to intentionally fool people. It also surfaces new harms that deal with the threat to intellectual property and copyright, where laws have yet to account for. And from your perspective this was not the intention when the model was conceived. What was your motivation in your early conception of what is now GANs? Schmidhuber: My old motivation for GANs was actually very important and it was not to create deepfakes or fake news but to enable AIs to be curious and invent their own goals, to make them explore their environment and make them creative. Suppose you have a robot that executes one action, then something happens, then it executes another action, and so on, because it wants to achieve certain goals in the environment. For example, when the battery is low, this will trigger “pain” through hunger sensors, so it wants to go to the charging station, without running into obstacles, which will trigger other pain sensors. It will seek to minimize pain (encoded through numbers). Now the robot has a friend, the second network, which is a world model ––it’s a prediction machine that learns to predict the consequences of the robot’s actions. Once the robot has a good model of the world, it can use it for planning. It can be used as a simulation of the real world. And then it can determine what is a good action sequence. If the robot imagines this sequence of actions, the model will predict a lot of pain, which it wants to avoid. If it plays this alternative action sequence in its mental model of the world, then it will predict a rewarding situation where it’s going to sit on the charging station and its battery is going to load again. So, it'll prefer to execute the latter action sequence. In the beginning, however, the model of the world knows nothing, so how can we motivate the first network to generate experiments that lead to data that helps the world model learn something it didn’t already know? That’s what artificial curiosity is about. The dueling two network systems effectively explore uncharted environments by creating experiments so that over time the curious AI gets a better sense of how the environment works. This can be applied to all kinds of environments, and has medical applications. Jones: Let’s talk about the future. You have said, “Traditional humans won’t play a significant role in spreading intelligence across the universe.” Schmidhuber: Let’s first conceptually separate two types of AIs. The first type of AI are tools directed by humans. They are trained to do specific things like accurately detect diabetes or heart disease and prevent attacks before they happen. In these cases, the goal is coming from the human. More interesting AIs are setting their own goals. They are inventing their own experiments and learning from them. Their horizons expand and eventually they become more and more general problem solvers in the real world. They are not controlled by their parents, but much of what they learn is through self-invented experiments. A robot, for example, is rotating a toy, and as it is doing this, the video coming in through the camera eyes, changes over time and it begins to learn how this video changes and learns how the 3D nature of the toy generates certain videos if you rotate it a certain way, and eventually, how gravity works, and how the physics of the world works. Like a little scientist! And I have predicted for decades that future scaled-up versions of such AI scientists will want to further expand their horizons, and eventually go where most of the physical resources are, to build more and bigger AIs. And of course, almost all of these resources are far away from earth out there in space, which is hostile to humans but friendly to appropriately designed AI-controlled robots and self-replicating robot factories. So here we are not talking any longer about our tiny biosphere; no, we are talking about the much bigger rest of the universe. Within a few tens of billions of years, curious self-improving AIs will colonize the visible cosmos in a way that’s infeasible for humans. Those who don’t won’t have an impact. Sounds like science fiction, but since the 1970s I have been unable to see a plausible alternative to this scenario, except for a global catastrophe such as an all-out nuclear war that stops this development before it takes off. Jones: How long have these AIs, which can set their own goals — how long have they existed? To what extent can they be independent of human interaction? Schmidhuber: Neural networks like that have existed for over 30 years. My first simple adversarial neural network system of this kind is the one from 1990 described above. You don’t need a teacher there; it's just a little agent running around in the world and trying to invent new experiments that surprise its own prediction machine. Once it has figured out certain parts of the world, the agent will become bored and will move on to more exciting experiments. The simple 1990 systems I mentioned have certain limitations, but in the past three decades, we have also built more sophisticated systems that are setting their own goals and such systems I think will be essential for achieving true intelligence. If you are only imitating humans, you will never go beyond them. So, you really must give AIs the freedom to explore previously unexplored regions of the world in a way that no human is really predefining. Jones: Where is this being done today? Schmidhuber: Variants of neural network-based artificial curiosity are used today for agents that learn to play video games in a human-competitive way. We have also started to use them for automatic design of experiments in fields such as materials science. I bet many other fields will be affected by it: chemistry, biology, drug design, you name it. However, at least for now, these artificial scientists, as I like to call them, cannot yet compete with human scientists. I don’t think it’s going to stay this way but, at the moment, it’s still the case. Sure, AI has made a lot of progress. Since 1997, there have been superhuman chess players, and since 2011, through the DanNet of my team, there have been superhuman visual pattern recognizers. But there are other things where humans, at the moment at least, are much better, in particular, science itself. In the lab we have many first examples of self-directed artificial scientists, but they are not yet convincing enough to appear on the radar screen of the public space, which is currently much more fascinated with simpler systems that just imitate humans and write texts based on previously seen human-written documents. Jones: You speak of these numerous instances dating back 30 years of these lab experiments where these self-driven agents are deciding and learning and moving on once they’ve learned. And I assume that that rate of learning becomes even faster over time. What kind of timeframe are we talking about when this eventually is taken outside of the lab and embedded into society? Schmidhuber: This could still take months or even years :-) Anyway, in the not-too-distant future, we will probably see artificial scientists who are good at devising experiments that allow them to discover new, previously unknown physical laws. As always, we are going to profit from the old trend that has held at least since 1941: every decade compute is getting 100 times cheaper. Jones: How does this trend affect modern AI such as ChatGPT? Schmidhuber: Perhaps you know that all the recent famous AI applications such as ChatGPT and similar models are largely based on principles of artificial neural networks invented in the previous millennium. The main reason why they works so well now is the incredible acceleration of compute per dollar. ChatGPT is driven by a neural network called “Transformer” described in 2017 by Google. I am happy about that because a quarter century earlier in 1991 I had a particular Transformer variant which is now called the “Transformer with linearized self-attention”. Back then, not much could be done with it, because the compute cost was a million times higher than today. But today, one can train such models on half the internet and achieve much more interesting results. Jones: And for how long will this acceleration continue? Schmidhuber: There's no reason to believe that in the next 30 years, we won't have another factor of 1 million and that's going to be really significant. In the near future, for the first time we will have many not-so expensive devices that can compute as much as a human brain. The physical limits of computation, however, are much further out so even if the trend of a factor of 100 every decade continues, the physical limits (of 1051 elementary instructions per second and kilogram of matter) won’t be hit until, say, the mid-next century. Even in our current century, however, we’ll probably have many machines that compute more than all 10 billion human brains collectively and you can imagine, everything will change then! Jones: That is the big question. Is everything going to change? If so, what do you say to the next generation of leaders, currently coming out of college and university. So much of this change is already impacting how they study, how they will work, or how the future of work and livelihood is defined. What is their purpose and how do we change our systems so they will adapt to this new version of intelligence? Schmidhuber: For decades, people have asked me questions like that, because you know what I'm saying now, I have basically said since the 1970s, it’s just that today, people are paying more attention because, back then, they thought this was science fiction. They didn't think that I would ever come close to achieving my crazy life goal of building a machine that learns to become smarter than myself such that I can retire. But now many have changed their minds and think it's conceivable. And now I have two daughters, 23 and 25. People ask me: what do I tell them? They know that Daddy always said, “It seems likely that within your lifetimes, you will have new types of intelligence that are probably going to be superior in many ways, and probably all kinds of interesting ways.” How should they prepare for that? And I kept telling them the obvious: Learn how to learn new things! It's not like in the previous millennium where within 20 years someone learned to be a useful member of society, and then took a job for 40 years and performed in this job until she received her pension. Now things are changing much faster and we must learn continuously just to keep up. I also told my girls that no matter how smart AIs are going to get, learn at least the basics of math and physics, because that’s the essence of our universe, and anybody who understands this will have an advantage, and learn all kinds of new things more easily. I also told them that social skills will remain important, because most future jobs for humans will continue to involve interactions with other humans, but I couldn’t teach them anything about that; they know much more about social skills than I do. You touched on the big philosophical question about people’s purpose. Can this be answered without answering the even grander question: What’s the purpose of the entire universe? We don’t know. But what’s happening right now might be connected to the unknown answer. Don’t think of humans as the crown of creation. Instead view human civilization as part of a much grander scheme, an important step (but not the last one) on the path of the universe from very simple initial conditions towards more and more unfathomable complexity. Now it seems ready to take its next step, a step comparable to the invention of life itself over 3.5 billion years ago. Alas, don’t worry, in the end, all will be good! Jones: Let’s get back to this transformation happening right now with OpenAI. There are many questioning the efficacy and accuracy of ChatGPT, and are concerned its release has been premature. In light of the rampant adoption, educators have banned its use over concerns of plagiarism and how it stifles individual development. Should large language models like ChatGPT be used in school? Schmidhuber: When the calculator was first introduced, instructors forbade students from using it in school. Today, the consensus is that kids should learn the basic methods of arithmetic, but they should also learn to use the “artificial multipliers” aka calculators, even in exams, because laziness and efficiency is a hallmark of intelligence. Any intelligent being wants to minimize its efforts to achieve things. And that's the reason why we have tools, and why our kids are learning to use these tools. The first stone tools were invented maybe 3.5 million years ago; tools just have become more sophisticated over time. In fact, humans have changed in response to the properties of their tools. Our anatomical evolution was shaped by tools such as spears and fire. So, it's going to continue this way. And there is no permanent way of preventing large language models from being used in school. Jones: And when our children, your children graduate, what does their future work look like? Schmidhuber: A single human trying to predict details of how 10 billion people and their machines will evolve in the future is like a single neuron in my brain trying to predict what the entire brain and its tens of billions of neurons will do next year. 40 years ago, before the WWW was created at CERN in Switzerland, who would have predicted all those young people making money as YouTube video bloggers? Nevertheless, let’s make a few limited job-related observations. For a long time, people have thought that desktop jobs may require more intelligence than skills trade or handicraft professions. But now, it turns out that it's much easier to replace certain aspects of desktop jobs than replacing a carpenter, for example. Because everything that works well in AI is happening behind the screen currently, but not so much in the physical world. There are now artificial systems that can read lots of documents and then make really nice summaries of these documents. That is a desktop job. Or you give them a description of an illustration that you want to have for your article and pretty good illustrations are being generated that may need some minimal fine-tuning. But you know, all these desktop jobs are much easier to facilitate than the real tough jobs in the physical world. And it's interesting that the things people thought required intelligence, like playing chess, or writing or summarizing documents, are much easier for machines than they thought. But for things like playing football or soccer, there is no physical robot that can remotely compete with the abilities of a little boy with these skills. So, AI in the physical world, interestingly, is much harder than AI behind the screen in virtual worlds. And it's really exciting, in my opinion, to see that jobs such as plumbers are much more challenging than playing chess or writing another tabloid story. Jones: The way data has been collected in these large language models does not guarantee personal information has not been excluded. Current consent laws already are outdated when it comes to these large language models (LLM). The concern, rightly so, is increasing surveillance and loss of privacy. What is your view on this? Schmidhuber: As I have indicated earlier: are surveillance and loss of privacy inevitable consequences of increasingly complex societies? Super-organisms such as cities and states and companies consist of numerous people, just like people consist of numerous cells. These cells enjoy little privacy. They are constantly monitored by specialized "police cells" and "border guard cells": Are you a cancer cell? Are you an external intruder, a pathogen? Individual cells sacrifice their freedom for the benefits of being part of a multicellular organism. Similarly, for super-organisms such as nations. Over 5000 years ago, writing enabled recorded history and thus became its inaugural and most important invention. Its initial purpose, however, was to facilitate surveillance, to track citizens and their tax payments. The more complex a super-organism, the more comprehensive its collection of information about its constituents. 200 years ago, at least, the parish priest in each village knew everything about all the village people, even about those who did not confess, because they appeared in the confessions of others. Also, everyone soon knew about the stranger who had entered the village, because some occasionally peered out of the window, and what they saw got around. Such control mechanisms were temporarily lost through anonymization in rapidly growing cities but are now returning with the help of new surveillance devices such as smartphones as part of digital nervous systems that tell companies and governments a lot about billions of users. Cameras and drones etc. are becoming increasingly tinier and more ubiquitous. More effective recognition of faces and other detection technology are becoming cheaper and cheaper, and many will use it to identify others anywhere on earth; the big wide world will not offer any more privacy than the local village. Is this good or bad? Some nations may find it easier than others to justify more complex kinds of super-organisms at the expense of the privacy rights of their constituents. Jones: So, there is no way to stop or change this process of collection, or how it continuously informs decisions over time? How do you see governance and rules responding to this, especially amid Italy’s ban on ChatGPT following suspected user data breach and the more recent news about the Meta’s record $1.3billion fine in the company’s handling of user information? Schmidhuber: Data collection has benefits and drawbacks, such as the loss of privacy. How to balance those? I have argued for addressing this through data ownership in data markets. If it is true that data is the new oil, then it should have a price, just like oil. At the moment, the major surveillance platforms such as Meta do not offer users any money for their data and the transitive loss of privacy. In the future, however, we will likely see attempts at creating efficient data markets to figure out the data's true financial value through the interplay between supply and demand. Even some of the sensitive medical data should not be priced by governmental regulators but by patients (and healthy persons) who own it and who may sell or license parts thereof as micro-entrepreneurs in a healthcare data market. Following a previous interview, I gave for one of the largest re-insurance companies , let's look at the different participants in such a data market: patients, hospitals, data companies. (1) Patients with a rare form of cancer can offer more valuable data than patients with a very common form of cancer. (2) Hospitals and their machines are needed to extract the data, e.g., through magnet spin tomography, radiology, evaluations through human doctors, and so on. (3) Companies such as Siemens, Google or IBM would like to buy annotated data to make better artificial neural networks that learn to predict pathologies and diseases and the consequences of therapies. Now the market’s invisible hand will decide about the data’s price through the interplay between demand and supply. On the demand side, you will have several companies offering something for the data, maybe through an app on the smartphone (a bit like a stock market app). On the supply side, each patient in this market should be able to profit from high prices for rare valuable types of data. Likewise, competing data extractors such as hospitals will profit from gaining recognition and trust for extracting data well at a reasonable price. The market will make the whole system efficient through incentives for all who are doing a good job. Soon there will be a flourishing ecosystem of commercial data market advisors and what not, just like the ecosystem surrounding the traditional stock market. The value of the data won’t be determined by governments or ethics committees, but by those who own the data and decide by themselves which parts thereof they want to license to others under certain conditions. At first glance, a market-based system seems to be detrimental to the interest of certain monopolistic companies, as they would have to pay for the data - some would prefer free data and keep their monopoly. However, since every healthy and sick person in the market would suddenly have an incentive to collect and share their data under self-chosen anonymity conditions, there will soon be many more useful data to evaluate all kinds of treatments. On average, people will live longer and healthier, and many companies and the entire healthcare system will benefit. Jones: Finally, what is your view on open source versus the private companies like Google and OpenAI? Is there a danger to supporting these private companies’ large language models versus trying to keep these models open source and transparent, very much like what LAION is doing? Schmidhuber: I signed this open letter by LAION because I strongly favor the open-source movement. And I think it's also something that is going to challenge whatever big tech dominance there might be at the moment. Sure, the best models today are run by big companies with huge budgets for computers, but the exciting fact is that open-source models are not so far behind, some people say maybe six to eight months only. Of course, the private company models are all based on stuff that was created in academia, often in little labs without so much funding, which publish without patenting their results and open source their code and others take it and improved it. Big tech has profited tremendously from academia; their main achievement being that they have scaled up everything greatly, sometimes even failing to credit the original inventors. So, it's very interesting to see that as soon as some big company comes up with a new scaled-up model, lots of students out there are competing, or collaborating, with each other, trying to come up with equal or better performance on smaller networks and smaller machines. And since they are open sourcing, the next guy can have another great idea to improve it, so now there’s tremendous competition also for the big companies. Because of that, and since AI is still getting exponentially cheaper all the time, I don't believe that big tech companies will dominate in the long run. They find it very hard to compete with the enormous open-source movement. As long as you can encourage the open-source community, I think you shouldn't worry too much. Now, of course, you might say if everything is open source, then the bad actors also will more easily have access to these AI tools. And there's truth to that. But as always since the invention of controlled fire, it was good that knowledge about how technology works quickly became public such that everybody could use it. And then, against any bad actor, there's almost immediately a counter actor trying to nullify his efforts. You see, I still believe in our old motto "AI∀" or "AI For All." Jones: Thank you, Juergen for sharing your perspective on this amazing time in history. It’s clear that with new technology, the enormous potential can be matched by disparate and troubling risks which we’ve yet to solve, and even those we have yet to identify. If we are to dispel the fear of a sentient system for which we have no control, humans, alone need to take steps for more responsible development and collaboration to ensure AI technology is used to ultimately benefit society. Humanity will be judged by what we do next.

[D] Why is the AI Hype Absolutely Bonkers
reddit
LLM Vibe Score0
Human Vibe Score1
good_riceThis week

[D] Why is the AI Hype Absolutely Bonkers

Edit 2: Both the repo and the post were deleted. Redacting identifying information as the author has appeared to make rectifications, and it’d be pretty damaging if this is what came up when googling their name / GitHub (hopefully they’ve learned a career lesson and can move on). TL;DR: A PhD candidate claimed to have achieved 97% accuracy for coronavirus from chest x-rays. Their post gathered thousands of reactions, and the candidate was quick to recruit branding, marketing, frontend, and backend developers for the project. Heaps of praise all around. He listed himself as a Director of XXXX (redacted), the new name for his project. The accuracy was based on a training dataset of ~30 images of lesion / healthy lungs, sharing of data between test / train / validation, and code to train ResNet50 from a PyTorch tutorial. Nonetheless, thousands of reactions and praise from the “AI | Data Science | Entrepreneur” community. Original Post: I saw this post circulating on LinkedIn: https://www.linkedin.com/posts/activity-6645711949554425856-9Dhm Here, a PhD candidate claims to achieve great performance with “ARTIFICIAL INTELLIGENCE” to predict coronavirus, asks for more help, and garners tens of thousands of views. The repo housing this ARTIFICIAL INTELLIGENCE solution already has a backend, front end, branding, a README translated in 6 languages, and a call to spread the word for this wonderful technology. Surely, I thought, this researcher has some great and novel tech for all of this hype? I mean dear god, we have branding, and the author has listed himself as the founder of an organization based on this project. Anything with this much attention, with dozens of “AI | Data Scientist | Entrepreneur” members of LinkedIn praising it, must have some great merit, right? Lo and behold, we have ResNet50, from torchvision.models import resnet50, with its linear layer replaced. We have a training dataset of 30 images. This should’ve taken at MAX 3 hours to put together - 1 hour for following a tutorial, and 2 for obfuscating the training with unnecessary code. I genuinely don’t know what to think other than this is bonkers. I hope I’m wrong, and there’s some secret model this author is hiding? If so, I’ll delete this post, but I looked through the repo and (REPO link redacted) that’s all I could find. I’m at a loss for thoughts. Can someone explain why this stuff trends on LinkedIn, gets thousands of views and reactions, and gets loads of praise from “expert data scientists”? It’s almost offensive to people who are like ... actually working to treat coronavirus and develop real solutions. It also seriously turns me off from pursuing an MS in CV as opposed to CS. Edit: It turns out there were duplicate images between test / val / training, as if ResNet50 on 30 images wasn’t enough already. He’s also posted an update signed as “Director of XXXX (redacted)”. This seems like a straight up sleazy way to capitalize on the pandemic by advertising himself to be the head of a made up organization, pulling resources away from real biomedical researchers.

[N] TheSequence Scope: When it comes to machine learning, size matters: Microsoft's DeepSpeed framework, which can train a model with up to a trillion parameters
reddit
LLM Vibe Score0
Human Vibe Score1
KseniaseThis week

[N] TheSequence Scope: When it comes to machine learning, size matters: Microsoft's DeepSpeed framework, which can train a model with up to a trillion parameters

Hi there! Offering to your attention the latest edition of a weekly ML-newsletter that focusing on three things: impactful ML research papers, cool ML tech solutions, and ML use cases supported by investors. Please, see it below. Reddit is a new thing for me, and I've been struggling a bit with it, so please don't judge me too harsh for this promotion. This weekly digest is free and I hope you'd find the format convenient for you. Your feedback is very appreciated, and please feel free to sign up if you like it. 📝 Editorial  The recent emergence of pre-trained language models and transformer architectures pushed the creation of larger and larger machine learning models. Google’s BERT presented attention mechanism and transformer architecture possibilities as the “next big thing” in ML, and the numbers seem surreal. OpenAI’s GPT-2 set a record by processing 1.5 billion parameters, followed by Microsoft’s Turing-NLG, which processed 17 billion parameters just to see the new GPT-3 processing an astonishing 175 billion parameters. To not feel complacent, just this week Microsoft announced a new release of its DeepSpeed framework (which powers Turing-NLG), which can train a model with up to a trillion parameters. That sounds insane but it really isn’t.   What we are seeing is a consequence of several factors. First, computation power and parallelization techniques have evolved to a point where it is relatively easy to train machine learning models in large clusters of machines. Second and most importantly, in the current state of machine learning, larger models have regularly outperformed smaller and more specialized models. Knowledge reusability methods like transfer learning are still in very nascent stages. As a result, it’s really hard to build small models that can operate in uncertain environments. Furthermore, as models like GPT-3 and Turing-NLG have shown, there is some unexplainable magic that happens after models go past a certain size. Many of the immediate machine learning problems might be solved by scaling the current generation of neural network architectures. Plain and simple, when it comes to machine learning, size matters.   We would love to hear your opinions about the debate between broader-larger vs. smaller and more specialized models.   Leave a comment Now, to the most important developments in the AI industry this week 🔎 ML Research GPT-3 Falls Short in Machine Comprehension Proposed by researchers from a few major American universities, a 57-task test to measure models’ ability to reason poses challenges even for sophisticated models like GPT-3 ->read more in the original paper Better Text Summarization OpenAI published a paper showing a reinforcement learning with human feedback technique that can surpass supervised models ->read more on OpenAI blog Reinforcement Learning with Offline Datasets Researchers from the Berkeley AI Research (BAIR) Lab published a paper unveiling a method that uses offline datasets to improve reinforcement learning models->read more on BAIR blog 🤖 Cool AI Tech Releases New Version of DeepSpeed Microsoft open-sourced a new version of DeepSpeed, an open-source library for parallelizable training that can scale up to models with 1 trillion parameters->read more on Microsoft Research blog 💸 Money in AI AI-powered customer experience management platform Sprinklr has raised $200 million (kudos to our subscribers from Sprinklr 👏). Sprinklr's “AI listening processing” solution allows companies to get structured and meaningful sentiments and insights from unstructured customer data that comes from public conversations on different websites and social platforms. Xometry, an on-demand industrial parts marketplace, raises $75 million in Series E funding. The company provides a digital way of creating the right combination of buyers and manufacturers. Another example of AI implementation into matching two sides for a deal. Real estate tech company Orchard raises $69 million in its recent funding round. Orchard aims to digitize the whole real estate market, by developing a solution that combines machine learning and rapid human assistance to smooth the search, match the right deal, and simplify buying and selling relationships. Cybersecurity startup Pcysys raised $25 million in its funding round. Pcysys’ platform, which doesn’t require installation or network reconfiguration, uses algorithms to scan and “ethically” attack enterprise networks. Robotics farming company Iron Ox raised $20 million in a funding round. The system of farming robots is still semi-autonomous, the company’s goal is to become fully autonomous.  Insurtech company Descartes Underwriting raised $18.5 million. The company applies AI and machine learning technologies to climate risk predicting and insurance underwriting. Legaltech startup ThoughtRiver raised $10 million in its Series A round. Its AI solution applied to contract pre-screening aims to boost operational efficiency. Medtech startup Skin Analytics raised $5.1 million in Series A funding. Skin Analytics has developed a clinically validated AI system that can identify not only the important skin cancers but also precancerous lesions that can be treated, as well as a range of lesions that are benign. Amazon, along with several government organizations and three other industry partners, helped fund the National Science Foundation, a high-priority AI research initiative. The amount of funding is not disclosed. The content of TheSequence is written by Jesus Rodriguez, one of the most-read contributors to KDNuggets and TDS. You can check his Medium here.

[D] Should We Be Concerned About The Failure Of Evolutionary Algorithms, And Its Implications?
reddit
LLM Vibe Score0
Human Vibe Score-1
mystikaldangerThis week

[D] Should We Be Concerned About The Failure Of Evolutionary Algorithms, And Its Implications?

https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6287292/ &#x200B; A number of possible explanations for \[why we can't evolve complex software\] could be considered. We tried to be as comprehensive as possible in this section, but it is possible that we have not considered some plausible explanations: Incompetent programmers—It is theoretically possible, but is highly unlikely, that out of thousands of scientists working on evolutionary computation, all failed to correctly implement the Darwinian algorithm. Nonrepresentative algorithms—Some have suggested that EAs do not accurately capture the theory of evolution, but of course that would imply that the theory itself is not specified in sufficient detail to make falsifiable predictions. If, however, such more detailed specifications are available to GP believers, it is up to them to implement them as computer simulations for testing purposes, but no successful examples of such work are known and the known ones have not been successful in evolving software. Inadequate fitness functions—Fitness function for a complex software product is difficult to outline and specify and may be as complex (or even more complex) as the software we want to evolve as it has to consider all the possible use cases and pass all unit tests. This may be the Achilles heel of GP, but it is also an objection to feasibility of programming in general and GP in particular, as both have to convert software specification into the source code. If human programmers and biological evolution succeed with such constraints, so should Darwinian simulations. The Halting problem—Turing proved that it is impossible to determine whether an arbitrary program halts, but this is also a problem for human programmers and could be easily addressed by placing time limits on considered solutions. Program correctness—If we require evolved software to be provably correct, this would present a problem as GP does not verify produced designs but only tests them against specific unit tests. Likewise, we cannot rely on automated software verification as it is still an unsolved problem in the general case. This is not really a problem as most of the human-written software is never proven to be correct and only a small portion of software engineering process relies of formal specification and Test Driven Development. Inappropriate solutions—Literature on EA is full of examples of surprising creativity of Darwinian algorithm resulting in solutions which match the letter of design specifications but not the spirit. This is similar to human-produced software and numerous examples of ways in which such software fails the goals of the initial design. Insufficient complexity of the environment (not enough data, poor fitness functions)—It is possible that the simulated environment is not complex enough to generate high complexity outputs in evolutionary simulations. This does not seem correct as Internet presents a highly complex landscape in which many self-modifying computer viruses roam. Likewise, virtual world such as Second Life and many others present close approximations to the real world and are certainly more complex than early Earth was: A skeptic might insist that an abstract environment would be inadequate for the evolution . . ., believing instead that the virtual environment would need to closely resemble the actual biological environment in which our ancestors evolved. Creating a physically realistic virtual world would require a far greater investment of computational resources than the simulation of a simple toy world or abstract problem domain (whereas evolution had access to a physically realistic real world “for free”). In the limiting case, if complete microphysical accuracy were insisted upon, the computational requirements would balloon to utterly infeasible proportions. Requiring more realistic environmental conditions may result in an increase in necessary computational resources, a problem addressed in the next bullet. Insufficient resources (compute, memory)—From the history of computer science, we know of many situations (speech recognition, NN training), where we had a correct algorithm but insufficient computational resources to run it to success. It is possible that we simply do not have hardware powerful enough to emulate evolution. We will address this possibility in section “Computational Complexity of Biological Evolution and Available Compute.” Software design is not amenable to evolutionary methods—Space of software designs may be discrete with no continuous path via incremental fitness to the desired solutions. This is possible, but this implies that original goals of GP are unattainable and misguided. In addition, because a clear mapping exists between solutions to problems and animals as solutions to environmental problems, this would also imply that current explanation for the origin of the species is incorrect. Darwinian algorithm is incomplete or wrong—Finally, we have to consider the possibility that the inspiration behind evolutionary computation, the Darwinian algorithm itself is wrong or at least partially incomplete. If that was true, computer simulations of such algorithm would fail to produce results comparable with observations we see in nature and a search for an alternative algorithm would need to take place. This would be an extraordinary claim and would require that we discard all the other possible explanations from this list. We challenge EA community to prove us wrong by producing an experiment, which evolves nontrivial software from scratch and without human help. That would be the only way in which our findings could be shown to be incorrect. Perhaps, reframing the problem in terms of maximizing negentropy of digital organisms, as suggested by Schrödinger, Michaelian, and Ulanowicz and Hannon, with respect to negative energy being a fundamental property of all life-forms may produce better results. On a positive side, the fact that it seems impossible to evolve complex software implies that we are unlikely to be able to evolve highly sophisticated artificially intelligent agents, which may present significant risk to our safety and security. Just imagine what would have happened, if the very first time we ran a simulation of evolution on a computer, it produced a superintelligent agent. Yampolskiy has shown that programming as a problem is AI-complete; if GP can solve programming that would imply that GP = AGI (artificial general intelligence), but we see no experimental evidence for such claim. In fact, it is more likely that once we have AGI, it could be used to create an intelligent fitness function for GP and so evolve software. Genetic programming will not be the cause of AI, but a product of it. However, neuroevolution methods for optimizing deep learning architectures and parameters remain a strong possibility for creation of AGI.

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!
reddit
LLM Vibe Score0
Human Vibe Score-0.5
regalalgorithmThis week

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!

Hi there, we at Skynet Today produce a weekly newsletter summarizing each week's major AI news, which seems like it'd be of interest to this subreddit. Here's what's in our latest one: Facebook’s AI for detecting hate speech is facing its biggest challenge yet Facebook has made significant progress recently to proactively take down content that violate its community standards. For example, in the second quarter of 2020, Facebook took down 104.6 million pieces of content. While reviews are typically performed by a vast workforce of human moderators, AI-powered tools have enabled Facebook to do this work at a greater scale for textual content. However, there’s a long way to go for these systems to match or exceed the capabilities of human moderators. This is because a large proportion of hate speech and misinformation is in the form of images and memes, and reasoning about the context and language-image interplay is an extremely difficult challenge for AI. Given Facebook’s scale and the speed at which some use it to spread hate, incite violence, and share lies with millions, Facebook will have to keep running to catch up. AI Slays Top F-16 Pilot In DARPA Dogfight Simulation The Defense Advanced Research Project Agency (DARPA) recently hosted a simulated F16 dogfight competition, with different AI bots competing with each other as well as with human pilots. The top AI bot was able to beat a human pilot 5-0 in the simulated contest. DARPA started this program “as a risk-reduction effort \[…\] to flesh out how human and machine pilots share operational control of a fighter jet to maximize its chances of mission success.” Competition runners are broadly optimistic about the demonstration of AI capabilities, even if they are not close to being deployed on a real aircraft. Of concern, the program had little discussion on the ethics of AI military applications, especially with the lethal autonomous weapon systems being considered. News Advances & Business Microsoft, Energy Dept. to Develop Disaster-Response AI Tools \- The U.S. Department of Energy and Microsoft Corp. on Tuesday announced a partnership to develop artificial-intelligence tools aimed at helping first-responders better react to fast-changing natural events, such as floods and wildfires. Coronavirus: Robot CERi is a bilingual Covid-19 expert \- Ceri is bilingual, clued-up on coronavirus and can tell what mood you are in. Ceri also happens to be a robot. Moscow DOH uses AI platform to detect lung cancer symptoms \- Moscow’s department of health is using an artificial intelligence (AI) platform to detect symptoms of lung cancer in CT scans, as part of a project to implement AI technology for radiology. Scientists develop artificial intelligence system for high precision recognition of hand gestures \- The recognition of human hand gestures by AI systems has been a valuable development over the last decade and has been adopted in high-precision surgical robots, health monitoring equipment and in gaming systems. Forget credit cards - now you can pay with your face. Creepy or cool? \- A new way to pay has arrived in Los Angeles: your face. Concerns & Hype The dystopian tech that companies are selling to help schools reopen sooner \- This fall, AI could be watching students social distance and checking their masks. Thousands of schools nationwide will not be reopening this fall. NYPD Used Facial Recognition Technology In Siege Of Black Lives Matter Activist’s Apartment \- The NYPD deployed facial recognition technology in its hunt for a prominent Black Lives Matter activist, whose home was besieged by dozens of officers and police dogs last week, a spokesperson confirmed to Gothamist. Machines can spot mental health issues - if you hand over your personal data \- Digital diagnosis could transform psychiatry by mining your most intimate data for clues. But is the privacy cost worth it? Supporting Black Artists Who Are Examining AI \- Technology has a complicated relationship with racial justice. Smartphones, internet platforms, and other digital tools can be used to document and expose racism. But digital tools can also fuel racism: smart doorbells surveil Black individuals. A-level and GCSE results in England to be based on teacher assessments in U-turn \- All A-level and GCSE results in England will be based on grades assesed by teachers instead of algorithms. Analysis & Policy GPT-3 and The Question of Automation \- Automation is not an all or nothing proposition. An AI model’s automation capability is highly conjoined with the task and application it is used in. An A.I. Movie Service Could One Day Serve You a New Custom Film Every Time \- How long will it be until an A.I. can make an actual feature film on demand? Fairness, evidence, and predictive equality \- How the causal fairness principle relates to predictive equality How robotics and automation could create new jobs in the new normal \- Depending on who you ask, AI and automation will either destroy jobs or create new ones. In reality, a greater push toward automation will probably both kill and create jobs - human workers will become redundant in certain spheres, sure, but many new roles will likely crop up. Expert Opinions & Discussion within the field Too many AI researchers think real-world problems are not relevant \- The community’s hyperfocus on novel methods ignores what’s really important.

[N] AI Robotics startup Covariant (founded by Peter Chen, Pieter Abbeel, other Berkeley / ex-OpenAI folks) just raised $40M in Series B funding round. “Covariant has recently seen increased usage from clients hoping to avoid supply chain disruption due to the coronavirus pandemic.”
reddit
LLM Vibe Score0
Human Vibe Score0
baylearnThis week

[N] AI Robotics startup Covariant (founded by Peter Chen, Pieter Abbeel, other Berkeley / ex-OpenAI folks) just raised $40M in Series B funding round. “Covariant has recently seen increased usage from clients hoping to avoid supply chain disruption due to the coronavirus pandemic.”

h/t their announcement, VB and WSJ article: Logistics AI Startup Covariant Reaps $40 Million in Funding Round Company plans to explore uses of machine learning for automation beyond warehouse operations Artificial-intelligence robotics startup Covariant raised $40 million to expand its logistics automation technology to new industries and ramp up hiring, the company said Wednesday. The Berkeley, Calif.-based company makes AI software that it says helps warehouse robots pick objects at a faster rate than human workers, with a roughly 95% accuracy rate. Covariant is working with Austrian logistics-automation company Knapp AG and the robotics business of Swiss industrial conglomerate ABB Ltd., which provide hardware such as robot arms or conveyor belts to pair with the startup’s technology platform. “What we’ve built is a universal brain for robotic manipulation tasks,” Covariant co-founder and Chief Executive Peter Chen said in an interview. “We provide the software, they provide the rest of the systems.” Logistics-sector appetite for such technology is growing as distribution and fulfillment operations that have relied on human labor look to speed output and meet rising digital commerce demand. The coronavirus pandemic has accelerated that interest as businesses have sought to adjust their operations to volatile swings in consumer demand and to new restrictions, such as spacing workers further apart to guard against contagion. That has provided a bright spot for some technology startups even as many big backers scale back venture-capital spending. Last month logistics delivery platform Bringg said it raised $30 million in a Series D funding round, for example, as demand for home delivery of food, household goods and e-commerce staples soared among homebound consumers. Covariant’s Series B round brings the company’s total funding to $67 million. New investor Index Ventures led the round, with participation from existing investor Amplify Partners and new investors including Radical Ventures. Mr. Chen said the funding will be used to explore the technology’s potential application in other markets such as manufacturing, recycling or agriculture “where there are repetitive manual processes.” Covariant also plans to hire more engineering and other staff, he said. Covariant was founded in 2017 and now has about 50 employees. The company’s technology uses camera systems to capture images of objects, and artificial intelligence to analyze objects and how to pick them up. Machine learning helps Covariant-powered robots learn from experience. The startup’s customers include a German electrical supplies wholesaler that uses the technology to control a mechanical arm that picks out orders of circuit boards, switches and other goods.

[N] Last Week in AI News Digest - Automated chemical synthesis, using heartbeats to detect deepfakes, and more!
reddit
LLM Vibe Score0
Human Vibe Score1
regalalgorithmThis week

[N] Last Week in AI News Digest - Automated chemical synthesis, using heartbeats to detect deepfakes, and more!

Hi there, just sharing the latest edition of our AI news digest newsletter! We're just a couple of AI grad students doing this for fun, so hope the self promotion is not too annoying (also, welcome feedback). See it below, and feel free to subscribe. Mini Briefs Robotics, AI, and Cloud Computing Combine to Supercharge Chemical and Drug Synthesis IBM recently demoed a complex system for chemical testing and drug synthesis. The system has an AI component that predicts the results of chemical reactions, and a fully automated robotic experiment setup that runs chemical tests 24/7. Users can access the remote robotics lab online, and IBM can also install the system on-premise. With these tools working together, IBM is hoping to reduce typical drug discovery and verification time by half. AI researchers use heartbeat detection to identify deepfake videos Researchers from multiple groups are tackling the challenge of detecting deepfake videos by analyzing the apparent heartbeat of the people depicted in the video. This is possible, because a person’s blood flow changes their skin color ever so slightly, and this change is often detectable via a process called photoplethysmography (PPG). Because deepfakes are not currently optimizing to generate realisitic heartbeats, temporal or spatial anomalies in PPG signals allow resesarchers to detect deepfakes with a 97% accuracy. Advances & Business This AI Expert From Senegal Is Helping Showcase Africans In STEM \- Adji Bousso Dieng will be Princeton’s School of Engineering’s first Black female faculty. Google’s AI-powered flood alerts now cover all of India and parts of Bangladesh \- India, the world’s second most populated nation, sees more than 20% of the global flood-related fatalities each year as overrun riverbanks sweep tens of thousands of homes with them. Two years ago, Google volunteered to help. Finding magnetic eruptions in space with an AI assistant \- MMS look for explosive reconnection events as it flies through the magnetopause - the boundary region where Earth’s magnetic butts up against the solar wind that flows throughout the solar system. This know-it-all AI learns by reading the entire web nonstop \- Diffbot is building the biggest-ever knowledge graph by applying image recognition and natural-language processing to billions of web pages. Bosch and Ford will test autonomous parking in Detroit \- Ford, Bosch, and Dan Gilbert’s real estate firm Bedrock today detailed an autonomous parking pilot scheduled to launch in September at The Assembly, a mixed-used building in Detroit’s Corktown neighborhood. Create your own moody quarantine music with Google’s AI \- Lo-Fi Player, the latest project out of Google Magenta, lets you mix tunes with the help of machine learning by interacting with a virtual room. Apple launches AI/ML residency program to attract niche experts \- As Apple’s artificial language and machine learning initiatives continue to expand, its interest in attracting talent has grown - a theme that’s barely under the surface of the company’s occasionally updated Machine Learning Research blog. Dusty Robotics CEO Tessa Lau Discusses Robotics Start-Ups and Autonomous Robots for Construction \- Tessa Lau is Founder/CEO at Dusty Robotics, whose mission is to increase construction industry productivity by introducing robotic automation on the jobsite. Concerns & Hype Google Offers to Help Others With the Tricky Ethics of AI \- Companies pay cloud computing providers like Amazon, Microsoft, and Google big money to avoid operating their own digital infrastructure. The Peace Dividends Of The Autonomous Vehicle Wars \- The rapid growth of the mobile market in the late 2000s and early 2010s led to a burst of technological progress. Ethics must be part of the development process’ \- The increasing use of AI (artificial intelligence) in the development of new medical technologies demands greater attention to ethical aspects. Analysis & Policy China’s new AI trade rules could hamper a TikTok sale \- TikTok’s attempt to sell itself and avert a possible US ban may run into some complications. The Wall Street Journal reports that China has unveiled new restrictions on AI technology exports that could affect TikTok. Podcast Check out our weekly podcast covering these stories! Website | RSS | iTunes | Spotify | YouTube

🌟 Introducing DarwinAI: An Open-Source Platform for the Evolution of Intelligent Agents 🚀 [Project]
reddit
LLM Vibe Score0
Human Vibe Score1
Interesting-Fox-6758This week

🌟 Introducing DarwinAI: An Open-Source Platform for the Evolution of Intelligent Agents 🚀 [Project]

🌱 The Vision: Evolutionary AI at Your Fingertips Imagine a world where AI agents aren't just programmed to perform tasks but evolve over time, adapting and improving through generations, much like living organisms. Welcome to DarwinAI, an open-source platform inspired by biological evolution, designed to breed, train, and evolve AI agents that can tackle complex, dynamic, and unpredictable challenges. 🧬 The Genetic Blueprint: Building Blocks of Intelligence At the core of DarwinAI is the concept of a digital DNA for each AI agent. This DNA is a modular structure that defines the agent's capabilities, behaviors, and adaptability. Here's what makes up this digital DNA: Genes of Ability: These are snippets of code that represent specific functions, like data classification, text analysis, or optimization. Think of them as the skills your AI agent possesses. Genes of Adaptation: These genes control how the agent responds to different environments or contexts. They determine its flexibility and resilience in the face of changing conditions. Genes of Connection: These define how the agent interacts with other agents or external resources. They are the social and collaborative aspects of the agent. This digital DNA is stored in a structured, version-controlled database, allowing us to track the evolution of each agent and ensure that beneficial mutations are preserved over time. 🛠️ The Evolutionary Process: From Genesis to Mastery The evolution of AI agents in DarwinAI happens through a series of generations, each building upon the strengths of the previous one: Selection of Parents: The fittest agents, those that excel at specific tasks, are chosen as parents. These agents have proven their worth in the simulated environment and are prime candidates for breeding the next generation. Genetic Crossover: The digital DNA of these parent agents is combined to create new agents. This can happen in two ways: Direct Crossover: Where entire genes are copied from the parents. Combinatorial Crossover: Where parts of different genes are fused to create entirely new abilities. Mutations: Random, small changes are introduced into the genes to promote diversity and explore new solutions. These mutations are the wildcards that can lead to breakthrough abilities. 🌍 The Simulated Environment: A Playground for Evolution Agents don't just exist in a vacuum; they operate in a dynamic, simulated environment where they must adapt and survive. This environment is designed to challenge the agents with: Evolutionary Tasks: Problems that agents must solve, such as data classification, prediction, or content generation. Changing Contexts: Factors like noisy data, resource constraints, or new rules that force agents to adapt on the fly. 🐣 The Life Cycle of an Agent: From Birth to Legacy Each agent goes through a life cycle that mirrors the process of natural selection: Initial Learning: Agents receive initial training based on their digital DNA. Task Execution: They perform tasks in the simulated environment, where their abilities are put to the test. Performance Evaluation: Their effectiveness, adaptability, and efficiency are measured. Reproduction: The top-performing agents produce offspring with improved genetic traits. Discard and Archive: Less effective agents are archived for future analysis, ensuring that their lessons are not lost. 🧩 Knowledge Transfer: Passing the Torch One of the key aspects of DarwinAI is the ability for agents to pass on their learned knowledge to future generations: Weight Persistence: Trained models retain their learned weights, allowing them to inherit capabilities from their ancestors. Modular Transfer: Optimized ability genes can be directly copied to new generations, ensuring that valuable skills are preserved. 🛠️ Modularity and Extensibility: Build, Mix, and Evolve DarwinAI is designed to be highly modular and extensible, allowing for: New Capabilities: Easily incorporate new genes to expand the agents' abilities over time. Hybridization: Combine agents from different specializations to create more complex and versatile agents. Directed Evolution: Introduce controlled mutations to address specific problems or challenges. 🚀 Innovative Use Cases: The Future is Bright The potential applications of DarwinAI are vast and varied: Adaptive Automation: Create agents that can adapt to new market conditions or evolving industrial requirements. Collaborative Robots: Develop robots that evolve to improve teamwork in dynamic environments. Scientific Discovery: Agents that combine skills to uncover patterns or solutions that were previously unknown. 🚀 Vision for the Future: An Ecosystem of Evolving Intelligence By fostering an ecosystem where knowledge is accumulated and adaptability is paramount, DarwinAI aims to produce agents that are not only intelligent but also diverse and efficient. These agents will be equipped to handle complex, unpredictable challenges, opening up new frontiers in AI research and application. 🌐 Join Us in Shaping the Future of AI! DarwinAI is more than just a project; it's a community-driven movement towards a new era of AI. We invite you to join us, contribute your ideas, and help shape the future of evolutionary AI. Whether you're a developer, researcher, or simply someone excited about the potential of AI, there's a place for you in this journey. Let's evolve together! 🌱💻

[D] Last Week in Medical AI: Top Research Papers/Models 🏅(September 21 - September 27, 2024)
reddit
LLM Vibe Score0
Human Vibe Score1
aadityauraThis week

[D] Last Week in Medical AI: Top Research Papers/Models 🏅(September 21 - September 27, 2024)

Last Week in Medical AI: Top Research Papers\/Models 🏅\(September 21 - September 27, 2024\) Medical AI Paper of the Week A Preliminary Study of o1 in Medicine: Are We Closer to an AI Doctor? This paper presents o1, a Large Language Model (LLM) evaluated across 37 medical datasets demonstrating superior performance in clinical understanding, reasoning, and multilinguality compared to GPT-4 and GPT-3.5. Medical LLM & Other Models: DREAMS: Python Framework for Medical LLMs A comprehensive deep learning framework for EEG data processing, model training, and report generation. SLaVA-CXR: A Small Language and Vision Assistant for Chest X-Ray Report Automation This paper introduces SLaVA-CXR, an innovative small-scale model designed for automating chest X-ray reports with high accuracy and efficiency. O1 in Medicine: AI Doctor Potential Genome Language Model : Opportunities & Challenge It highlights key gLM applications like functional constraint prediction, sequence design, and transfer learning, while discussing challenges in developing effective gLMs for complex genomes. Medical LLMs & Benchmarks: MEDICONFUSION: Probing Medical LLM Reliability This paper introduces MediConfusion, a challenging benchmark for probing the failure modes of multimodal large language models (MLLMs) in medical imaging. CHBench: Chinese LLM Health Evaluation This paper introduces CHBench, the first comprehensive Chinese health-related benchmark designed to evaluate large language models (LLMs) on their understanding of physical and mental health. LLMs for Mental Illness Evaluation PALLM: Evaluating Palliative Care LLMs Protein LMs: Scaling Necessity? Frameworks and Methodologies: Digital Twin for Oncology Operations Enhancing Guardrails for Healthcare AI InterMind: LLM-Powered Depression Assessment Conversational Health Agents: LLM Framework Medical LLM Applications: LLMs for Mental Health Severity Prediction Fine-tuning LLMs for Radiology Reports LLMs in Patient Education: Back Pain Boosting Healthcare LLMs with Retrieved Context Continuous Pretraining for Clinical LLMs AI in Healthcare Ethics: Confidence Intervals in Medical Imaging AI Generative AI Readiness for Clinical Use ... Check the full thread in detail: https://x.com/OpenlifesciAI/status/1840020394880667937 Thank you for reading! If you know of any interesting papers that were missed, feel free to share them in the comments. If you have insights or breakthroughs in Medical AI you'd like to share in next week's edition, connect with us on Twt/x: OpenlifesciAI

[D] Last Week in Medical AI: Top LLM Research Papers/Models (December 7 - December 14, 2024)
reddit
LLM Vibe Score0
Human Vibe Score0
aadityauraThis week

[D] Last Week in Medical AI: Top LLM Research Papers/Models (December 7 - December 14, 2024)

[\[D\] Last Week in Medical AI: Top LLM Research Papers\/Models \(December 7 - December 14, 2024\)](https://preview.redd.it/o23fp3csj07e1.jpg?width=1280&format=pjpg&auto=webp&s=69e19fc351b3aa5e34c4c00e66245583f88bd9bb) Medical LLM & Other Models PediaBench: Chinese Pediatric LLM This paper introduces PediaBench, the first Chinese pediatric dataset for evaluating Large Language Model (LLM) question-answering performance, containing 4,565 objective and 1,632 subjective questions across 12 disease groups. BiMediX: Bilingual Medical LLM This paper introduces BiMediX, the first bilingual (English-Arabic) medical Mixture of Experts LLM, along with BiMed1.3M, a 1.3M bilingual medical instruction dataset with over 632M tokens used for training. Diverse medical knowledge integration This paper introduces BiMediX2, a bilingual (Arabic-English) Large Multimodal Model (LMM) based on Llama3.1 architecture, trained on 1.6M medical interaction samples. BRAD: Digital Biology Language Model This paper introduces BRAD (Bioinformatics Retrieval Augmented Digital assistant), an LLM-powered chatbot and agent system integrating various bioinformatics tools. MMedPO: Vision-Language Medical LLM This paper introduces MMedPO, a multimodal medical preference optimization approach to improve factual accuracy in Medical Large Vision-Language Models (Med-LVLMs) by addressing modality misalignment. Frameworks & Methodologies \- TOP-Training: Medical Q&A Framework \- Hybrid RAG: Secure Medical Data Management \- Zero-Shot ATC Clinical Coding \- Chest X-Ray Diagnosis Architecture \- Medical Imaging AI Democratization Benchmarks & Evaluations \- KorMedMCQA: Korean Healthcare Licensing Benchmark \- Large Language Model Medical Tasks \- Clinical T5 Model Performance Study \- Radiology Report Quality Assessment \- Genomic Analysis Benchmarking LLM Applications \- TCM-FTP: Herbal Prescription Prediction \- LLaSA: Activity Analysis via Sensors \- Emergency Department Visit Predictions \- Neurodegenerative Disease AI Diagnosis \- Kidney Disease Explainable AI Model Ethical AI & Privacy \- Privacy-Preserving LLM Mechanisms \- AI-Driven Digital Organism Modeling \- Biomedical Research Automation \- Multimodality in Medical Practice Full thread in detail: https://x.com/OpenlifesciAI/status/1867999825721242101

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!
reddit
LLM Vibe Score0
Human Vibe Score-0.5
regalalgorithmThis week

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!

Hi there, we at Skynet Today produce a weekly newsletter summarizing each week's major AI news, which seems like it'd be of interest to this subreddit. Here's what's in our latest one: Facebook’s AI for detecting hate speech is facing its biggest challenge yet Facebook has made significant progress recently to proactively take down content that violate its community standards. For example, in the second quarter of 2020, Facebook took down 104.6 million pieces of content. While reviews are typically performed by a vast workforce of human moderators, AI-powered tools have enabled Facebook to do this work at a greater scale for textual content. However, there’s a long way to go for these systems to match or exceed the capabilities of human moderators. This is because a large proportion of hate speech and misinformation is in the form of images and memes, and reasoning about the context and language-image interplay is an extremely difficult challenge for AI. Given Facebook’s scale and the speed at which some use it to spread hate, incite violence, and share lies with millions, Facebook will have to keep running to catch up. AI Slays Top F-16 Pilot In DARPA Dogfight Simulation The Defense Advanced Research Project Agency (DARPA) recently hosted a simulated F16 dogfight competition, with different AI bots competing with each other as well as with human pilots. The top AI bot was able to beat a human pilot 5-0 in the simulated contest. DARPA started this program “as a risk-reduction effort \[…\] to flesh out how human and machine pilots share operational control of a fighter jet to maximize its chances of mission success.” Competition runners are broadly optimistic about the demonstration of AI capabilities, even if they are not close to being deployed on a real aircraft. Of concern, the program had little discussion on the ethics of AI military applications, especially with the lethal autonomous weapon systems being considered. News Advances & Business Microsoft, Energy Dept. to Develop Disaster-Response AI Tools \- The U.S. Department of Energy and Microsoft Corp. on Tuesday announced a partnership to develop artificial-intelligence tools aimed at helping first-responders better react to fast-changing natural events, such as floods and wildfires. Coronavirus: Robot CERi is a bilingual Covid-19 expert \- Ceri is bilingual, clued-up on coronavirus and can tell what mood you are in. Ceri also happens to be a robot. Moscow DOH uses AI platform to detect lung cancer symptoms \- Moscow’s department of health is using an artificial intelligence (AI) platform to detect symptoms of lung cancer in CT scans, as part of a project to implement AI technology for radiology. Scientists develop artificial intelligence system for high precision recognition of hand gestures \- The recognition of human hand gestures by AI systems has been a valuable development over the last decade and has been adopted in high-precision surgical robots, health monitoring equipment and in gaming systems. Forget credit cards - now you can pay with your face. Creepy or cool? \- A new way to pay has arrived in Los Angeles: your face. Concerns & Hype The dystopian tech that companies are selling to help schools reopen sooner \- This fall, AI could be watching students social distance and checking their masks. Thousands of schools nationwide will not be reopening this fall. NYPD Used Facial Recognition Technology In Siege Of Black Lives Matter Activist’s Apartment \- The NYPD deployed facial recognition technology in its hunt for a prominent Black Lives Matter activist, whose home was besieged by dozens of officers and police dogs last week, a spokesperson confirmed to Gothamist. Machines can spot mental health issues - if you hand over your personal data \- Digital diagnosis could transform psychiatry by mining your most intimate data for clues. But is the privacy cost worth it? Supporting Black Artists Who Are Examining AI \- Technology has a complicated relationship with racial justice. Smartphones, internet platforms, and other digital tools can be used to document and expose racism. But digital tools can also fuel racism: smart doorbells surveil Black individuals. A-level and GCSE results in England to be based on teacher assessments in U-turn \- All A-level and GCSE results in England will be based on grades assesed by teachers instead of algorithms. Analysis & Policy GPT-3 and The Question of Automation \- Automation is not an all or nothing proposition. An AI model’s automation capability is highly conjoined with the task and application it is used in. An A.I. Movie Service Could One Day Serve You a New Custom Film Every Time \- How long will it be until an A.I. can make an actual feature film on demand? Fairness, evidence, and predictive equality \- How the causal fairness principle relates to predictive equality How robotics and automation could create new jobs in the new normal \- Depending on who you ask, AI and automation will either destroy jobs or create new ones. In reality, a greater push toward automation will probably both kill and create jobs - human workers will become redundant in certain spheres, sure, but many new roles will likely crop up. Expert Opinions & Discussion within the field Too many AI researchers think real-world problems are not relevant \- The community’s hyperfocus on novel methods ignores what’s really important.

[N] New Trends to Power Faster Artificial Intelligence and Machine Learning Adoption?
reddit
LLM Vibe Score0
Human Vibe Score1
EsotericaCapitalThis week

[N] New Trends to Power Faster Artificial Intelligence and Machine Learning Adoption?

In 2012, Google X lab created a neural network that can identify cats. Since then, technology companies have been increasingly adopting AI/ML on a large scale to build better applications and services for consumers (ToC). On the other hand, AI/ML's adoption on the enterprises' side (ToB) has yet to see the same growth trajectory due to the costs and complexities in both hardware and software. However, Since 2020, we started noticing three emerging tech trends that can help accelerate enterprises' adoption of AI/ML. Breakthrough in semiconductors: In 2020, Nvidia debut the concept of "Data Processing Unit," a new class of programmable processors that combine high-performance CPU with SmartNiC (network interface controller). Data centers can deploy DPUs to optimize computing offload and frees up CPUs to focus on intended tasks, such as machine learning. DPUs help resolve a significant bottleneck for ML training, where models, sometimes with billions of parameters, are way too big for traditional CPUs and GPUs to handle. Other leading semiconductor players, such as Marvell and Xilinx, follow suit with their in-house or partner-designed DPUs. Industry analysts have forecasted that the market size for DPUs in data centers alone could reach $50 billion by 2025. &#x200B; https://preview.redd.it/l436muluhnn61.png?width=1430&format=png&auto=webp&s=ba8d1298056ea31bddd25f1596ff64b7e107580a Breakthrough in software: we also saw significant progress of "Conversational AI," a new form of AI that can understand and speak languages with human-like accuracy, in 2020. Conversational AI allows two-way interactions and provides a much better user experience than traditional AI-powered Chatbot, mostly a one-way response system. The secret of conversational AI is its ability to handle lots of human conversation variance. Developers have designed innovative algorithms such as "Switch transformers" and "Sparse training" to enable models to handle vast amounts of data. The size of conversational AI training models is enormous and keeps expanding. For example, in February 2021, Google Brain announced a model with 1.6 trillion parameters, nine times the size of the famous Open AI GPT-3 (175 billion parameters) unveiled in July 2020. GPT-3 is 100+ times bigger than GPT-2 introduced in 2019. &#x200B; https://preview.redd.it/oajpi2yvhnn61.png?width=1430&format=png&auto=webp&s=1482913a98e17ddc1d62cc79864598d4012ad6f7 Cloud giants are expanding machine-learning platforms for developers. Andy Jassy famously said that "AI is shifting from a niche experiment inside technical departments to becoming more mainstream in business processes." in the 2020 AWS reInvent. During the conference, AWS rolled out many AI products across the technology stack, including AI chips (AWS Trainium), database (Aurora Machine Learning), and vertical solutions (Amazon Healthlake), etc. However, the most significant development is the drastic expansion of "Amazon SageMaker," one of the largest cloud machine-learning platforms. SageMaker has been offering new features to make it easier for developers to automate machine learning workflow. Microsoft Azure and Google Cloud are also growing their ML developer platforms. &#x200B; https://preview.redd.it/z9wf2o8xhnn61.png?width=1430&format=png&auto=webp&s=9f607acfe8f0dbf36fb9b472f3cb40b80f13879e Witnessing these breakthroughs in semiconductor and software, coupled with cloud giants' effort to democratize AI, we see a coming inflection point of accelerated AI adoption in both ToC and ToB markets. So how do we benefit from this megatrend? In semiconductors, we like companies with DPUs exposure. In AI development and processing, we favor multi-cloud AI platforms such as Databricks. In enterprise software, we believe there will be a strong wave of new AI-based enterprise applications that can be creative and efficient in solving real-world problems.

[D] Using AI to navigate the complexities of regulatory frameworks
reddit
LLM Vibe Score0
Human Vibe Score1
cryptobooty_This week

[D] Using AI to navigate the complexities of regulatory frameworks

I would be interested in hearing opinions for using AI for regulatory assurance and compliance in regulated industries, what are your thoughts? Explanation: An AI-driven compliance system ensuring adherence to evolving regulations, minimizing risks, and enabling businesses to operate confidently within legal boundaries. Pairing Large Language Models (LLMs) with blockchain technology to offer a range of benefits, particularly in the context of regulatory compliance. LLMs, powered by advanced natural language processing and machine learning capabilities, can enhance regulatory compliance processes in several ways. Firstly, they can automate the analysis of regulatory documents, helping businesses stay updated with evolving compliance requirements. LLMs can also assist in generating compliance reports, simplifying complex legal language into understandable summaries. Furthermore, by integrating LLMs into smart contracts, businesses can ensure that contract terms adhere to regulatory guidelines automatically. The integration of LLMs with blockchain can significantly improve regulatory compliance by automating document analysis, simplifying legal language, monitoring compliance in real-time, and enhancing customer interactions—all contributing to greater efficiency and accuracy in adhering to regulatory standards. I have a whole technical whitepaper with this stuff on hand, if anyone would like to review it let me know..

[N] AI Robotics startup Covariant (founded by Peter Chen, Pieter Abbeel, other Berkeley / ex-OpenAI folks) just raised $40M in Series B funding round. “Covariant has recently seen increased usage from clients hoping to avoid supply chain disruption due to the coronavirus pandemic.”
reddit
LLM Vibe Score0
Human Vibe Score0
baylearnThis week

[N] AI Robotics startup Covariant (founded by Peter Chen, Pieter Abbeel, other Berkeley / ex-OpenAI folks) just raised $40M in Series B funding round. “Covariant has recently seen increased usage from clients hoping to avoid supply chain disruption due to the coronavirus pandemic.”

h/t their announcement, VB and WSJ article: Logistics AI Startup Covariant Reaps $40 Million in Funding Round Company plans to explore uses of machine learning for automation beyond warehouse operations Artificial-intelligence robotics startup Covariant raised $40 million to expand its logistics automation technology to new industries and ramp up hiring, the company said Wednesday. The Berkeley, Calif.-based company makes AI software that it says helps warehouse robots pick objects at a faster rate than human workers, with a roughly 95% accuracy rate. Covariant is working with Austrian logistics-automation company Knapp AG and the robotics business of Swiss industrial conglomerate ABB Ltd., which provide hardware such as robot arms or conveyor belts to pair with the startup’s technology platform. “What we’ve built is a universal brain for robotic manipulation tasks,” Covariant co-founder and Chief Executive Peter Chen said in an interview. “We provide the software, they provide the rest of the systems.” Logistics-sector appetite for such technology is growing as distribution and fulfillment operations that have relied on human labor look to speed output and meet rising digital commerce demand. The coronavirus pandemic has accelerated that interest as businesses have sought to adjust their operations to volatile swings in consumer demand and to new restrictions, such as spacing workers further apart to guard against contagion. That has provided a bright spot for some technology startups even as many big backers scale back venture-capital spending. Last month logistics delivery platform Bringg said it raised $30 million in a Series D funding round, for example, as demand for home delivery of food, household goods and e-commerce staples soared among homebound consumers. Covariant’s Series B round brings the company’s total funding to $67 million. New investor Index Ventures led the round, with participation from existing investor Amplify Partners and new investors including Radical Ventures. Mr. Chen said the funding will be used to explore the technology’s potential application in other markets such as manufacturing, recycling or agriculture “where there are repetitive manual processes.” Covariant also plans to hire more engineering and other staff, he said. Covariant was founded in 2017 and now has about 50 employees. The company’s technology uses camera systems to capture images of objects, and artificial intelligence to analyze objects and how to pick them up. Machine learning helps Covariant-powered robots learn from experience. The startup’s customers include a German electrical supplies wholesaler that uses the technology to control a mechanical arm that picks out orders of circuit boards, switches and other goods.

[N] Last Week in AI News Digest - Automated chemical synthesis, using heartbeats to detect deepfakes, and more!
reddit
LLM Vibe Score0
Human Vibe Score1
regalalgorithmThis week

[N] Last Week in AI News Digest - Automated chemical synthesis, using heartbeats to detect deepfakes, and more!

Hi there, just sharing the latest edition of our AI news digest newsletter! We're just a couple of AI grad students doing this for fun, so hope the self promotion is not too annoying (also, welcome feedback). See it below, and feel free to subscribe. Mini Briefs Robotics, AI, and Cloud Computing Combine to Supercharge Chemical and Drug Synthesis IBM recently demoed a complex system for chemical testing and drug synthesis. The system has an AI component that predicts the results of chemical reactions, and a fully automated robotic experiment setup that runs chemical tests 24/7. Users can access the remote robotics lab online, and IBM can also install the system on-premise. With these tools working together, IBM is hoping to reduce typical drug discovery and verification time by half. AI researchers use heartbeat detection to identify deepfake videos Researchers from multiple groups are tackling the challenge of detecting deepfake videos by analyzing the apparent heartbeat of the people depicted in the video. This is possible, because a person’s blood flow changes their skin color ever so slightly, and this change is often detectable via a process called photoplethysmography (PPG). Because deepfakes are not currently optimizing to generate realisitic heartbeats, temporal or spatial anomalies in PPG signals allow resesarchers to detect deepfakes with a 97% accuracy. Advances & Business This AI Expert From Senegal Is Helping Showcase Africans In STEM \- Adji Bousso Dieng will be Princeton’s School of Engineering’s first Black female faculty. Google’s AI-powered flood alerts now cover all of India and parts of Bangladesh \- India, the world’s second most populated nation, sees more than 20% of the global flood-related fatalities each year as overrun riverbanks sweep tens of thousands of homes with them. Two years ago, Google volunteered to help. Finding magnetic eruptions in space with an AI assistant \- MMS look for explosive reconnection events as it flies through the magnetopause - the boundary region where Earth’s magnetic butts up against the solar wind that flows throughout the solar system. This know-it-all AI learns by reading the entire web nonstop \- Diffbot is building the biggest-ever knowledge graph by applying image recognition and natural-language processing to billions of web pages. Bosch and Ford will test autonomous parking in Detroit \- Ford, Bosch, and Dan Gilbert’s real estate firm Bedrock today detailed an autonomous parking pilot scheduled to launch in September at The Assembly, a mixed-used building in Detroit’s Corktown neighborhood. Create your own moody quarantine music with Google’s AI \- Lo-Fi Player, the latest project out of Google Magenta, lets you mix tunes with the help of machine learning by interacting with a virtual room. Apple launches AI/ML residency program to attract niche experts \- As Apple’s artificial language and machine learning initiatives continue to expand, its interest in attracting talent has grown - a theme that’s barely under the surface of the company’s occasionally updated Machine Learning Research blog. Dusty Robotics CEO Tessa Lau Discusses Robotics Start-Ups and Autonomous Robots for Construction \- Tessa Lau is Founder/CEO at Dusty Robotics, whose mission is to increase construction industry productivity by introducing robotic automation on the jobsite. Concerns & Hype Google Offers to Help Others With the Tricky Ethics of AI \- Companies pay cloud computing providers like Amazon, Microsoft, and Google big money to avoid operating their own digital infrastructure. The Peace Dividends Of The Autonomous Vehicle Wars \- The rapid growth of the mobile market in the late 2000s and early 2010s led to a burst of technological progress. Ethics must be part of the development process’ \- The increasing use of AI (artificial intelligence) in the development of new medical technologies demands greater attention to ethical aspects. Analysis & Policy China’s new AI trade rules could hamper a TikTok sale \- TikTok’s attempt to sell itself and avert a possible US ban may run into some complications. The Wall Street Journal reports that China has unveiled new restrictions on AI technology exports that could affect TikTok. Podcast Check out our weekly podcast covering these stories! Website | RSS | iTunes | Spotify | YouTube

[P] I built an open SotA image tagging model to do what CLIP won't
reddit
LLM Vibe Score0
Human Vibe Score1
fpgaminerThis week

[P] I built an open SotA image tagging model to do what CLIP won't

I'm a hobbyist ML researcher and finally, after a year of work, built a state of the art machine vision model from scratch. It's ViT-B/16 based, 448x448x3 input, 91M parameters, trained for 660M samples, with multi-label classification as the target task, on over 5000 unique tags. All the big foundation vision models today were trained on heavily filtered datasets, greatly limiting the concepts they can represent, in line with arbitrary sets of rules for what is deemed "wholesome" by leading tech companies. Everything from innocuous to spicy is on the chopping block of those filters. And because CLIP pervades the industry, from StableDiffusion to LLaVA, so does OpenAI's sensibilities. My goal was to build a vision model for tagging images, mainly for labelling images for SD finetunes, but which wasn't as heavily filtered and handicapped as CLIP/BLIP/LLaVA. Something more inclusive, diverse, and sex positive. Starting from the wonderful work of SmilingWolf (https://github.com/SmilingWolf/SW-CV-ModelZoo) and the Danbooru2021 dataset, I iterated for a year on the model, training, and manually labeling a thousand images to help the model generalize beyond the danbooru domain. I'm releasing the first version of this model, dubbed JoyTag, today: https://github.com/fpgaminer/joytag It achieves a mean F1 score of 0.578 across all of its over 5000 tags and across both the anime/manga styled images of the original danbooru dataset, but also photographs and other mediums thanks to the auxiliary training data I provided to it. It was quite the struggle getting to this point, and I probably spent more time and money than any sane person should have. I learned a lot about dealing with datasets as large as danbooru2021, training models at scale, and how to keep yourself awake all night so your 8xA100 rental doesn't crash and blow all your money. In my manual testing outside of even the validation set, the model has generalized well to unseen images, so I'm quite happy with the results thus far. There's plenty more work to do expanding its dataset to improve that F1 score further, and roundout its weak points. With inclusivity and diversity being a major goal of this project, I'm disappointed by some of its remaining limitations (as documented in the GitHub README). But I'm already busy manually tagging more images using my model-augmented workflow. I'm happy to answer questions about the project, the training procedure, anything. All the training parameters are documented on GitHub, but there are so many little details that were hard won over the year. Like that damned loss multiplier. Ugh. Github: https://github.com/fpgaminer/joytag Model download: https://huggingface.co/fancyfeast/joytag/tree/main Demo: https://huggingface.co/spaces/fancyfeast/joytag

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.
reddit
LLM Vibe Score0
Human Vibe Score0.765
hardmaruThis week

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.

Schmidhuber interview expressing his views on the future of AI and AGI. Original source. I think the interview is of interest to r/MachineLearning, and presents an alternate view, compared to other influential leaders in AI. Juergen Schmidhuber, Renowned 'Father Of Modern AI,' Says His Life’s Work Won't Lead To Dystopia May 23, 2023. Contributed by Hessie Jones. Amid the growing concern about the impact of more advanced artificial intelligence (AI) technologies on society, there are many in the technology community who fear the implications of the advancements in Generative AI if they go unchecked. Dr. Juergen Schmidhuber, a renowned scientist, artificial intelligence researcher and widely regarded as one of the pioneers in the field, is more optimistic. He declares that many of those who suddenly warn against the dangers of AI are just seeking publicity, exploiting the media’s obsession with killer robots which has attracted more attention than “good AI” for healthcare etc. The potential to revolutionize various industries and improve our lives is clear, as are the equal dangers if bad actors leverage the technology for personal gain. Are we headed towards a dystopian future, or is there reason to be optimistic? I had a chance to sit down with Dr. Juergen Schmidhuber to understand his perspective on this seemingly fast-moving AI-train that will leap us into the future. As a teenager in the 1970s, Juergen Schmidhuber became fascinated with the idea of creating intelligent machines that could learn and improve on their own, becoming smarter than himself within his lifetime. This would ultimately lead to his groundbreaking work in the field of deep learning. In the 1980s, he studied computer science at the Technical University of Munich (TUM), where he earned his diploma in 1987. His thesis was on the ultimate self-improving machines that, not only, learn through some pre-wired human-designed learning algorithm, but also learn and improve the learning algorithm itself. Decades later, this became a hot topic. He also received his Ph.D. at TUM in 1991 for work that laid some of the foundations of modern AI. Schmidhuber is best known for his contributions to the development of recurrent neural networks (RNNs), the most powerful type of artificial neural network that can process sequential data such as speech and natural language. With his students Sepp Hochreiter, Felix Gers, Alex Graves, Daan Wierstra, and others, he published architectures and training algorithms for the long short-term memory (LSTM), a type of RNN that is widely used in natural language processing, speech recognition, video games, robotics, and other applications. LSTM has become the most cited neural network of the 20th century, and Business Week called it "arguably the most commercial AI achievement." Throughout his career, Schmidhuber has received various awards and accolades for his groundbreaking work. In 2013, he was awarded the Helmholtz Prize, which recognizes significant contributions to the field of machine learning. In 2016, he was awarded the IEEE Neural Network Pioneer Award for "pioneering contributions to deep learning and neural networks." The media have often called him the “father of modern AI,” because the most cited neural networks all build on his lab’s work. He is quick to point out, however, that AI history goes back centuries. Despite his many accomplishments, at the age of 60, he feels mounting time pressure towards building an Artificial General Intelligence within his lifetime and remains committed to pushing the boundaries of AI research and development. He is currently director of the KAUST AI Initiative, scientific director of the Swiss AI Lab IDSIA, and co-founder and chief scientist of AI company NNAISENSE, whose motto is "AI∀" which is a math-inspired way of saying "AI For All." He continues to work on cutting-edge AI technologies and applications to improve human health and extend human lives and make lives easier for everyone. The following interview has been edited for clarity. Jones: Thank you Juergen for joining me. You have signed letters warning about AI weapons. But you didn't sign the recent publication, "Pause Gigantic AI Experiments: An Open Letter"? Is there a reason? Schmidhuber: Thank you Hessie. Glad to speak with you. I have realized that many of those who warn in public against the dangers of AI are just seeking publicity. I don't think the latest letter will have any significant impact because many AI researchers, companies, and governments will ignore it completely. The proposal frequently uses the word "we" and refers to "us," the humans. But as I have pointed out many times in the past, there is no "we" that everyone can identify with. Ask 10 different people, and you will hear 10 different opinions about what is "good." Some of those opinions will be completely incompatible with each other. Don't forget the enormous amount of conflict between the many people. The letter also says, "If such a pause cannot be quickly put in place, governments should intervene and impose a moratorium." The problem is that different governments have ALSO different opinions about what is good for them and for others. Great Power A will say, if we don't do it, Great Power B will, perhaps secretly, and gain an advantage over us. The same is true for Great Powers C and D. Jones: Everyone acknowledges this fear surrounding current generative AI technology. Moreover, the existential threat of this technology has been publicly acknowledged by Sam Altman, CEO of OpenAI himself, calling for AI regulation. From your perspective, is there an existential threat? Schmidhuber: It is true that AI can be weaponized, and I have no doubt that there will be all kinds of AI arms races, but AI does not introduce a new quality of existential threat. The threat coming from AI weapons seems to pale in comparison to the much older threat from nuclear hydrogen bombs that don’t need AI at all. We should be much more afraid of half-century-old tech in the form of H-bomb rockets. The Tsar Bomba of 1961 had almost 15 times more destructive power than all weapons of WW-II combined. Despite the dramatic nuclear disarmament since the 1980s, there are still more than enough nuclear warheads to wipe out human civilization within two hours, without any AI I’m much more worried about that old existential threat than the rather harmless AI weapons. Jones: I realize that while you compare AI to the threat of nuclear bombs, there is a current danger that a current technology can be put in the hands of humans and enable them to “eventually” exact further harms to individuals of group in a very precise way, like targeted drone attacks. You are giving people a toolset that they've never had before, enabling bad actors, as some have pointed out, to be able to do a lot more than previously because they didn't have this technology. Schmidhuber: Now, all that sounds horrible in principle, but our existing laws are sufficient to deal with these new types of weapons enabled by AI. If you kill someone with a gun, you will go to jail. Same if you kill someone with one of these drones. Law enforcement will get better at understanding new threats and new weapons and will respond with better technology to combat these threats. Enabling drones to target persons from a distance in a way that requires some tracking and some intelligence to perform, which has traditionally been performed by skilled humans, to me, it seems is just an improved version of a traditional weapon, like a gun, which is, you know, a little bit smarter than the old guns. But, in principle, all of that is not a new development. For many centuries, we have had the evolution of better weaponry and deadlier poisons and so on, and law enforcement has evolved their policies to react to these threats over time. So, it's not that we suddenly have a new quality of existential threat and it's much more worrisome than what we have had for about six decades. A large nuclear warhead doesn’t need fancy face recognition to kill an individual. No, it simply wipes out an entire city with ten million inhabitants. Jones: The existential threat that’s implied is the extent to which humans have control over this technology. We see some early cases of opportunism which, as you say, tends to get more media attention than positive breakthroughs. But you’re implying that this will all balance out? Schmidhuber: Historically, we have a long tradition of technological breakthroughs that led to advancements in weapons for the purpose of defense but also for protection. From sticks, to rocks, to axes to gunpowder to cannons to rockets… and now to drones… this has had a drastic influence on human history but what has been consistent throughout history is that those who are using technology to achieve their own ends are themselves, facing the same technology because the opposing side is learning to use it against them. And that's what has been repeated in thousands of years of human history and it will continue. I don't see the new AI arms race as something that is remotely as existential a threat as the good old nuclear warheads. You said something important, in that some people prefer to talk about the downsides rather than the benefits of this technology, but that's misleading, because 95% of all AI research and AI development is about making people happier and advancing human life and health. Jones: Let’s touch on some of those beneficial advances in AI research that have been able to radically change present day methods and achieve breakthroughs. Schmidhuber: All right! For example, eleven years ago, our team with my postdoc Dan Ciresan was the first to win a medical imaging competition through deep learning. We analyzed female breast cells with the objective to determine harmless cells vs. those in the pre-cancer stage. Typically, a trained oncologist needs a long time to make these determinations. Our team, who knew nothing about cancer, were able to train an artificial neural network, which was totally dumb in the beginning, on lots of this kind of data. It was able to outperform all the other methods. Today, this is being used not only for breast cancer, but also for radiology and detecting plaque in arteries, and many other things. Some of the neural networks that we have developed in the last 3 decades are now prevalent across thousands of healthcare applications, detecting Diabetes and Covid-19 and what not. This will eventually permeate across all healthcare. The good consequences of this type of AI are much more important than the click-bait new ways of conducting crimes with AI. Jones: Adoption is a product of reinforced outcomes. The massive scale of adoption either leads us to believe that people have been led astray, or conversely, technology is having a positive effect on people’s lives. Schmidhuber: The latter is the likely case. There's intense commercial pressure towards good AI rather than bad AI because companies want to sell you something, and you are going to buy only stuff you think is going to be good for you. So already just through this simple, commercial pressure, you have a tremendous bias towards good AI rather than bad AI. However, doomsday scenarios like in Schwarzenegger movies grab more attention than documentaries on AI that improve people’s lives. Jones: I would argue that people are drawn to good stories – narratives that contain an adversary and struggle, but in the end, have happy endings. And this is consistent with your comment on human nature and how history, despite its tendency for violence and destruction of humanity, somehow tends to correct itself. Let’s take the example of a technology, which you are aware – GANs – General Adversarial Networks, which today has been used in applications for fake news and disinformation. In actuality, the purpose in the invention of GANs was far from what it is used for today. Schmidhuber: Yes, the name GANs was created in 2014 but we had the basic principle already in the early 1990s. More than 30 years ago, I called it artificial curiosity. It's a very simple way of injecting creativity into a little two network system. This creative AI is not just trying to slavishly imitate humans. Rather, it’s inventing its own goals. Let me explain: You have two networks. One network is producing outputs that could be anything, any action. Then the second network is looking at these actions and it’s trying to predict the consequences of these actions. An action could move a robot, then something happens, and the other network is just trying to predict what will happen. Now we can implement artificial curiosity by reducing the prediction error of the second network, which, at the same time, is the reward of the first network. The first network wants to maximize its reward and so it will invent actions that will lead to situations that will surprise the second network, which it has not yet learned to predict well. In the case where the outputs are fake images, the first network will try to generate images that are good enough to fool the second network, which will attempt to predict the reaction of the environment: fake or real image, and it will try to become better at it. The first network will continue to also improve at generating images whose type the second network will not be able to predict. So, they fight each other. The 2nd network will continue to reduce its prediction error, while the 1st network will attempt to maximize it. Through this zero-sum game the first network gets better and better at producing these convincing fake outputs which look almost realistic. So, once you have an interesting set of images by Vincent Van Gogh, you can generate new images that leverage his style, without the original artist having ever produced the artwork himself. Jones: I see how the Van Gogh example can be applied in an education setting and there are countless examples of artists mimicking styles from famous painters but image generation from this instance that can happen within seconds is quite another feat. And you know this is how GANs has been used. What’s more prevalent today is a socialized enablement of generating images or information to intentionally fool people. It also surfaces new harms that deal with the threat to intellectual property and copyright, where laws have yet to account for. And from your perspective this was not the intention when the model was conceived. What was your motivation in your early conception of what is now GANs? Schmidhuber: My old motivation for GANs was actually very important and it was not to create deepfakes or fake news but to enable AIs to be curious and invent their own goals, to make them explore their environment and make them creative. Suppose you have a robot that executes one action, then something happens, then it executes another action, and so on, because it wants to achieve certain goals in the environment. For example, when the battery is low, this will trigger “pain” through hunger sensors, so it wants to go to the charging station, without running into obstacles, which will trigger other pain sensors. It will seek to minimize pain (encoded through numbers). Now the robot has a friend, the second network, which is a world model ––it’s a prediction machine that learns to predict the consequences of the robot’s actions. Once the robot has a good model of the world, it can use it for planning. It can be used as a simulation of the real world. And then it can determine what is a good action sequence. If the robot imagines this sequence of actions, the model will predict a lot of pain, which it wants to avoid. If it plays this alternative action sequence in its mental model of the world, then it will predict a rewarding situation where it’s going to sit on the charging station and its battery is going to load again. So, it'll prefer to execute the latter action sequence. In the beginning, however, the model of the world knows nothing, so how can we motivate the first network to generate experiments that lead to data that helps the world model learn something it didn’t already know? That’s what artificial curiosity is about. The dueling two network systems effectively explore uncharted environments by creating experiments so that over time the curious AI gets a better sense of how the environment works. This can be applied to all kinds of environments, and has medical applications. Jones: Let’s talk about the future. You have said, “Traditional humans won’t play a significant role in spreading intelligence across the universe.” Schmidhuber: Let’s first conceptually separate two types of AIs. The first type of AI are tools directed by humans. They are trained to do specific things like accurately detect diabetes or heart disease and prevent attacks before they happen. In these cases, the goal is coming from the human. More interesting AIs are setting their own goals. They are inventing their own experiments and learning from them. Their horizons expand and eventually they become more and more general problem solvers in the real world. They are not controlled by their parents, but much of what they learn is through self-invented experiments. A robot, for example, is rotating a toy, and as it is doing this, the video coming in through the camera eyes, changes over time and it begins to learn how this video changes and learns how the 3D nature of the toy generates certain videos if you rotate it a certain way, and eventually, how gravity works, and how the physics of the world works. Like a little scientist! And I have predicted for decades that future scaled-up versions of such AI scientists will want to further expand their horizons, and eventually go where most of the physical resources are, to build more and bigger AIs. And of course, almost all of these resources are far away from earth out there in space, which is hostile to humans but friendly to appropriately designed AI-controlled robots and self-replicating robot factories. So here we are not talking any longer about our tiny biosphere; no, we are talking about the much bigger rest of the universe. Within a few tens of billions of years, curious self-improving AIs will colonize the visible cosmos in a way that’s infeasible for humans. Those who don’t won’t have an impact. Sounds like science fiction, but since the 1970s I have been unable to see a plausible alternative to this scenario, except for a global catastrophe such as an all-out nuclear war that stops this development before it takes off. Jones: How long have these AIs, which can set their own goals — how long have they existed? To what extent can they be independent of human interaction? Schmidhuber: Neural networks like that have existed for over 30 years. My first simple adversarial neural network system of this kind is the one from 1990 described above. You don’t need a teacher there; it's just a little agent running around in the world and trying to invent new experiments that surprise its own prediction machine. Once it has figured out certain parts of the world, the agent will become bored and will move on to more exciting experiments. The simple 1990 systems I mentioned have certain limitations, but in the past three decades, we have also built more sophisticated systems that are setting their own goals and such systems I think will be essential for achieving true intelligence. If you are only imitating humans, you will never go beyond them. So, you really must give AIs the freedom to explore previously unexplored regions of the world in a way that no human is really predefining. Jones: Where is this being done today? Schmidhuber: Variants of neural network-based artificial curiosity are used today for agents that learn to play video games in a human-competitive way. We have also started to use them for automatic design of experiments in fields such as materials science. I bet many other fields will be affected by it: chemistry, biology, drug design, you name it. However, at least for now, these artificial scientists, as I like to call them, cannot yet compete with human scientists. I don’t think it’s going to stay this way but, at the moment, it’s still the case. Sure, AI has made a lot of progress. Since 1997, there have been superhuman chess players, and since 2011, through the DanNet of my team, there have been superhuman visual pattern recognizers. But there are other things where humans, at the moment at least, are much better, in particular, science itself. In the lab we have many first examples of self-directed artificial scientists, but they are not yet convincing enough to appear on the radar screen of the public space, which is currently much more fascinated with simpler systems that just imitate humans and write texts based on previously seen human-written documents. Jones: You speak of these numerous instances dating back 30 years of these lab experiments where these self-driven agents are deciding and learning and moving on once they’ve learned. And I assume that that rate of learning becomes even faster over time. What kind of timeframe are we talking about when this eventually is taken outside of the lab and embedded into society? Schmidhuber: This could still take months or even years :-) Anyway, in the not-too-distant future, we will probably see artificial scientists who are good at devising experiments that allow them to discover new, previously unknown physical laws. As always, we are going to profit from the old trend that has held at least since 1941: every decade compute is getting 100 times cheaper. Jones: How does this trend affect modern AI such as ChatGPT? Schmidhuber: Perhaps you know that all the recent famous AI applications such as ChatGPT and similar models are largely based on principles of artificial neural networks invented in the previous millennium. The main reason why they works so well now is the incredible acceleration of compute per dollar. ChatGPT is driven by a neural network called “Transformer” described in 2017 by Google. I am happy about that because a quarter century earlier in 1991 I had a particular Transformer variant which is now called the “Transformer with linearized self-attention”. Back then, not much could be done with it, because the compute cost was a million times higher than today. But today, one can train such models on half the internet and achieve much more interesting results. Jones: And for how long will this acceleration continue? Schmidhuber: There's no reason to believe that in the next 30 years, we won't have another factor of 1 million and that's going to be really significant. In the near future, for the first time we will have many not-so expensive devices that can compute as much as a human brain. The physical limits of computation, however, are much further out so even if the trend of a factor of 100 every decade continues, the physical limits (of 1051 elementary instructions per second and kilogram of matter) won’t be hit until, say, the mid-next century. Even in our current century, however, we’ll probably have many machines that compute more than all 10 billion human brains collectively and you can imagine, everything will change then! Jones: That is the big question. Is everything going to change? If so, what do you say to the next generation of leaders, currently coming out of college and university. So much of this change is already impacting how they study, how they will work, or how the future of work and livelihood is defined. What is their purpose and how do we change our systems so they will adapt to this new version of intelligence? Schmidhuber: For decades, people have asked me questions like that, because you know what I'm saying now, I have basically said since the 1970s, it’s just that today, people are paying more attention because, back then, they thought this was science fiction. They didn't think that I would ever come close to achieving my crazy life goal of building a machine that learns to become smarter than myself such that I can retire. But now many have changed their minds and think it's conceivable. And now I have two daughters, 23 and 25. People ask me: what do I tell them? They know that Daddy always said, “It seems likely that within your lifetimes, you will have new types of intelligence that are probably going to be superior in many ways, and probably all kinds of interesting ways.” How should they prepare for that? And I kept telling them the obvious: Learn how to learn new things! It's not like in the previous millennium where within 20 years someone learned to be a useful member of society, and then took a job for 40 years and performed in this job until she received her pension. Now things are changing much faster and we must learn continuously just to keep up. I also told my girls that no matter how smart AIs are going to get, learn at least the basics of math and physics, because that’s the essence of our universe, and anybody who understands this will have an advantage, and learn all kinds of new things more easily. I also told them that social skills will remain important, because most future jobs for humans will continue to involve interactions with other humans, but I couldn’t teach them anything about that; they know much more about social skills than I do. You touched on the big philosophical question about people’s purpose. Can this be answered without answering the even grander question: What’s the purpose of the entire universe? We don’t know. But what’s happening right now might be connected to the unknown answer. Don’t think of humans as the crown of creation. Instead view human civilization as part of a much grander scheme, an important step (but not the last one) on the path of the universe from very simple initial conditions towards more and more unfathomable complexity. Now it seems ready to take its next step, a step comparable to the invention of life itself over 3.5 billion years ago. Alas, don’t worry, in the end, all will be good! Jones: Let’s get back to this transformation happening right now with OpenAI. There are many questioning the efficacy and accuracy of ChatGPT, and are concerned its release has been premature. In light of the rampant adoption, educators have banned its use over concerns of plagiarism and how it stifles individual development. Should large language models like ChatGPT be used in school? Schmidhuber: When the calculator was first introduced, instructors forbade students from using it in school. Today, the consensus is that kids should learn the basic methods of arithmetic, but they should also learn to use the “artificial multipliers” aka calculators, even in exams, because laziness and efficiency is a hallmark of intelligence. Any intelligent being wants to minimize its efforts to achieve things. And that's the reason why we have tools, and why our kids are learning to use these tools. The first stone tools were invented maybe 3.5 million years ago; tools just have become more sophisticated over time. In fact, humans have changed in response to the properties of their tools. Our anatomical evolution was shaped by tools such as spears and fire. So, it's going to continue this way. And there is no permanent way of preventing large language models from being used in school. Jones: And when our children, your children graduate, what does their future work look like? Schmidhuber: A single human trying to predict details of how 10 billion people and their machines will evolve in the future is like a single neuron in my brain trying to predict what the entire brain and its tens of billions of neurons will do next year. 40 years ago, before the WWW was created at CERN in Switzerland, who would have predicted all those young people making money as YouTube video bloggers? Nevertheless, let’s make a few limited job-related observations. For a long time, people have thought that desktop jobs may require more intelligence than skills trade or handicraft professions. But now, it turns out that it's much easier to replace certain aspects of desktop jobs than replacing a carpenter, for example. Because everything that works well in AI is happening behind the screen currently, but not so much in the physical world. There are now artificial systems that can read lots of documents and then make really nice summaries of these documents. That is a desktop job. Or you give them a description of an illustration that you want to have for your article and pretty good illustrations are being generated that may need some minimal fine-tuning. But you know, all these desktop jobs are much easier to facilitate than the real tough jobs in the physical world. And it's interesting that the things people thought required intelligence, like playing chess, or writing or summarizing documents, are much easier for machines than they thought. But for things like playing football or soccer, there is no physical robot that can remotely compete with the abilities of a little boy with these skills. So, AI in the physical world, interestingly, is much harder than AI behind the screen in virtual worlds. And it's really exciting, in my opinion, to see that jobs such as plumbers are much more challenging than playing chess or writing another tabloid story. Jones: The way data has been collected in these large language models does not guarantee personal information has not been excluded. Current consent laws already are outdated when it comes to these large language models (LLM). The concern, rightly so, is increasing surveillance and loss of privacy. What is your view on this? Schmidhuber: As I have indicated earlier: are surveillance and loss of privacy inevitable consequences of increasingly complex societies? Super-organisms such as cities and states and companies consist of numerous people, just like people consist of numerous cells. These cells enjoy little privacy. They are constantly monitored by specialized "police cells" and "border guard cells": Are you a cancer cell? Are you an external intruder, a pathogen? Individual cells sacrifice their freedom for the benefits of being part of a multicellular organism. Similarly, for super-organisms such as nations. Over 5000 years ago, writing enabled recorded history and thus became its inaugural and most important invention. Its initial purpose, however, was to facilitate surveillance, to track citizens and their tax payments. The more complex a super-organism, the more comprehensive its collection of information about its constituents. 200 years ago, at least, the parish priest in each village knew everything about all the village people, even about those who did not confess, because they appeared in the confessions of others. Also, everyone soon knew about the stranger who had entered the village, because some occasionally peered out of the window, and what they saw got around. Such control mechanisms were temporarily lost through anonymization in rapidly growing cities but are now returning with the help of new surveillance devices such as smartphones as part of digital nervous systems that tell companies and governments a lot about billions of users. Cameras and drones etc. are becoming increasingly tinier and more ubiquitous. More effective recognition of faces and other detection technology are becoming cheaper and cheaper, and many will use it to identify others anywhere on earth; the big wide world will not offer any more privacy than the local village. Is this good or bad? Some nations may find it easier than others to justify more complex kinds of super-organisms at the expense of the privacy rights of their constituents. Jones: So, there is no way to stop or change this process of collection, or how it continuously informs decisions over time? How do you see governance and rules responding to this, especially amid Italy’s ban on ChatGPT following suspected user data breach and the more recent news about the Meta’s record $1.3billion fine in the company’s handling of user information? Schmidhuber: Data collection has benefits and drawbacks, such as the loss of privacy. How to balance those? I have argued for addressing this through data ownership in data markets. If it is true that data is the new oil, then it should have a price, just like oil. At the moment, the major surveillance platforms such as Meta do not offer users any money for their data and the transitive loss of privacy. In the future, however, we will likely see attempts at creating efficient data markets to figure out the data's true financial value through the interplay between supply and demand. Even some of the sensitive medical data should not be priced by governmental regulators but by patients (and healthy persons) who own it and who may sell or license parts thereof as micro-entrepreneurs in a healthcare data market. Following a previous interview, I gave for one of the largest re-insurance companies , let's look at the different participants in such a data market: patients, hospitals, data companies. (1) Patients with a rare form of cancer can offer more valuable data than patients with a very common form of cancer. (2) Hospitals and their machines are needed to extract the data, e.g., through magnet spin tomography, radiology, evaluations through human doctors, and so on. (3) Companies such as Siemens, Google or IBM would like to buy annotated data to make better artificial neural networks that learn to predict pathologies and diseases and the consequences of therapies. Now the market’s invisible hand will decide about the data’s price through the interplay between demand and supply. On the demand side, you will have several companies offering something for the data, maybe through an app on the smartphone (a bit like a stock market app). On the supply side, each patient in this market should be able to profit from high prices for rare valuable types of data. Likewise, competing data extractors such as hospitals will profit from gaining recognition and trust for extracting data well at a reasonable price. The market will make the whole system efficient through incentives for all who are doing a good job. Soon there will be a flourishing ecosystem of commercial data market advisors and what not, just like the ecosystem surrounding the traditional stock market. The value of the data won’t be determined by governments or ethics committees, but by those who own the data and decide by themselves which parts thereof they want to license to others under certain conditions. At first glance, a market-based system seems to be detrimental to the interest of certain monopolistic companies, as they would have to pay for the data - some would prefer free data and keep their monopoly. However, since every healthy and sick person in the market would suddenly have an incentive to collect and share their data under self-chosen anonymity conditions, there will soon be many more useful data to evaluate all kinds of treatments. On average, people will live longer and healthier, and many companies and the entire healthcare system will benefit. Jones: Finally, what is your view on open source versus the private companies like Google and OpenAI? Is there a danger to supporting these private companies’ large language models versus trying to keep these models open source and transparent, very much like what LAION is doing? Schmidhuber: I signed this open letter by LAION because I strongly favor the open-source movement. And I think it's also something that is going to challenge whatever big tech dominance there might be at the moment. Sure, the best models today are run by big companies with huge budgets for computers, but the exciting fact is that open-source models are not so far behind, some people say maybe six to eight months only. Of course, the private company models are all based on stuff that was created in academia, often in little labs without so much funding, which publish without patenting their results and open source their code and others take it and improved it. Big tech has profited tremendously from academia; their main achievement being that they have scaled up everything greatly, sometimes even failing to credit the original inventors. So, it's very interesting to see that as soon as some big company comes up with a new scaled-up model, lots of students out there are competing, or collaborating, with each other, trying to come up with equal or better performance on smaller networks and smaller machines. And since they are open sourcing, the next guy can have another great idea to improve it, so now there’s tremendous competition also for the big companies. Because of that, and since AI is still getting exponentially cheaper all the time, I don't believe that big tech companies will dominate in the long run. They find it very hard to compete with the enormous open-source movement. As long as you can encourage the open-source community, I think you shouldn't worry too much. Now, of course, you might say if everything is open source, then the bad actors also will more easily have access to these AI tools. And there's truth to that. But as always since the invention of controlled fire, it was good that knowledge about how technology works quickly became public such that everybody could use it. And then, against any bad actor, there's almost immediately a counter actor trying to nullify his efforts. You see, I still believe in our old motto "AI∀" or "AI For All." Jones: Thank you, Juergen for sharing your perspective on this amazing time in history. It’s clear that with new technology, the enormous potential can be matched by disparate and troubling risks which we’ve yet to solve, and even those we have yet to identify. If we are to dispel the fear of a sentient system for which we have no control, humans, alone need to take steps for more responsible development and collaboration to ensure AI technology is used to ultimately benefit society. Humanity will be judged by what we do next.

[N] Inside DeepMind's secret plot to break away from Google
reddit
LLM Vibe Score0
Human Vibe Score0
MassivePellfishThis week

[N] Inside DeepMind's secret plot to break away from Google

Article https://www.businessinsider.com/deepmind-secret-plot-break-away-from-google-project-watermelon-mario-2021-9 by Hugh Langley and Martin Coulter For a while, some DeepMind employees referred to it as "Watermelon." Later, executives called it "Mario." Both code names meant the same thing: a secret plan to break away from parent company Google. DeepMind feared Google might one day misuse its technology, and executives worked to distance the artificial-intelligence firm from its owner for years, said nine current and former employees who were directly familiar with the plans. This included plans to pursue an independent legal status that would distance the group's work from Google, said the people, who asked not to be identified discussing private matters. One core tension at DeepMind was that it sold the business to people it didn't trust, said one former employee. "Everything that happened since that point has been about them questioning that decision," the person added. Efforts to separate DeepMind from Google ended in April without a deal, The Wall Street Journal reported. The yearslong negotiations, along with recent shake-ups within Google's AI division, raise questions over whether the search giant can maintain control over a technology so crucial to its future. "DeepMind's close partnership with Google and Alphabet since the acquisition has been extraordinarily successful — with their support, we've delivered research breakthroughs that transformed the AI field and are now unlocking some of the biggest questions in science," a DeepMind spokesperson said in a statement. "Over the years, of course we've discussed and explored different structures within the Alphabet group to find the optimal way to support our long-term research mission. We could not be prouder to be delivering on this incredible mission, while continuing to have both operational autonomy and Alphabet's full support." When Google acquired DeepMind in 2014, the deal was seen as a win-win. Google got a leading AI research organization, and DeepMind, in London, won financial backing for its quest to build AI that can learn different tasks the way humans do, known as artificial general intelligence. But tensions soon emerged. Some employees described a cultural conflict between researchers who saw themselves firstly as academics and the sometimes bloated bureaucracy of Google's colossal business. Others said staff were immediately apprehensive about putting DeepMind's work under the control of a tech giant. For a while, some employees were encouraged to communicate using encrypted messaging apps over the fear of Google spying on their work. At one point, DeepMind's executives discovered that work published by Google's internal AI research group resembled some of DeepMind's codebase without citation, one person familiar with the situation said. "That pissed off Demis," the person added, referring to Demis Hassabis, DeepMind's CEO. "That was one reason DeepMind started to get more protective of their code." After Google restructured as Alphabet in 2015 to give riskier projects more freedom, DeepMind's leadership started to pursue a new status as a separate division under Alphabet, with its own profit and loss statement, The Information reported. DeepMind already enjoyed a high level of operational independence inside Alphabet, but the group wanted legal autonomy too. And it worried about the misuse of its technology, particularly if DeepMind were to ever achieve AGI. Internally, people started referring to the plan to gain more autonomy as "Watermelon," two former employees said. The project was later formally named "Mario" among DeepMind's leadership, these people said. "Their perspective is that their technology would be too powerful to be held by a private company, so it needs to be housed in some other legal entity detached from shareholder interest," one former employee who was close to the Alphabet negotiations said. "They framed it as 'this is better for society.'" In 2017, at a company retreat at the Macdonald Aviemore Resort in Scotland, DeepMind's leadership disclosed to employees its plan to separate from Google, two people who were present said. At the time, leadership said internally that the company planned to become a "global interest company," three people familiar with the matter said. The title, not an official legal status, was meant to reflect the worldwide ramifications DeepMind believed its technology would have. Later, in negotiations with Google, DeepMind pursued a status as a company limited by guarantee, a corporate structure without shareholders that is sometimes used by nonprofits. The agreement was that Alphabet would continue to bankroll the firm and would get an exclusive license to its technology, two people involved in the discussions said. There was a condition: Alphabet could not cross certain ethical redlines, such as using DeepMind technology for military weapons or surveillance. In 2019, DeepMind registered a new company called DeepMind Labs Limited, as well as a new holding company, filings with the UK's Companies House showed. This was done in anticipation of a separation from Google, two former employees involved in those registrations said. Negotiations with Google went through peaks and valleys over the years but gained new momentum in 2020, one person said. A senior team inside DeepMind started to hold meetings with outside lawyers and Google to hash out details of what this theoretical new formation might mean for the two companies' relationship, including specifics such as whether they would share a codebase, internal performance metrics, and software expenses, two people said. From the start, DeepMind was thinking about potential ethical dilemmas from its deal with Google. Before the 2014 acquisition closed, both companies signed an "Ethics and Safety Review Agreement" that would prevent Google from taking control of DeepMind's technology, The Economist reported in 2019. Part of the agreement included the creation of an ethics board that would supervise the research. Despite years of internal discussions about who should sit on this board, and vague promises to the press, this group "never existed, never convened, and never solved any ethics issues," one former employee close to those discussions said. A DeepMind spokesperson declined to comment. DeepMind did pursue a different idea: an independent review board to convene if it were to separate from Google, three people familiar with the plans said. The board would be made up of Google and DeepMind executives, as well as third parties. Former US president Barack Obama was someone DeepMind wanted to approach for this board, said one person who saw a shortlist of candidates. DeepMind also created an ethical charter that included bans on using its technology for military weapons or surveillance, as well as a rule that its technology should be used for ways that benefit society. In 2017, DeepMind started a unit focused on AI ethics research composed of employees and external research fellows. Its stated goal was to "pave the way for truly beneficial and responsible AI." A few months later, a controversial contract between Google and the Pentagon was disclosed, causing an internal uproar in which employees accused Google of getting into "the business of war." Google's Pentagon contract, known as Project Maven, "set alarm bells ringing" inside DeepMind, a former employee said. Afterward, Google published a set of principles to govern its work in AI, guidelines that were similar to the ethical charter that DeepMind had already set out internally, rankling some of DeepMind's senior leadership, two former employees said. In April, Hassabis told employees in an all-hands meeting that negotiations to separate from Google had ended. DeepMind would maintain its existing status inside Alphabet. DeepMind's future work would be overseen by Google's Advanced Technology Review Council, which includes two DeepMind executives, Google's AI chief Jeff Dean, and the legal SVP Kent Walker. But the group's yearslong battle to achieve more independence raises questions about its future within Google. Google's commitment to AI research has also come under question, after the company forced out two of its most senior AI ethics researchers. That led to an industry backlash and sowed doubt over whether it could allow truly independent research. Ali Alkhatib, a fellow at the Center for Applied Data Ethics, told Insider that more public accountability was "desperately needed" to regulate the pursuit of AI by large tech companies. For Google, its investment in DeepMind may be starting to pay off. Late last year, DeepMind announced a breakthrough to help scientists better understand the behavior of microscopic proteins, which has the potential to revolutionize drug discovery. As for DeepMind, Hassabis is holding on to the belief that AI technology should not be controlled by a single corporation. Speaking at Tortoise's Responsible AI Forum in June, he proposed a "world institute" of AI. Such a body might sit under the jurisdiction of the United Nations, Hassabis theorized, and could be filled with top researchers in the field. "It's much stronger if you lead by example," he told the audience, "and I hope DeepMind can be part of that role-modeling for the industry."

[Discussion]: Mark Zuckerberg on Meta's Strategy on Open Source and AI during the earnings call
reddit
LLM Vibe Score0
Human Vibe Score1
noiseinvacuumThis week

[Discussion]: Mark Zuckerberg on Meta's Strategy on Open Source and AI during the earnings call

During the recent earnings call, Mark Zuckerberg answered a question from Eric Sheridan of Goldman Sachs on Meta's AI strategy, opportunities to integrate into products, and why they open source models and how it would benefit their business. I found the reasoning to be very sound and promising for the OSS and AI community. The biggest risk from AI, in my opinion, is not the doomsday scenarios that intuitively come to mind but rather that the most powerful AI systems will only be accessible to the most powerful and resourceful corporations. Quote copied from Ben Thompson's write up on Meta's earning in his Stratechery blog post which goes beyond AI. It's behind a paywall but I highly recommend it personally. Some noteworthy quotes that signal the thought process at Meta FAIR and more broadly We’re just playing a different game on the infrastructure than companies like Google or Microsoft or Amazon We would aspire to and hope to make even more open than that. So, we’ll need to figure out a way to do that. ...lead us to do more work in terms of open sourcing, some of the lower level models and tools Open sourcing low level tools make the way we run all this infrastructure more efficient over time. On PyTorch: It’s generally been very valuable for us to provide that because now all of the best developers across the industry are using tools that we’re also using internally. I would expect us to be pushing and helping to build out an open ecosystem. For all the negative that comes out of the popular discourse on Meta, I think their work to open source key tech tools over the last 10 years has been exceptional, here's hoping it continues into this decade of AI and pushes other tech giants to also realize the benefits of Open Source. Full Transcript: Right now most of the companies that are training large language models have business models that lead them to a closed approach to development. I think there’s an important opportunity to help create an open ecosystem. If we can help be a part of this, then much of the industry will standardize on using these open tools and help improve them further. So this will make it easier for other companies to integrate with our products and platforms as we enable more integrations, and that will help our products stay at the leading edge as well. Our approach to AI and our infrastructure has always been fairly open. We open source many of our state of the art models so people can experiment and build with them. This quarter we released our LLaMa LLM to researchers. It has 65 billion parameters but outperforms larger models and has proven quite popular. We’ve also open-sourced three other groundbreaking visual models along with their training data and model weights — Segment Anything, DinoV2, and our Animated Drawings tool — and we’ve gotten positive feedback on all of those as well. I think that there’s an important distinction between the products we offer and a lot of the technical infrastructure, especially the software that we write to support that. And historically, whether it’s the Open Compute project that we’ve done or just open sourcing a lot of the infrastructure that we’ve built, we’ve historically open sourced a lot of that infrastructure, even though we haven’t open sourced the code for our core products or anything like that. And the reason why I think why we do this is that unlike some of the other companies in the space, we’re not selling a cloud computing service where we try to keep the different software infrastructure that we’re building proprietary. For us, it’s way better if the industry standardizes on the basic tools that we’re using and therefore we can benefit from the improvements that others make and others’ use of those tools can, in some cases like Open Compute, drive down the costs of those things which make our business more efficient too. So I think to some degree we’re just playing a different game on the infrastructure than companies like Google or Microsoft or Amazon, and that creates different incentives for us. So overall, I think that that’s going to lead us to do more work in terms of open sourcing, some of the lower level models and tools. But of course, a lot of the product work itself is going to be specific and integrated with the things that we do. So it’s not that everything we do is going to be open. Obviously, a bunch of this needs to be developed in a way that creates unique value for our products, but I think in terms of the basic models, I would expect us to be pushing and helping to build out an open ecosystem here, which I think is something that’s going to be important. On the AI tools, and we have a bunch of history here, right? So if you if you look at what we’ve done with PyTorch, for example, which has generally become the standard in the industry as a tool that a lot of folks who are building AI models and different things in that space use, it’s generally been very valuable for us to provide that because now all of the best developers across the industry are using tools that we’re also using internally. So the tool chain is the same. So when they create some innovation, we can easily integrate it into the things that we’re doing. When we improve something, it improves other products too. Because it’s integrated with our technology stack, when there are opportunities to make integrations with products, it’s much easier to make sure that developers and other folks are compatible with the things that we need in the way that our systems work. So there are a lot of advantages, but I view this more as a kind of back end infrastructure advantage with potential integrations on the product side, but one that should hopefully enable us to stay at the leading edge and integrate more broadly with the community and also make the way we run all this infrastructure more efficient over time. There are a number of models. I just gave PyTorch as an example. Open Compute is another model that has worked really well for us in this way, both to incorporate both innovation and scale efficiency into our own infrastructure. So I think that there’s, our incentives I think are basically aligned towards moving in this direction. Now that said, there’s a lot to figure out, right? So when you asked if there are going to be other opportunities, I hope so. I can’t speak to what all those things might be now. This is all quite early in getting developed. The better we do at the foundational work, the more opportunities I think that will come and present themselves. So I think that that’s all stuff that we need to figure out. But at least at the base level, I think we’re generally incentivized to move in this direction. And we also need to figure out how to go in that direction over time. I mean, I mentioned LLaMA before and I also want to be clear that while I’m talking about helping contribute to an open ecosystem, LLaMA is a model that we only really made available to researchers and there’s a lot of really good stuff that’s happening there. But a lot of the work that we’re doing, I think, we would aspire to and hope to make even more open than that. So, we’ll need to figure out a way to do that.

[D] How Facebook got addicted to spreading misinformation
reddit
LLM Vibe Score0
Human Vibe Score0
proof_requiredThis week

[D] How Facebook got addicted to spreading misinformation

Behind paywall: With new machine-learning models coming online daily, the company created a new system to track their impact and maximize user engagement. The process is still the same today. Teams train up a new machine-learning model on FBLearner, whether to change the ranking order of posts or to better catch content that violates Facebook’s community standards (its rules on what is and isn’t allowed on the platform). Then they test the new model on a small subset of Facebook’s users to measure how it changes engagement metrics, such as the number of likes, comments, and shares, says Krishna Gade, who served as the engineering manager for news feed from 2016 to 2018. If a model reduces engagement too much, it’s discarded. Otherwise, it’s deployed and continually monitored. On Twitter, Gade explained that his engineers would get notifications every few days when metrics such as likes or comments were down. Then they’d decipher what had caused the problem and whether any models needed retraining. But this approach soon caused issues. The models that maximize engagement also favor controversy, misinformation, and extremism: put simply, people just like outrageous stuff. Sometimes this inflames existing political tensions. The most devastating example to date is the case of Myanmar, where viral fake news and hate speech about the Rohingya Muslim minority escalated the country’s religious conflict into a full-blown genocide. Facebook admitted in 2018, after years of downplaying its role, that it had not done enough “to help prevent our platform from being used to foment division and incite offline violence.” While Facebook may have been oblivious to these consequences in the beginning, it was studying them by 2016. In an internal presentation from that year, reviewed by the Wall Street Journal, a company researcher, Monica Lee, found that Facebook was not only hosting a large number of extremist groups but also promoting them to its users: “64% of all extremist group joins are due to our recommendation tools,” the presentation said, predominantly thanks to the models behind the “Groups You Should Join” and “Discover” features. https://www.technologyreview.com/2021/03/11/1020600/facebook-responsible-ai-misinformation/

[N] Montreal-based Element AI sold for $230-million as founders saw value mostly wiped out
reddit
LLM Vibe Score0
Human Vibe Score1
sensetimeThis week

[N] Montreal-based Element AI sold for $230-million as founders saw value mostly wiped out

According to Globe and Mail article: Element AI sold for $230-million as founders saw value mostly wiped out, document reveals Montreal startup Element AI Inc. was running out of money and options when it inked a deal last month to sell itself for US$230-milion to Silicon Valley software company ServiceNow Inc., a confidential document obtained by the Globe and Mail reveals. Materials sent to Element AI shareholders Friday reveal that while many of its institutional shareholders will make most if not all of their money back from backing two venture financings, employees will not fare nearly as well. Many have been terminated and had their stock options cancelled. Also losing out are co-founders Jean-François Gagné, the CEO, his wife Anne Martel, the chief administrative officer, chief science officer Nick Chapados and Yoshua Bengio, the University of Montreal professor known as a godfather of “deep learning,” the foundational science behind today’s AI revolution. Between them, they owned 8.8 million common shares, whose value has been wiped out with the takeover, which goes to a shareholder vote Dec 29 with enough investor support already locked up to pass before the takeover goes to a Canadian court to approve a plan of arrangement with ServiceNow. The quartet also owns preferred shares worth less than US$300,000 combined under the terms of the deal. The shareholder document, a management proxy circular, provides a rare look inside efforts by a highly hyped but deeply troubled startup as it struggled to secure financing at the same time as it was failing to live up to its early promises. The circular states the US$230-million purchase price is subject to some adjustments and expenses which could bring the final price down to US$195-million. The sale is a disappointing outcome for a company that burst onto the Canadian tech scene four years ago like few others, promising to deliver AI-powered operational improvements to a range of industries and anchor a thriving domestic AI sector. Element AI became the self-appointed representative of Canada’s AI sector, lobbying politicians and officials and landing numerous photo ops with them, including Prime Minister Justin Trudeau. It also secured $25-million in federal funding – $20-million of which was committed earlier this year and cancelled by the government with the ServiceNow takeover. Element AI invested heavily in hype and and earned international renown, largely due to its association with Dr. Bengio. It raised US$102-million in venture capital in 2017 just nine months after its founding, an unheard of amount for a new Canadian company, from international backers including Microsoft Corp., Intel Corp., Nvidia Corp., Tencent Holdings Ltd., Fidelity Investments, a Singaporean sovereign wealth fund and venture capital firms. Element AI went on a hiring spree to establish what the founders called “supercredibility,” recruiting top AI talent in Canada and abroad. It opened global offices, including a British operation that did pro bono work to deliver “AI for good,” and its ranks swelled to 500 people. But the swift hiring and attention-seeking were at odds with its success in actually building a software business. Element AI took two years to focus on product development after initially pursuing consulting gigs. It came into 2019 with a plan to bring several AI-based products to market, including a cybersecurity offering for financial institutions and a program to help port operators predict waiting times for truck drivers. It was also quietly shopping itself around. In December 2018, the company asked financial adviser Allen & Co LLC to find a potential buyer, in addition to pursuing a private placement, the circular reveals. But Element AI struggled to advance proofs-of-concept work to marketable products. Several client partnerships faltered in 2019 and 2020. Element did manage to reach terms for a US$151.4-million ($200-million) venture financing in September, 2019 led by the Caisse de dépôt et placement du Québec and backed by the Quebec government and consulting giant McKinsey and Co. However, the circular reveals the company only received the first tranche of the financing – roughly half of the amount – at the time, and that it had to meet unspecified conditions to get the rest. A fairness opinion by Deloitte commissioned as part of the sale process estimated Element AI’s enterprises value at just US$76-million around the time of the 2019 financing, shrinking to US$45-million this year. “However, the conditions precedent the closing of the second tranche … were not going to be met in a timely manner,” the circular reads. It states “new terms were proposed” for a round of financing that would give incoming investors ranking ahead of others and a cumulative dividend of 12 per cent on invested capital and impose “other operating and governance constraints and limitations on the company.” Management instead decided to pursue a sale, and Allen contacted prospective buyers in June. As talks narrowed this past summer to exclusive negotiations with ServiceNow, “the company’s liquidity was diminishing as sources of capital on acceptable terms were scarce,” the circular reads. By late November, it was generating revenue at an annualized rate of just $10-million to $12-million, Deloitte said. As part of the deal – which will see ServiceNow keep Element AI’s research scientists and patents and effectively abandon its business – the buyer has agreed to pay US$10-million to key employees and consultants including Mr. Gagne and Dr. Bengio as part of a retention plan. The Caisse and Quebec government will get US$35.45-million and US$11.8-million, respectively, roughly the amount they invested in the first tranche of the 2019 financing.

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.
reddit
LLM Vibe Score0
Human Vibe Score0.6
AlexSnakeKingThis week

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.

TD;LR: At Company A, Team X does advanced analytics using on-prem ERP tools and older programming languages. Their tools work very well and are designed based on very deep business and domain expertise. Team Y is a new and ambitious Data Science team that thinks they can replace Team X's tools with a bunch of R scripts and a custom built ML platform. Their models are simplistic, but more "fashionable" compared to the econometric models used by Team X, and team Y benefits from the ML/DS moniker so leadership is allowing Team Y to start a large scale overhaul of the analytics platform in question. Team Y doesn't have the experience for such a larger scale transformation, and is refusing to collaborate with team X. This project is very likely going to fail, and cause serious harm to the company as a whole financially and from a people perspective. I argue that this is not just because of bad leadership, but also because of various trends and mindsets in the DS community at large. Update (Jump to below the line for the original story): Several people in the comments are pointing out that this just a management failure, not something due to ML/DS, and that you can replace DS with any buzz tech and the story will still be relevant. My response: Of course, any failure at an organization level is ultimately a management failure one way or the other. Moreover, it is also the case that ML/DS when done correctly, will always improve a company's bottom line. There is no scenario where the proper ML solution, delivered at a reasonable cost and in a timely fashion, will somehow hurt the company's bottom line. My point is that in this case management is failing because of certain trends and practices that are specific to the ML/DS community, namely: The idea that DS teams should operate independently of tech and business orgs -- too much autonomy for DS teams The disregard for domain knowledge that seems prevalent nowadays thanks to the ML hype, that DS can be generalists and someone with good enough ML chops can solve any business problem. That wasn't the case when I first left academia for the industry in 2009 (back then nobody would even bother with a phone screen if you didn't have the right domain knowledge). Over reliance on resources who check all the ML hype related boxes (knows Python, R, Tensorflow, Shiny, etc..., has the right Coursera certifications, has blogged on the topic, etc...), but are lacking in depth of experience. DS interviews nowadays all seem to be: Can you tell me what a p-value is? What is elastic net regression? Show me how to fit a model in sklearn? How do you impute NAs in an R dataframe? Any smart person can look those up on Stackoverflow or Cross-Validated,.....Instead teams should be asking stuff like: why does portfolio optimization use QP not LP? How does a forecast influence a customer service level? When should a recommendation engine be content based and when should it use collaborative filtering? etc... (This is a true story, happening to the company I currently work for. Names, domains, algorithms, and roles have been shuffled around to protect my anonymity)  Company A has been around for several decades. It is not the biggest name in its domain, but it is a well respected one. Risk analysis and portfolio optimization have been a core of Company A's business since the 90s. They have a large team of 30 or so analysts who perform those tasks on a daily basis. These analysts use ERP solutions implemented for them by one the big ERP companies (SAP, Teradata, Oracle, JD Edwards,...) or one of the major tech consulting companies (Deloitte, Accenture, PWC, Capgemini, etc...) in collaboration with their own in house engineering team. The tools used are embarrassingly old school: Classic RDBMS running on on-prem servers or maybe even on mainframes, code written in COBOL, Fortran, weird proprietary stuff like ABAP or SPSS.....you get the picture. But the models and analytic functions were pretty sophisticated, and surprisingly cutting edge compared to the published academic literature. Most of all, they fit well with the company's enterprise ecosystem, and were honed based on years of deep domain knowledge.  They have a tech team of several engineers (poached from the aforementioned software and consulting companies) and product managers (who came from the experienced pools of analysts and managers who use the software, or poached from business rivals) maintaining and running this software. Their technology might be old school, but collectively, they know the domain and the company's overall architecture very, very well. They've guided the company through several large scale upgrades and migrations and they have a track record of delivering on time, without too much overhead. The few times they've stumbled, they knew how to pick themselves up very quickly. In fact within their industry niche, they have a reputation for their expertise, and have very good relations with the various vendors they've had to deal with. They were the launching pad of several successful ERP consulting careers.  Interestingly, despite dealing on a daily basis with statistical modeling and optimization algorithms, none of the analysts, engineers, or product managers involved describe themselves as data scientists or machine learning experts. It is mostly a cultural thing: Their expertise predates the Data Science/ML hype that started circa 2010, and they got most of their chops using proprietary enterprise tools instead of the open source tools popular nowadays. A few of them have formal statistical training, but most of them came from engineering or domain backgrounds and learned stats on the fly while doing their job. Call this team "Team X".  Sometime around the mid 2010s, Company A started having some serious anxiety issues: Although still doing very well for a company its size, overall economic and demographic trends were shrinking its customer base, and a couple of so called disruptors came up with a new app and business model that started seriously eating into their revenue. A suitable reaction to appease shareholders and Wall Street was necessary. The company already had a decent website and a pretty snazzy app, what more could be done? Leadership decided that it was high time that AI and ML become a core part of the company's business. An ambitious Manager, with no science or engineering background, but who had very briefly toyed with a recommender system a couple of years back, was chosen to build a data science team, call it team "Y" (he had a bachelor's in history from the local state college and worked for several years in the company's marketing org). Team "Y" consists mostly of internal hires who decided they wanted to be data scientists and completed a Coursera certification or a Galvanize boot camp, before being brought on to the team, along with a few of fresh Ph.D or M.Sc holders who didn't like academia and wanted to try their hand at an industry role. All of them were very bright people, they could write great Medium blog posts and give inspiring TED talks, but collectively they had very little real world industry experience. As is the fashion nowadays, this group was made part of a data science org that reported directly to the CEO and Board, bypassing the CIO and any tech or business VPs, since Company A wanted to claim the monikers "data driven" and "AI powered" in their upcoming shareholder meetings. In 3 or 4 years of existence, team Y produced a few Python and R scripts. Their architectural experience  consisted almost entirely in connecting Flask to S3 buckets or Redshift tables, with a couple of the more resourceful ones learning how to plug their models into Tableau or how to spin up a Kuberneties pod.  But they needn't worry: The aforementioned manager, who was now a director (and was also doing an online Masters to make up for his qualifications gap and bolster his chances of becoming VP soon - at least he now understands what L1 regularization is), was a master at playing corporate politics and self-promotion. No matter how few actionable insights team Y produced or how little code they deployed to production, he always had their back and made sure they had ample funding. In fact he now had grandiose plans for setting up an all-purpose machine learning platform that can be used to solve all of the company's data problems.  A couple of sharp minded members of team Y, upon googling their industry name along with the word "data science", realized that risk analysis was a prime candidate for being solved with Bayesian models, and there was already a nifty R package for doing just that, whose tutorial they went through on R-Bloggers.com. One of them had even submitted a Bayesian classifier Kernel for a competition on Kaggle (he was 203rd on the leaderboard), and was eager to put his new-found expertise to use on a real world problem. They pitched the idea to their director, who saw a perfect use case for his upcoming ML platform. They started work on it immediately, without bothering to check whether anybody at Company A was already doing risk analysis. Since their org was independent, they didn't really need to check with anybody else before they got funding for their initiative. Although it was basically a Naive Bayes classifier, the term ML was added to the project tile, to impress the board.  As they progressed with their work however, tensions started to build. They had asked the data warehousing and CA analytics teams to build pipelines for them, and word eventually got out to team X about their project. Team X was initially thrilled: They offered to collaborate whole heartedly, and would have loved to add an ML based feather to their already impressive cap. The product owners and analysts were totally onboard as well: They saw a chance to get in on the whole Data Science hype that they kept hearing about. But through some weird mix of arrogance and insecurity, team Y refused to collaborate with them or share any of their long term goals with them, even as they went to other parts of the company giving brown bag presentations and tutorials on the new model they created.  Team X got resentful: from what they saw of team Y's model, their approach was hopelessly naive and had little chances of scaling or being sustainable in production, and they knew exactly how to help with that. Deploying the model to production would have taken them a few days, given how comfortable they were with DevOps and continuous delivery (team Y had taken several months to figure out how to deploy a simple R script to production). And despite how old school their own tech was, team X were crafty enough to be able to plug it in to their existing architecture. Moreover, the output of the model was such that it didn't take into account how the business will consume it or how it was going to be fed to downstream systems, and the product owners could have gone a long way in making the model more amenable to adoption by the business stakeholders. But team Y wouldn't listen, and their leads brushed off any attempts at communication, let alone collaboration. The vibe that team Y was giving off was "We are the cutting edge ML team, you guys are the legacy server grunts. We don't need your opinion.", and they seemed to have a complete disregard for domain knowledge, or worse, they thought that all that domain knowledge consisted of was being able to grasp the definitions of a few business metrics.  Team X got frustrated and tried to express their concerns to leadership. But despite owning a vital link in Company A's business process, they were only \~50 people in a large 1000 strong technology and operations org, and they were several layers removed from the C-suite, so it was impossible for them to get their voices heard.  Meanwhile, the unstoppable director was doing what he did best: Playing corporate politics. Despite how little his team had actually delivered, he had convinced the board that all analysis and optimization tasks should now be migrated to his yet to be delivered ML platform. Since most leaders now knew that there was overlap between team Y and team X's objectives, his pitch was no longer that team Y was going to create a new insight, but that they were going to replace (or modernize) the legacy statistics based on-prem tools with more accurate cloud based ML tools. Never mind that there was no support in the academic literature for the idea that Naive Bayes works better than the Econometric approaches used by team X, let alone the additional wacky idea that Bayesian Optimization would definitely outperform the QP solvers that were running in production.  Unbeknownst to team X, the original Bayesian risk analysis project has now grown into a multimillion dollar major overhaul initiative, which included the eventual replacement of all of the tools and functions supported by team X along with the necessary migration to the cloud. The CIO and a couple of business VPs are on now board, and tech leadership is treating it as a done deal. An outside vendor, a startup who nobody had heard of, was contracted to help build the platform, since team Y has no engineering skills. The choice was deliberate, as calling on any of the established consulting or software companies would have eventually led leadership to the conclusion that team X was better suited for a transformation on this scale than team Y.  Team Y has no experience with any major ERP deployments, and no domain knowledge, yet they are being tasked with fundamentally changing the business process that is at the core of Company A's business. Their models actually perform worse than those deployed by team X, and their architecture is hopelessly simplistic, compared to what is necessary for running such a solution in production.  Ironically, using Bayesian thinking and based on all the evidence, the likelihood that team Y succeeds is close to 0%. At best, the project is going to end up being a write off of 50 million dollars or more. Once the !@#$!@hits the fan, a couple of executive heads are going to role, and dozens of people will get laid off. At worst, given how vital risk analysis and portfolio optimization is to Company A's revenue stream, the failure will eventually sink the whole company. It probably won't go bankrupt, but it will lose a significant portion of its business and work force. Failed ERP implementations can and do sink large companies: Just see what happened to National Grid US, SuperValu or Target Canada.  One might argue that this is more about corporate disfunction and bad leadership than about data science and AI. But I disagree. I think the core driver of this debacle is indeed the blind faith in Data Scientists, ML models and the promise of AI, and the overall culture of hype and self promotion that is very common among the ML crowd.  We haven't seen the end of this story: I sincerely hope that this ends well for the sake of my colleagues and all involved. Company A is a good company, and both its customers and its employees deserver better. But the chances of that happening are negligible given all the information available, and this failure will hit my company hard.

[D] LLMs causing more harm than good for the field?
reddit
LLM Vibe Score0
Human Vibe Score1
Stevens97This week

[D] LLMs causing more harm than good for the field?

This post might be a bit ranty, but i feel more and more share this sentiment with me as of late. If you bother to read this whole post feel free to share how you feel about this. When OpenAI put the knowledge of AI in the everyday household, I was at first optimistic about it. In smaller countries outside the US, companies were very hesitant before about AI, they thought it felt far away and something only big FANG companies were able to do. Now? Its much better. Everyone is interested in it and wants to know how they can use AI in their business. Which is great! Pre-ChatGPT-times, when people asked me what i worked with and i responded "Machine Learning/AI" they had no clue and pretty much no further interest (Unless they were a tech-person) Post-ChatGPT-times, when I get asked the same questions I get "Oh, you do that thing with the chatbots?" Its a step in the right direction, I guess. I don't really have that much interest in LLMs and have the privilege to work exclusively on vision related tasks unlike some other people who have had to pivot to working full time with LLMs. However, right now I think its almost doing more harm to the field than good. Let me share some of my observations, but before that I want to highlight I'm in no way trying to gatekeep the field of AI in any way. I've gotten job offers to be "ChatGPT expert", What does that even mean? I strongly believe that jobs like these don't really fill a real function and is more of a "hypetrain"-job than a job that fills any function at all. Over the past years I've been going to some conferences around Europe, one being last week, which has usually been great with good technological depth and a place for Data-scientists/ML Engineers to network, share ideas and collaborate. However, now the talks, the depth, the networking has all changed drastically. No longer is it new and exiting ways companies are using AI to do cool things and push the envelope, its all GANs and LLMs with surface level knowledge. The few "old-school" type talks being sent off to a 2nd track in a small room The panel discussions are filled with philosophists with no fundamental knowledge of AI talking about if LLMs will become sentient or not. The spaces for data-scientists/ML engineers are quickly dissapearing outside the academic conferences, being pushed out by the current hypetrain. The hypetrain evangelists also promise miracles and gold with LLMs and GANs, miracles that they will never live up to. When the investors realize that the LLMs cant live up to these miracles they will instantly get more hesitant with funding for future projects within AI, sending us back into an AI-winter once again. EDIT: P.S. I've also seen more people on this reddit appearing claiming to be "Generative AI experts". But when delving deeper it turns out they are just "good prompters" and have no real knowledge, expertice or interest in the actual field of AI or Generative AI.

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[D] Overwhelmed by fast advances in recent weeks
reddit
LLM Vibe Score0
Human Vibe Score1
iamx9000againThis week

[D] Overwhelmed by fast advances in recent weeks

I was watching the GTC keynote and became entirely overwhelmed by the amount of progress achieved from last year. I'm wondering how everyone else feels. &#x200B; Firstly, the entire ChatGPT, GPT-3/GPT-4 chaos has been going on for a few weeks, with everyone scrambling left and right to integrate chatbots into their apps, products, websites. Twitter is flooded with new product ideas, how to speed up the process from idea to product, countless promp engineering blogs, tips, tricks, paid courses. &#x200B; Not only was ChatGPT disruptive, but a few days later, Microsoft and Google also released their models and integrated them into their search engines. Microsoft also integrated its LLM into its Office suite. It all happenned overnight. I understand that they've started integrating them along the way, but still, it seems like it hapenned way too fast. This tweet encompases the past few weeks perfectly https://twitter.com/AlphaSignalAI/status/1638235815137386508 , on a random Tuesday countless products are released that seem revolutionary. &#x200B; In addition to the language models, there are also the generative art models that have been slowly rising in mainstream recognition. Now Midjourney AI is known by a lot of people who are not even remotely connected to the AI space. &#x200B; For the past few weeks, reading Twitter, I've felt completely overwhelmed, as if the entire AI space is moving beyond at lightning speed, whilst around me we're just slowly training models, adding some data, and not seeing much improvement, being stuck on coming up with "new ideas, that set us apart". &#x200B; Watching the GTC keynote from NVIDIA I was again, completely overwhelmed by how much is being developed throughout all the different domains. The ASML EUV (microchip making system) was incredible, I have no idea how it does lithography and to me it still seems like magic. The Grace CPU with 2 dies (although I think Apple was the first to do it?) and 100 GB RAM, all in a small form factor. There were a lot more different hardware servers that I just blanked out at some point. The omniverse sim engine looks incredible, almost real life (I wonder how much of a domain shift there is between real and sim considering how real the sim looks). Beyond it being cool and usable to train on synthetic data, the car manufacturers use it to optimize their pipelines. This change in perspective, of using these tools for other goals than those they were designed for I find the most interesting. &#x200B; The hardware part may be old news, as I don't really follow it, however the software part is just as incredible. NVIDIA AI foundations (language, image, biology models), just packaging everything together like a sandwich. Getty, Shutterstock and Adobe will use the generative models to create images. Again, already these huge juggernauts are already integrated. &#x200B; I can't believe the point where we're at. We can use AI to write code, create art, create audiobooks using Britney Spear's voice, create an interactive chatbot to converse with books, create 3D real-time avatars, generate new proteins (?i'm lost on this one), create an anime and countless other scenarios. Sure, they're not perfect, but the fact that we can do all that in the first place is amazing. &#x200B; As Huang said in his keynote, companies want to develop "disruptive products and business models". I feel like this is what I've seen lately. Everyone wants to be the one that does something first, just throwing anything and everything at the wall and seeing what sticks. &#x200B; In conclusion, I'm feeling like the world is moving so fast around me whilst I'm standing still. I want to not read anything anymore and just wait until everything dies down abit, just so I can get my bearings. However, I think this is unfeasible. I fear we'll keep going in a frenzy until we just burn ourselves at some point. &#x200B; How are you all fairing? How do you feel about this frenzy in the AI space? What are you the most excited about?

[D] Last Week in Medical AI: Top LLM Research Papers/Models (December 7 - December 14, 2024)
reddit
LLM Vibe Score0
Human Vibe Score0
aadityauraThis week

[D] Last Week in Medical AI: Top LLM Research Papers/Models (December 7 - December 14, 2024)

[\[D\] Last Week in Medical AI: Top LLM Research Papers\/Models \(December 7 - December 14, 2024\)](https://preview.redd.it/o23fp3csj07e1.jpg?width=1280&format=pjpg&auto=webp&s=69e19fc351b3aa5e34c4c00e66245583f88bd9bb) Medical LLM & Other Models PediaBench: Chinese Pediatric LLM This paper introduces PediaBench, the first Chinese pediatric dataset for evaluating Large Language Model (LLM) question-answering performance, containing 4,565 objective and 1,632 subjective questions across 12 disease groups. BiMediX: Bilingual Medical LLM This paper introduces BiMediX, the first bilingual (English-Arabic) medical Mixture of Experts LLM, along with BiMed1.3M, a 1.3M bilingual medical instruction dataset with over 632M tokens used for training. Diverse medical knowledge integration This paper introduces BiMediX2, a bilingual (Arabic-English) Large Multimodal Model (LMM) based on Llama3.1 architecture, trained on 1.6M medical interaction samples. BRAD: Digital Biology Language Model This paper introduces BRAD (Bioinformatics Retrieval Augmented Digital assistant), an LLM-powered chatbot and agent system integrating various bioinformatics tools. MMedPO: Vision-Language Medical LLM This paper introduces MMedPO, a multimodal medical preference optimization approach to improve factual accuracy in Medical Large Vision-Language Models (Med-LVLMs) by addressing modality misalignment. Frameworks & Methodologies \- TOP-Training: Medical Q&A Framework \- Hybrid RAG: Secure Medical Data Management \- Zero-Shot ATC Clinical Coding \- Chest X-Ray Diagnosis Architecture \- Medical Imaging AI Democratization Benchmarks & Evaluations \- KorMedMCQA: Korean Healthcare Licensing Benchmark \- Large Language Model Medical Tasks \- Clinical T5 Model Performance Study \- Radiology Report Quality Assessment \- Genomic Analysis Benchmarking LLM Applications \- TCM-FTP: Herbal Prescription Prediction \- LLaSA: Activity Analysis via Sensors \- Emergency Department Visit Predictions \- Neurodegenerative Disease AI Diagnosis \- Kidney Disease Explainable AI Model Ethical AI & Privacy \- Privacy-Preserving LLM Mechanisms \- AI-Driven Digital Organism Modeling \- Biomedical Research Automation \- Multimodality in Medical Practice Full thread in detail: https://x.com/OpenlifesciAI/status/1867999825721242101

[D] Working with Various OpenAI Models - My Thoughts and Experiences
reddit
LLM Vibe Score0
Human Vibe Score1
bart_soThis week

[D] Working with Various OpenAI Models - My Thoughts and Experiences

I'd like to share some of my insights from working with OpenAI models on my project. I'm not exactly a tech person, so some of these observations might be obvious to some of you, but I think they're worth sharing for those with less experience or who aren't directly in the field. Intro: In early February, my friends and I started a side project where we aimed to build an AI portal called DoMoreAI. For the first two months, we focused on creating an AI tools catalog. Our experiment is based on the idea that in the future, companies will be "Managed by AI, and Driven by Humans." So, our goal was to leave as much as possible to AI and automation, with all the consequences that come with it. As mentioned before, I'm not a tech guy, but I've been playing with OpenAI models for the past few years, so I had some experience when starting this project. Tasks We Assigned to AI: Based on an AI tool's front page, we had the AI write a one-sentence summary of an AI project + write a more in-depth review of the project, categorize the project into different categories (WHAT category, like blog; TASK category, like writing; FOR category, like content creator), decide if the project offers iOS app, Android app, browser extension, API, find social media links, process information about prices and pricing policy, and more. Interesting Findings: When working on a more complex prompt, particularly one with several tasks, you have to be patient when crafting it. You might eventually find the right wording to achieve the desired results, but it takes time and lots of trial and error. You might even be surprised by what works and what doesn't. If cost isn't an issue, you can always break up one complex prompt into several smaller prompts. However, the more requests you send, the higher the chance of encountering errors like the 429 error, which may require setting up more sophisticated error handlers for the whole process. You need error handlers because, without them, the automation process will suffer. With more complex prompts, there are no prompts that always yield the expected results, so you have to plan for what to do if the results aren't satisfactory and how to determine if the result meets your expectations or not. GPT-3.0 struggled with outputting JSON strings as requested, but GPT-3.5 is much better at this task. I'd say the number of errors from improperly formatting the response in JSON is 3-4 times lower for GPT-3.5. AI models have trouble distinguishing words singular forms from plural forms. Just because you can use AI for a given task doesn't mean you should. Often, standard techniques like using regex can yield better results when extracting something from text than relying solely on AI. A hybrid solution often provides the best results. We're using ADA vector embeddings and Pinecone for semantic search in our catalog, and I was really surprised to find that this kind of semantic search works in any language. Even if all the content on our page is in English, you can search in another language and still get decent results. The Best Mishaps: As you may know, there's a token limit for requests, so we have to ensure that we don't send too long a part of the front page to the model. Sometimes, this led to funny situations. If the HTML of the page consists mainly of styles and the model is fed only with styles, then when you ask the AI to write a review of the project, it writes about how beautiful, mobile-friendly, etc., the project is. For one project, instead of writing the one-sentence summary, the model's output only included the prompt we were using to generate the summary (needless to say, it was automatically published on our website ;)) &#x200B; I hope this post will be useful. We are currently running a campaign on Product Hunt: https://www.producthunt.com/posts/domore-ai So, if you have any feedback for us or think what we're doing is cool, don't hesitate to support us :)

[D] Misuse of Deep Learning in Nature Journal’s Earthquake Aftershock Paper
reddit
LLM Vibe Score0
Human Vibe Score0.333
milaworldThis week

[D] Misuse of Deep Learning in Nature Journal’s Earthquake Aftershock Paper

Recently, I saw a post by Rajiv Shah, Chicago-based data-scientist, regarding an article published in Nature last year called Deep learning of aftershock patterns following large earthquakes, written by scientists at Harvard in collaboration with Google. Below is the article: Stand Up for Best Practices: Misuse of Deep Learning in Nature’s Earthquake Aftershock Paper The Dangers of Machine Learning Hype Practitioners of AI, machine learning, predictive modeling, and data science have grown enormously over the last few years. What was once a niche field defined by its blend of knowledge is becoming a rapidly growing profession. As the excitement around AI continues to grow, the new wave of ML augmentation, automation, and GUI tools will lead to even more growth in the number of people trying to build predictive models. But here’s the rub: While it becomes easier to use the tools of predictive modeling, predictive modeling knowledge is not yet a widespread commodity. Errors can be counterintuitive and subtle, and they can easily lead you to the wrong conclusions if you’re not careful. I’m a data scientist who works with dozens of expert data science teams for a living. In my day job, I see these teams striving to build high-quality models. The best teams work together to review their models to detect problems. There are many hard-to-detect-ways that lead to problematic models (say, by allowing target leakage into their training data). Identifying issues is not fun. This requires admitting that exciting results are “too good to be true” or that their methods were not the right approach. In other words, it’s less about the sexy data science hype that gets headlines and more about a rigorous scientific discipline. Bad Methods Create Bad Results Almost a year ago, I read an article in Nature that claimed unprecedented accuracy in predicting earthquake aftershocks by using deep learning. Reading the article, my internal radar became deeply suspicious of their results. Their methods simply didn’t carry many of the hallmarks of careful predicting modeling. I started to dig deeper. In the meantime, this article blew up and became widely recognized! It was even included in the release notes for Tensorflow as an example of what deep learning could do. However, in my digging, I found major flaws in the paper. Namely, data leakage which leads to unrealistic accuracy scores and a lack of attention to model selection (you don’t build a 6 layer neural network when a simpler model provides the same level of accuracy). To my earlier point: these are subtle, but incredibly basic predictive modeling errors that can invalidate the entire results of an experiment. Data scientists are trained to recognize and avoid these issues in their work. I assumed that this was simply overlooked by the author, so I contacted her and let her know so that she could improve her analysis. Although we had previously communicated, she did not respond to my email over concerns with the paper. Falling On Deaf Ears So, what was I to do? My coworkers told me to just tweet it and let it go, but I wanted to stand up for good modeling practices. I thought reason and best practices would prevail, so I started a 6-month process of writing up my results and shared them with Nature. Upon sharing my results, I received a note from Nature in January 2019 that despite serious concerns about data leakage and model selection that invalidate their experiment, they saw no need to correct the errors, because “Devries et al. are concerned primarily with using machine learning as [a] tool to extract insight into the natural world, and not with details of the algorithm design.” The authors provided a much harsher response. You can read the entire exchange on my github. It’s not enough to say that I was disappointed. This was a major paper (it’s Nature!) that bought into AI hype and published a paper despite it using flawed methods. Then, just this week, I ran across articles by Arnaud Mignan and Marco Broccardo on shortcomings that they found in the aftershocks article. Here are two more data scientists with expertise in earthquake analysis who also noticed flaws in the paper. I also have placed my analysis and reproducible code on github. Standing Up For Predictive Modeling Methods I want to make it clear: my goal is not to villainize the authors of the aftershocks paper. I don’t believe that they were malicious, and I think that they would argue their goal was to just show how machine learning could be applied to aftershocks. Devries is an accomplished earthquake scientist who wanted to use the latest methods for her field of study and found exciting results from it. But here’s the problem: their insights and results were based on fundamentally flawed methods. It’s not enough to say, “This isn’t a machine learning paper, it’s an earthquake paper.” If you use predictive modeling, then the quality of your results are determined by the quality of your modeling. Your work becomes data science work, and you are on the hook for your scientific rigor. There is a huge appetite for papers that use the latest technologies and approaches. It becomes very difficult to push back on these papers. But if we allow papers or projects with fundamental issues to advance, it hurts all of us. It undermines the field of predictive modeling. Please push back on bad data science. Report bad findings to papers. And if they don’t take action, go to twitter, post about it, share your results and make noise. This type of collective action worked to raise awareness of p-values and combat the epidemic of p-hacking. We need good machine learning practices if we want our field to continue to grow and maintain credibility. Link to Rajiv's Article Original Nature Publication (note: paywalled) GitHub repo contains an attempt to reproduce Nature's paper Confrontational correspondence with authors

[N] Last Week in AI News Digest - Automated chemical synthesis, using heartbeats to detect deepfakes, and more!
reddit
LLM Vibe Score0
Human Vibe Score1
regalalgorithmThis week

[N] Last Week in AI News Digest - Automated chemical synthesis, using heartbeats to detect deepfakes, and more!

Hi there, just sharing the latest edition of our AI news digest newsletter! We're just a couple of AI grad students doing this for fun, so hope the self promotion is not too annoying (also, welcome feedback). See it below, and feel free to subscribe. Mini Briefs Robotics, AI, and Cloud Computing Combine to Supercharge Chemical and Drug Synthesis IBM recently demoed a complex system for chemical testing and drug synthesis. The system has an AI component that predicts the results of chemical reactions, and a fully automated robotic experiment setup that runs chemical tests 24/7. Users can access the remote robotics lab online, and IBM can also install the system on-premise. With these tools working together, IBM is hoping to reduce typical drug discovery and verification time by half. AI researchers use heartbeat detection to identify deepfake videos Researchers from multiple groups are tackling the challenge of detecting deepfake videos by analyzing the apparent heartbeat of the people depicted in the video. This is possible, because a person’s blood flow changes their skin color ever so slightly, and this change is often detectable via a process called photoplethysmography (PPG). Because deepfakes are not currently optimizing to generate realisitic heartbeats, temporal or spatial anomalies in PPG signals allow resesarchers to detect deepfakes with a 97% accuracy. Advances & Business This AI Expert From Senegal Is Helping Showcase Africans In STEM \- Adji Bousso Dieng will be Princeton’s School of Engineering’s first Black female faculty. Google’s AI-powered flood alerts now cover all of India and parts of Bangladesh \- India, the world’s second most populated nation, sees more than 20% of the global flood-related fatalities each year as overrun riverbanks sweep tens of thousands of homes with them. Two years ago, Google volunteered to help. Finding magnetic eruptions in space with an AI assistant \- MMS look for explosive reconnection events as it flies through the magnetopause - the boundary region where Earth’s magnetic butts up against the solar wind that flows throughout the solar system. This know-it-all AI learns by reading the entire web nonstop \- Diffbot is building the biggest-ever knowledge graph by applying image recognition and natural-language processing to billions of web pages. Bosch and Ford will test autonomous parking in Detroit \- Ford, Bosch, and Dan Gilbert’s real estate firm Bedrock today detailed an autonomous parking pilot scheduled to launch in September at The Assembly, a mixed-used building in Detroit’s Corktown neighborhood. Create your own moody quarantine music with Google’s AI \- Lo-Fi Player, the latest project out of Google Magenta, lets you mix tunes with the help of machine learning by interacting with a virtual room. Apple launches AI/ML residency program to attract niche experts \- As Apple’s artificial language and machine learning initiatives continue to expand, its interest in attracting talent has grown - a theme that’s barely under the surface of the company’s occasionally updated Machine Learning Research blog. Dusty Robotics CEO Tessa Lau Discusses Robotics Start-Ups and Autonomous Robots for Construction \- Tessa Lau is Founder/CEO at Dusty Robotics, whose mission is to increase construction industry productivity by introducing robotic automation on the jobsite. Concerns & Hype Google Offers to Help Others With the Tricky Ethics of AI \- Companies pay cloud computing providers like Amazon, Microsoft, and Google big money to avoid operating their own digital infrastructure. The Peace Dividends Of The Autonomous Vehicle Wars \- The rapid growth of the mobile market in the late 2000s and early 2010s led to a burst of technological progress. Ethics must be part of the development process’ \- The increasing use of AI (artificial intelligence) in the development of new medical technologies demands greater attention to ethical aspects. Analysis & Policy China’s new AI trade rules could hamper a TikTok sale \- TikTok’s attempt to sell itself and avert a possible US ban may run into some complications. The Wall Street Journal reports that China has unveiled new restrictions on AI technology exports that could affect TikTok. Podcast Check out our weekly podcast covering these stories! Website | RSS | iTunes | Spotify | YouTube

[D] What is your honest experience with reinforcement learning?
reddit
LLM Vibe Score0
Human Vibe Score1
Starks-TechnologyThis week

[D] What is your honest experience with reinforcement learning?

In my personal experience, SOTA RL algorithms simply don't work. I've tried working with reinforcement learning for over 5 years. I remember when Alpha Go defeated the world famous Go player, Lee Sedol, and everybody thought RL would take the ML community by storm. Yet, outside of toy problems, I've personally never found a practical use-case of RL. What is your experience with it? Aside from Ad recommendation systems and RLHF, are there legitimate use-cases of RL? Or, was it all hype? Edit: I know a lot about AI. I built NexusTrade, an AI-Powered automated investing tool that lets non-technical users create, update, and deploy their trading strategies. I’m not an idiot nor a noob; RL is just ridiculously hard. Edit 2: Since my comments are being downvoted, here is a link to my article that better describes my position. It's not that I don't understand RL. I released my open-source code and wrote a paper on it. It's the fact that it's EXTREMELY difficult to understand. Other deep learning algorithms like CNNs (including ResNets), RNNs (including GRUs and LSTMs), Transformers, and GANs are not hard to understand. These algorithms work and have practical use-cases outside of the lab. Traditional SOTA RL algorithms like PPO, DDPG, and TD3 are just very hard. You need to do a bunch of research to even implement a toy problem. In contrast, the decision transformer is something anybody can implement, and it seems to match or surpass the SOTA. You don't need two networks battling each other. You don't have to go through hell to debug your network. It just naturally learns the best set of actions in an auto-regressive manner. I also didn't mean to come off as arrogant or imply that RL is not worth learning. I just haven't seen any real-world, practical use-cases of it. I simply wanted to start a discussion, not claim that I know everything. Edit 3: There's a shockingly number of people calling me an idiot for not fully understanding RL. You guys are wayyy too comfortable calling people you disagree with names. News-flash, not everybody has a PhD in ML. My undergraduate degree is in biology. I self-taught myself the high-level maths to understand ML. I'm very passionate about the field; I just have VERY disappointing experiences with RL. Funny enough, there are very few people refuting my actual points. To summarize: Lack of real-world applications Extremely complex and inaccessible to 99% of the population Much harder than traditional DL algorithms like CNNs, RNNs, and GANs Sample inefficiency and instability Difficult to debug Better alternatives, such as the Decision Transformer Are these not legitimate criticisms? Is the purpose of this sub not to have discussions related to Machine Learning? To the few commenters that aren't calling me an idiot...thank you! Remember, it costs you nothing to be nice! Edit 4: Lots of people seem to agree that RL is over-hyped. Unfortunately those comments are downvoted. To clear up some things: We've invested HEAVILY into reinforcement learning. All we got from this investment is a robot that can be super-human at (some) video games. AlphaFold did not use any reinforcement learning. SpaceX doesn't either. I concede that it can be useful for robotics, but still argue that it's use-cases outside the lab are extremely limited. If you're stumbling on this thread and curious about an RL alternative, check out the Decision Transformer. It can be used in any situation that a traditional RL algorithm can be used. Final Edit: To those who contributed more recently, thank you for the thoughtful discussion! From what I learned, model-based models like Dreamer and IRIS MIGHT have a future. But everybody who has actually used model-free models like DDPG unanimously agree that they suck and don’t work.

[R] AutoDev: Automated AI-Driven Development - Microsoft 2024
reddit
LLM Vibe Score0
Human Vibe Score0
Singularian2501This week

[R] AutoDev: Automated AI-Driven Development - Microsoft 2024

Paper: https://arxiv.org/abs/2403.08299 Sorry posted a wrong github link. The real code sadly isnt public yet! Thank you for everyone who pointed that out to me! ~~Github includes Code + AutoDev Coder Model:~~ ~~https://github.com/unit-mesh/auto-dev~~ Abstract: The landscape of software development has witnessed a paradigm shift with the advent of AI-powered assistants, exemplified by GitHub Copilot. However, existing solutions are not leveraging all the potential capabilities available in an IDE such as building, testing, executing code, git operations, etc. Therefore, they are constrained by their limited capabilities, primarily focusing on suggesting code snippets and file manipulation within a chat-based interface. To fill this gap, we present AutoDev, a fully automated AI-driven software development framework, designed for autonomous planning and execution of intricate software engineering tasks. AutoDev enables users to define complex software engineering objectives, which are assigned to AutoDev's autonomous AI Agents to achieve. These AI agents can perform diverse operations on a codebase, including file editing, retrieval, build processes, execution, testing, and git operations. They also have access to files, compiler output, build and testing logs, static analysis tools, and more. This enables the AI Agents to execute tasks in a fully automated manner with a comprehensive understanding of the contextual information required. Furthermore, AutoDev establishes a secure development environment by confining all operations within Docker containers. This framework incorporates guardrails to ensure user privacy and file security, allowing users to define specific permitted or restricted commands and operations within AutoDev. In our evaluation, we tested AutoDev on the HumanEval dataset, obtaining promising results with 91.5% and 87.8% of Pass@1 for code generation and test generation respectively, demonstrating its effectiveness in automating software engineering tasks while maintaining a secure and user-controlled development environment. https://preview.redd.it/5nxqajnvbkoc1.jpg?width=924&format=pjpg&auto=webp&s=8343c5fb33d2914bbfbf2dd9c164b5970b9743ab https://preview.redd.it/z5fkkjnvbkoc1.jpg?width=1364&format=pjpg&auto=webp&s=bc434ff384d2ed67ea0382dbbb68b9a90313cd44

Have You Used AI Tools for Your Research? Which Ones Are Your Favorite and Why?
reddit
LLM Vibe Score0
Human Vibe Score1
somdipdeyThis week

Have You Used AI Tools for Your Research? Which Ones Are Your Favorite and Why?

Over a decade ago, I wrote two articles: "A B\ginner’s Guide to Computer Science Research" and "How to Start a Research Work in Computer Science"*. These articles were widely used in universities worldwide to help students and early-career researchers navigate academic research in Computer Science (CS). Fast forward to 2025, the research landscape has evolved significantly, especially in AI and CS, with the advent of AI-powered research tools, open-access repositories, and real-time collaboration platforms. These tools have made research more accessible, enabling students and professionals to work more efficiently while focusing on real innovation. I recently published an updated article in The Times of India, presenting an Eight-Step Approach to Research framework designed for modern AI and CS research. This framework integrates AI-powered literature review tools, reference management systems, open science platforms, and collaborative research methods to enhance the research workflow. 🚀 Would love to hear from the ML research community: 1️⃣ Have you used any AI-powered tools or automation techniques in your research? Which ones do you find most useful? 2️⃣ Do you have recommendations for other AI tools that weren’t covered in the article but could benefit researchers? 3️⃣ How do you think AI will shape the future of academic research and discovery? 📖 Read the article here: How to Start Research in Computer Science & AI in 2025 – An Updated Framework Block Diagram of “Eight-Step Approach to Research” in 2025 Let’s discuss! What are your go-to tools for making research more efficient in 2025?

Is being a solopreneur really that fatal?
reddit
LLM Vibe Score0
Human Vibe Score1
Upbeat_Challenge5460This week

Is being a solopreneur really that fatal?

Okay, so I need to get something off my chest... People love to say that solopreneurship is a death sentence. That if you can’t find a cofounder, you’ll never build a team, never scale, never succeed. But I wonder about the other side of the coin—something that, browsing here and in other subs, doesn’t seem to get nearly as much attention—how fatal cofounder conflicts can be. I’ve personally seen three startups fail before even getting to an MVP because of cofounder issues. One of them was a company I was briefly a cofounder for. The other two are startups coworkers were previous cofounders for that fell apart before they even got to an MVP. In each case, it wasn’t lack of funding or product-market fit that killed them—it was the people. Yet, somehow, the startup world keeps pushing the idea that finding a cofounder is the most important thing you can do. But here’s the thing: if you can’t find a cofounder, that doesn’t mean you can’t build a business. It doesn’t even mean you can’t build a team. With the tools available today (no-code, AI, fractional hiring), a single person can get an MVP off the ground, validate demand, and take those first steps without needing to rush into a partnership with someone they barely know. And also—I wonder how many people actually succeed with a cofounder they met casually at a networking event or online? People talk about the risks of going solo, but not enough about the risks of tying your company’s future to someone you just met. (If you’re going to have a cofounder, IMO it should be someone you trust deeply, someone whose skills and working style you know complement yours—not just someone you brought on because startup X/YouTube told you to.). At the end of the day, I honestly think it’s about the product. If you can build something valuable and find market fit—whether solo or with a team—you’ll have the leverage to hire, partner, and grow. That’s what actually matters. That said—I know how incredibly hard it is to be a solopreneur—and not to have someone along the journey with you who can take half of the emotional and psychological burden, in addition to the actual work... What do you think? Any thoughts here appreciated.

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company
reddit
LLM Vibe Score0
Human Vibe Score0.778
wutangsamThis week

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company

I’ve learned so much over the years from this subreddit. I thought I’d return the favour and share some of my own learnings. In November 2020 my best friend and I had an idea. “What if we could find out which stocks the Internet is talking about?” This formed the origins of Ticker Nerd. 9 months later we sold Ticker Nerd to Finder (an Australian fintech company valued at around $500m). In this post, I am going to lay out how we got there. How we came up with the idea First off, like other posts have covered - you don’t NEED a revolutionary or original idea to build a business. There are tonnes of “boring” businesses making over 7 figures a year e.g. law firms, marketing agencies, real estate companies etc. If you’re looking for an exact formula to come up with a great business idea I’m sorry, but it doesn’t exist. Finding new business opportunities is more of an art than a science. Although, there are ways you can make it easier to find inspiration. Below are the same resources I use for inspiration. I rarely ever come up with ideas without first searching one of the resources below for inspiration: Starter Story Twitter Startup Ideas My First Million Trends by the Hustle Trends VC To show how you how messy, random and unpredictable it can be to find an idea - let me explain how my co-founder and I came up with the idea for Ticker Nerd: We discovered a new product on Twitter called Exploding Topics. It was a newsletter that uses a bunch of software and algorithms to find trends that are growing quickly before they hit the mainstream. I had recently listened to a podcast episode from My First Million where they spoke about Motley Fool making hundreds of millions from their investment newsletters. We asked ourselves what if we could build a SaaS platform similar to Exploding Topics but it focused on stocks? We built a quick landing page using Carrd + Gumroad that explained what our new idea will do and included a payment option to get early access for $49. We called it Exploding Stock (lol). We shared it around a bunch of Facebook groups and subreddits. We made $1,000 in pre-sales within a couple days. My co-founder and I can’t code so we had to find a developer to build our idea. We interviewed a bunch of potential candidates. Meanwhile, I was trawling through Wall Street Bets and found a bunch of free tools that did roughly what we wanted to build. Instead of building another SaaS tool that did the same thing as these free tools we decided to pivot from our original idea. Our new idea = a paid newsletter that sends a weekly report that summarises 2 of the best stocks that are growing in interest on the Internet. We emailed everyone who pre-ordered access, telling them about the change and offered a full refund if they wanted. tl;dr: We essentially combined two existing businesses (Exploding Topics and Motley Fool) and made it way better. We validated the idea by finding out if people will actually pay money for it BEFORE we decided to build it. The idea we started out with changed over time. How to work out if your idea will actually make money It’s easy to get hung up on designing the logo or choosing the perfect domain name for your new idea. At this stage none of that matters. The most important thing is working out if people will pay money for it. This is where validation comes in. We usually validate ideas using Carrd. It lets you build a simple one page site without having to code. The Ticker Nerd site was actually built using a Carrd template. Here’s how you can do it yourself (at a high level): Create a Carrd pro account (yes it's a $49 one off payment but you’ll get way more value out of it). Buy a cheap template and send it to your Carrd account. You can build your own template but this will save you a lot of time. Once the template reaches your Carrd account, duplicate it. Leave the original so it can be duplicated for other ideas. Jump onto Canva (free) and create a logo using the free logos provided. Import your logo. Add copy to the page that explains your idea. Use the AIDA formula. Sign up to Gumroad (free) and create a pre-sale campaign. Create a discounted lifetime subscription or version of the product. This will be used pre-sales. Add the copy from the site into the pre-sale campaign on Gumroad. Add a ‘widget’ to Carrd and connect it to Gumroad using the existing easy integration feature. Purchase a domain name. Connect it to Carrd. Test the site works. Share your website Now the site is ready you can start promoting it in various places to see how the market reacts. An easy method is to find relevant subreddits using Anvaka (Github tool) or Subreddit Stats. The Anvaka tool provides a spider map of all the connected subreddits that users are active in. The highlighted ones are most relevant. You can post a thread in these subreddits that offer value or can generate discussion. For example: ‘I’m creating a tool that can write all your copy, would anyone actually use this?’ ‘What does everything think of using AI to get our copy written faster?’ ‘It’s time to scratch my own itch, I’m creating a tool that writes marketing copy using GPT-3. What are the biggest problems you face writing marketing copy? I’ll build a solution for it’ Reddit is pretty brutal these days so make sure the post is genuine and only drop your link in the comments or in the post if it seems natural. If people are interested they’ll ask for the link. Another great place to post is r/entrepreuerridealong and r/business_ideas. These subreddits expect people to share their ideas and you’ll likely make some sales straight off the bat. I also suggest posting in some Facebook groups (related to your idea) as well just for good measure. Assess the results If people are paying you for early access you can assume that it’s worth building your idea. The beauty of posting your idea on Reddit or in Facebook groups is you’ll quickly learn why people love/hate your idea. This can help you decide how to tweak the idea or if you should drop it and move on to the next one. How we got our first 100 customers (for free) By validating Ticker Nerd using subreddits and Facebook groups this gave us our first paying customers. But we knew this wouldn’t be sustainable. We sat down and brainstormed every organic strategy we could use to get traction as quickly as possible. The winner: a Product Hunt launch. A successful Product Hunt launch isn’t easy. You need: Someone that has a solid reputation and audience to “hunt” your product (essentially an endorsement). An aged Product Hunt account - you can’t post any products if your account is less than a week old. To be following relevant Product Hunt members - since they get notified when you launch a new product if they’re following you. Relationships with other builders and makers on Product Hunt that also have a solid reputation and following. Although, if you can pull it off you can get your idea in front of tens of thousands of people actively looking for new products. Over the next few weeks, I worked with my co-founder on connecting with different founders, indie hackers and entrepreneurs mainly via Twitter. We explained to them our plans for the Product Hunt launch and managed to get a small army of people ready to upvote our product on launch day. We were both nervous on the day of the launch. We told ourselves to have zero expectations. The worst that could happen was no one signed up and we were in the same position as we’re in now. Luckily, within a couple of hours Ticker Nerd was on the homepage of Product Hunt and in the top 10. The results were instant. After 24 hours we had around 200 people enter their payment details to sign up for our free trial. These signups were equal to around $5,800 in monthly recurring revenue. \-- I hope this post was useful! Drop any questions you have below and I’ll do my best to respond :)

tools I use to not have to hire anyone
reddit
LLM Vibe Score0
Human Vibe Score1
Pio_SceThis week

tools I use to not have to hire anyone

I’ve spent unreasonable amount of time with AI tools and here’s curated list of ones I recommend for productivity (honestly, some of them can replace an employee): General assistants ChatGPT \- You probably know it. It’s a great tool for ideating, brainstorming, document summarization and quick question-answer work. There’s a desktop app available so you can quickly pop it up by pressing control + space, which makes it even better for productivity. Claude \- Another chat interface, similar to ChatGPT. It’s a different model provider so the answers and behavior might be different. From my experience, Claude 3.5 Sonnet is performing better than GPT-4o (but not o1) in tasks that focus on reasoning, code writing and copywriting. There’s also a desktop app available. Gemini \- Honestly, I’m not even sure where to put it. It’s Google’s model, one of the most powerful in terms of multimodal capabilities (text, image, audio). And it’s tailored for your Google Workspace. Email, docs, spreadsheets, meets, presentation. Anything. Research Perplexity \- Perplexity is an AI search engine that provides answers to questions with up-to-date information. So, forget Google. Use Perplexity to get answers to questions and dive down the rabbit hole. Exa AI \- Exa is another advanced search engine that combines AI-driven neural search with traditional keyword search. It understands the semantic meaning of queries and documents. And you can also choose what you want to search: academic articles, news, reports, tweets etc. Meetings, calendar and email Granola \- Great AI notepad for meetings. It’s a desktop app, so there’s no bot joining your meetings. It automatically transcribes and enhances meeting notes, helping organize and summarize key takeaways and generates action items, follow-up emails, etc. It also allows you to ask questions about the transcript and get answers. Reclaim \- AI-powered calendar that optimizes for productivity. Essentially, it automates meetings, tracks tasks, and protects deep work time. Cool thing is that it syncs with Google Calendar and Slack. Cora \- Batch processing emails is one of the main productivity tactics. Cora enables that. You only see emails that you need to respond to. And it generates automatic replies for you. All other emails are summarized twice a day. Knowledge summarization Particle News \- Short summaries of the daily news. Pretty straightforward. Notebook LM \- Notebook LM helps process and summarize various types of content, such as PDFs, websites, videos, and more. The cool thing is that it provides insights and connections between topics, cites sources and offers audio summaries. I use it when the content to read is too long and I’m on the go. Napkin \- For creating visuals from text. You can easily generate and customize infographics, diagrams etc. So, if you’re brainstorming, writing or preparing for a presentation, Napkin will work well. Writing and brainstorming Grammarly \- Well known grammar checker. It helps improve writing by focusing on clarity and tone. Sometimes the Grammarly icon popping up is annoying though. Flow \- Flow helps you write and edit notes by speaking. And it integrates across all the apps you use, adapts to your tone and style. Cool tool for just yapping! Automations Gumloop \- Think AI-first Zapier, but 100x more powerful. It's is a platform for automating complex work using AI via a no-code drag and drop interface. It’s very easy to automate work without needing engineers. And they have loads of templates. Wordware \- A platform for building AI agents with natural language. Honestly, for folks who are a bit more technical. You simply prompt LLM to perform a task for you. And you can build any integration you want. If you’re a builder, you can later on connect the agent via API. I strongly believe that technology is leverage. And with AI we can be in top 0.1% of people. If you want bit deeper dive into the topic, I shared that on my substack (available via link in my profile) Any other recommendations for apps I could use? What works if you want to keep the team super lean in early days?

Built a Free AI Fitness Planner - From Passion to Product with No Traditional Coding
reddit
LLM Vibe Score0
Human Vibe Score1
jhojnac2This week

Built a Free AI Fitness Planner - From Passion to Product with No Traditional Coding

I wanted to share my journey of creating a free ai-powered workout planning tool with bolt. new and very minimal coding skills. It has taken me probably 4 days in total to complete and get to a point I am happy with. Many improvements coming but want to get it out there for some feedback and testing. I have been going to the gym for years and at this point my routines have gotten stale. I end up doing the same sets of exercises and repetitions over and over. I figured why not let chat gpt or some AI software help me develop or at least recommend different exercises. I was then was recommended youtube videos on creating your own web application without any coding. I will say it does take some coding knowledge, not that I am editing it myself, but I know what its trying to do and can prompt it correctly. I am still struggling with some things like integrating stripe for subscriptions so I only have it set up for donations currently. I dont mind it being free as I would like everyone the opportunity to help develop their own workouts. current cost breakdown to create: bolt. new credits - $100/month (gonna drop to the $20 now that its complete) supabase database - $35/month netlify domain - $11.99/year If anyone is interested or has questions feel free to let me know. It is called fitfocuscalendar. com Edit: title and 1st sentence came from AI everything else was typed by me.

If only someone told me this before my first startup
reddit
LLM Vibe Score0
Human Vibe Score0.625
johnrushxThis week

If only someone told me this before my first startup

If only someone told me this before my first startup: Validate idea first. I wasted a decade building stuff nobody needed. Incubators and VCs served to me as a validation, but I was so wrong. Kill my EGO. It’s not about me, but the user. I must want what the user wants, not what I want. My taste isn't important. The user has expectations, and I must fulfill them. Don’t chaise investors. Chase users, and then investors will be chasing me. I've never had more incoming interest from VC than now when I'm the least interested in them. Never hire managers. Only hire doers until PMF. So many people know how to manage people and so few can actually get sh\*t done barehand. Landing page is the least important thing in a startup. Pick a simple template, edit texts with a no-code website builder in less than an hour and that's it! At the early stage, I win traffic outside of my website, people are already interested, so don't make them search for the signup button among the texts! Focus on conversion optimization only when the traffic is consistent. Keep it to one page. Nobody gonna browse this website. Hire only fullstack devs. There is nothing less productive in this world than a team of developers for an early-stage product. One full stack dev building the whole product. That’s it. Chase global market from day 1. If the product and marketing are good, it will work on the global market too, if it’s bad, it won’t work on the local market too. So better go global from day 1, so that if it works, the upside is 100x bigger. I launched all startups for the Norwegian market, hoping we will scale to international at some point. I wish I launched to international from day 1 as I do now. The size of the market is 10000x bigger. I can validate and grow products in days, not in years as it used to be. Do SEO from day 2. As early as I can. I ignored this for 14 years. It’s my biggest regret. It takes just 5 minutes to get it done on my landing page. I go to Google Keyword Planner, enter a few keywords around my product, sort them by traffic, filter out high competition kws, pick the top 10, and place them natively on my home page and meta tags. Add one blog article every week. Either manually or by paying for an AI blogging tool. Sell features, before building them. Ask existing users if they want this feature. I run DMs with 10-20 users every day, where I chat about all my ideas and features I wanna add. I clearly see what resonates with me most and only go build those. If I don't have followers, try HN, Reddit, or just search on X for posts and ask it in the replies. People are helpful, they will reply if the question is easy to understand. Hire only people I would wanna hug. My cofounder, an old Danish man said this to me in 2015. And it was a big shift. I realized that if I don’t wanna hug the person, it means I dislike them on a chemical/animal level. Even if I can’t say why, but that’s the fact. Sooner or later, we would have a conflict and eventually break up. It takes up to 10 years to build a startup, make sure I do it with people I have this connection with. Invest all money into my startups and friends. Not crypt0, not stockmarket, not properties. I did some math, if I kept investing all my money into all my friends’ startups, that would be about 70 investments. 3 of them turned into unicorns eventually. Even 1 would have made the bank. Since 2022, I have invested all my money into my products, friends, and network. If I don't have friends who do startups, invest it in myself. Post on Twitter daily. I started posting here in March last year. It’s my primary source of new connections and growth. I could have started it earlier, I don't know why I didn't. Don’t work/partner with corporates. Corporations always seem like an amazing opportunity. They’re big and rich, they promise huge stuff, millions of users, etc. But every single time none of this happens. Because I talk to a regular employees there. They waste my time, destroy focus, shift priorities, and eventually bring in no users/money. Don’t get ever distracted by hype e.g. crypt0. I lost 1.5 years of my life this way. I met the worst people along the way. Fricks, scammers, thieves. Some of my close friends turned into thieves along the way, just because it was so common in that space. I wish this didn’t happen to me. I wish I was stronger and stayed on my mission. Don’t build consumer apps. Only b2b. Consumer apps are so hard, like a lottery. It’s just 0.00001% who make it big. The rest don’t. Even if I got many users, then there is a monetization challenge. I’ve spent 4 years in consumer apps and regret it. Don’t hold on bad project for too long, max 1 year. Some projects just don’t work. In most cases, it’s either the idea that’s so wrong that I can’t even pivot it or it’s a team that is good one by one but can’t make it as a team. Don’t drag this out for years. Tech conferences are a waste of time. They cost money, take energy, and time and I never really meet anyone there. Most people there are the “good” employees of corporations who were sent there as a perk for being loyal to the corporation. Very few fellow makers. Scrum is a Scam. For small teams and bootstrapped teams. If I had a team that had to be nagged every morning with questions as if they were children in kindergarten, then things would eventually fail. The only good stuff I managed to do happened with people who were grownups and could manage their stuff on their own. We would just do everything over chat as a sync on goals and plans. Outsource nothing at all until PMF. In a startup, almost everything needs to be done in a slightly different way, more creative, and more integrated into the vision. When outsourcing, the external members get no love and no case for the product. It’s just yet another assignment in their boring job. Instead of coming up with great ideas for my project they will be just focusing on ramping up their skills to get a promotion or a better job offer. Bootstrap. I spent way too much time raising money. I raised more than 10 times, preseed, seeded, and series A. But each time it was a 3-9 month project, meetings every week, and lots of destruction. I could afford to bootstrap, but I still went the VC-funded way, I don’t know why. To be honest, I didn’t know bootstrapping was a thing I could do or anyone does. It may take a decade. When I was 20, I was convinced it takes a few years to build and succeed with a startup. So I kept pushing my plans forward, to do it once I exited. Family, kids. I wish I married earlier. I wish I had kids earlier. No Free Tier. I'd launch a tool with a free tier, and it'd get sign-ups, but very few would convert. I'd treat free sign-ups as KPIs and run on it for years. I'd brag about signups and visitors. I'd even raise VC money with these stats. But eventually, I would fail to reach PMF. Because my main feedback would come from free users and the product turned into a perfect free product. Once I switched to "paid only" until I validated the product, things went really well. Free and paid users often need different products. Don't fall into this trap as I did. Being To Cheap. I always started by checking all competitors and setting the lowest price. I thought this would be one of the key advantages of my product. But no, I was wrong. The audience on $5 and $50 are totally different. $5: pain in the \*ss, never happy, never recommend me to a friend, leave in 4 months. $50: polite, give genuine feedback, happy, share with friends, become my big fan if I solve their request. I will fail. When I started my first startup. I thought if I did everything right, it would work out. But it turned out that almost every startup fails. I wish I knew that and I tried to fail faster, to get to the second iteration, then to the third, and keep going on, until I either find out nothing works or make it work. Use boilerplates. I wasted years of dev time and millions of VC money to pay for basic things. To build yet another sidebar, yet another dashboard, and payment integration... I had too much pride, I couldn't see myself taking someone else code as a basis for my product. I wanted it to be 100% mine, original, from scratch. Because my product seems special to me. Spend more time with Family & Friends. I missed the weddings of all my best friends and family. I was so busy. I thought if I didn't do it on time, the world would end. Looking back today, it was so wrong. I meet my friends and can't share those memories with them, which makes me very sad. I realized now, that spending 10% of my time with family and friends would practically make no negative impact on my startups. Build Products For Audiences I Love. I never thought of this. I'd often build products either for corporates, consumers, or for developers. It turns out I have no love for all 3. But I deeply love indie founders. Because they are risk-takers and partly kids in their hearts. Once I switched the focus to indie makers on my products, my level of joy increased by 100x for me. Ignore Badges and Awards I was chasing those awards just like everyone else. Going to ceremonies, signing up for events and stuff. I've won tons of awards, but none of those were eventually useful to my business. I better focused on my business and users. Write Every Single Day. When I was a kid, I loved writing stories. In school, they would give an assignment, and I'd often write a long story for it, however, the teacher would put an F on it. The reason was simple, I had an issue with the direction of the letters and the sequence of letters in the words. I still have it, it's just the Grammarly app helping me to correct these issues. So the teacher would fail my stories because almost every sentence had a spelling mistake that I couldn't even see. It made me think I'm made at writing. So I stopped, for 15 years. But I kept telling stories all these years. Recently I realized that in any group, the setup ends up turning into me telling stories to everyone. So I tried it all again, here on X 10 months ago. I love it, the process, the feedback from people. I write every day. I wish I had done it all these years. The End. \ this is an updated version of my post on the same topic from 2 months ago. I've edited some of the points and added 9 new ones.* \\ This is not advice, it's my self-reflection that might help you avoid same mistakes if you think those were mistakes

I run an AI automation agency (AAA). My honest overview and review of this new business model
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

I run an AI automation agency (AAA). My honest overview and review of this new business model

I started an AI tools directory in February, and then branched off that to start an AI automation agency (AAA) in June. So far I've come across a lot of unsustainable "ideas" to make money with AI, but at the same time a few diamonds in the rough that aren't fully tapped into yet- especially the AAA model. Thought I'd share this post to shine light into this new business model and share some ways you could potentially start your own agency, or at the very least know who you are dealing with and how to pick and choose when you (inevitably) get bombarded with cold emails from them down the line. Foreword Running an AAA does NOT involve using AI tools directly to generate and sell content directly. That ship has sailed, and unless you are happy with $5 from Fiverr every month or so, it is not a real business model. Cry me a river but generating generic art with AI and slapping it onto a T-shirt to sell on Etsy won't make you a dime. At the same time, the AAA model will NOT require you to have a deep theoretical knowledge of AI, or any academic degree, as we are more so dealing with the practical applications of generative AI and how we can implement these into different workflows and tech-stacks, rather than building AI models from the ground up. Regardless of all that, common sense and a willingness to learn will help (a shit ton), as with anything. Keep in mind - this WILL involve work and motivation as well. The mindset that AI somehow means everything can be done for you on autopilot is not the right way to approach things. The common theme of businesses I've seen who have successfully implemented AI into their operations is the willingess to work with AI in a way that augments their existing operations, rather than flat out replace a worker or team. And this is exactly the train of thought you need when working with AI as a business model. However, as the field is relatively unsaturated and hype surrounding AI is still fresh for enterprises, right now is the prime time to start something new if generative AI interests you at all. With that being said, I'll be going over three of the most successful AI-adjacent businesses I've seen over this past year, in addition to some tips and resources to point you in the right direction. so.. WTF is an AI Automation Agency? The AI automation agency (or as some YouTubers have coined it, the AAA model) at its core involves creating custom AI solutions for businesses. I have over 1500 AI tools listed in my directory, however the feedback I've received from some enterprise users is that ready-made SaaS tools are too generic to meet their specific needs. Combine this with the fact virtually no smaller companies have the time or skills required to develop custom solutions right off the bat, and you have yourself real demand. I would say in practice, the AAA model is quite similar to Wordpress and even web dev agencies, with the major difference being all solutions you develop will incorporate key aspects of AI AND automation. Which brings me to my second point- JUST AI IS NOT ENOUGH. Rather than reducing the amount of time required to complete certain tasks, I've seen many AI agencies make the mistake of recommending and (trying to) sell solutions that more likely than not increase the workload of their clients. For example, if you were to make an internal tool that has AI answer questions based on their knowledge base, but this knowledge base has to be updated manually, this is creating unnecessary work. As such I think one of the key components of building successful AI solutions is incorporating the new (Generative AI/LLMs) with the old (programmtic automation- think Zapier, APIs, etc.). Finally, for this business model to be successful, ideally you should target a niche in which you have already worked and understand pain points and needs. Not only does this make it much easier to get calls booked with prospects, the solutions you build will have much greater value to your clients (meaning you get paid more). A mistake I've seen many AAA operators make (and I blame this on the "Get Rich Quick" YouTubers) is focusing too much on a specific productized service, rather than really understanding the needs of businesses. The former is much done via a SaaS model, but when going the agency route the only thing that makes sense is building custom solutions. This is why I always take a consultant-first approach. You can only build once you understand what they actually need and how certain solutions may impact their operations, workflows, and bottom-line. Basics of How to Get Started Pick a niche. As I mentioned previously, preferably one that you've worked in before. Niches I know of that are actively being bombarded with cold emails include real estate, e-commerce, auto-dealerships, lawyers, and medical offices. There is a reason for this, but I will tell you straight up this business model works well if you target any white-collar service business (internal tools approach) or high volume businesses (customer facing tools approach). Setup your toolbox. If you wanted to start a pressure washing business, you would need a pressure-washer. This is no different. For those without programming knowledge, I've seen two common ways AAA get setup to build- one is having a network of on-call web developers, whether its personal contacts or simply going to Upwork or any talent sourcing agency. The second is having an arsenal of no-code tools. I'll get to this more in a second, but this works beecause at its core, when we are dealing with the practical applications of AI, the code is quite simple, simply put. Start cold sales. Unless you have a network already, this is not a step you can skip. You've already picked a niche, so all you have to do is find the right message. Keep cold emails short, sweet, but enticing- and it will help a lot if you did step 1 correctly and intimately understand who your audience is. I'll be touching base later about how you can leverage AI yourself to help you with outreach and closing. The beauty of gen AI and the AAA model You don't need to be a seasoned web developer to make this business model work. The large majority of solutions that SME clients want is best done using an API for an LLM for the actual AI aspect. The value we create with the solutions we build comes with the conceptual framework and design that not only does what they need it to but integrates smoothly with their existing tech-stack and workflow. The actual implementation is quite straightforward once you understand the high level design and know which tools you are going to use. To give you a sense, even if you plan to build out these apps yourself (say in Python) the large majority of the nitty gritty technical work has already been done for you, especially if you leverage Python libraries and packages that offer high level abstraction for LLM-related functions. For instance, calling GPT can be as little as a single line of code. (And there are no-code tools where these functions are simply an icon on a GUI). Aside from understanding the capabilities and limitations of these tools and frameworks, the only thing that matters is being able to put them in a way that makes sense for what you want to build. Which is why outsourcing and no-code tools both work in our case. Okay... but how TF am I suppposed to actually build out these solutions? Now the fun part. I highly recommend getting familiar with Langchain and LlamaIndex. Both are Python libraires that help a lot with the high-level LLM abstraction I mentioned previously. The two most important aspects include being able to integrate internal data sources/knowledge bases with LLMs, and have LLMs perform autonomous actions. The two most common methods respectively are RAG and output parsing. RAG (retrieval augmented Generation) If you've ever seen a tool that seemingly "trains" GPT on your own data, and wonder how it all works- well I have an answer from you. At a high level, the user query is first being fed to what's called a vector database to run vector search. Vector search basically lets you do semantic search where you are searching data based on meaning. The vector databases then retrieves the most relevant sections of text as it relates to the user query, and this text gets APPENDED to your GPT prompt to provide extra context to the AI. Further, with prompt engineering, you can limit GPT to only generate an answer if it can be found within this extra context, greatly limiting the chance of hallucination (this is where AI makes random shit up). Aside from vector databases, we can also implement RAG with other data sources and retrieval methods, for example SQL databses (via parsing the outputs of LLM's- more on this later). Autonomous Agents via Output Parsing A common need of clients has been having AI actually perform tasks, rather than simply spitting out text. For example, with autonomous agents, we can have an e-commerce chatbot do the work of a basic customer service rep (i.e. look into orders, refunds, shipping). At a high level, what's going on is that the response of the LLM is being used programmtically to determine which API to call. Keeping on with the e-commerce example, if I wanted a chatbot to check shipping status, I could have a LLM response within my app (not shown to the user) with a prompt that outputs a random hash or string, and programmatically I can determine which API call to make based on this hash/string. And using the same fundamental concept as with RAG, I can append the the API response to a final prompt that would spit out the answer for the user. How No Code Tools Can Fit In (With some example solutions you can build) With that being said, you don't necessarily need to do all of the above by coding yourself, with Python libraries or otherwise. However, I will say that having that high level overview will help IMMENSELY when it comes to using no-code tools to do the actual work for you. Regardless, here are a few common solutions you might build for clients as well as some no-code tools you can use to build them out. Ex. Solution 1: AI Chatbots for SMEs (Small and Medium Enterprises) This involves creating chatbots that handle user queries, lead gen, and so forth with AI, and will use the principles of RAG at heart. After getting the required data from your client (i.e. product catalogues, previous support tickets, FAQ, internal documentation), you upload this into your knowledge base and write a prompt that makes sense for your use case. One no-code tool that does this well is MyAskAI. The beauty of it especially for building external chatbots is the ability to quickly ingest entire websites into your knowledge base via a sitemap, and bulk uploading files. Essentially, they've covered the entire grunt work required to do this manually. Finally, you can create a inline or chat widget on your client's website with a few lines of HTML, or altneratively integrate it with a Slack/Teams chatbot (if you are going for an internal Q&A chatbot approach). Other tools you could use include Botpress and Voiceflow, however these are less for RAG and more for building out complete chatbot flows that may or may not incorporate LLMs. Both apps are essentially GUIs that eliminate the pain and tears and trying to implement complex flows manually, and both natively incoporate AI intents and a knowledge base feature. Ex. Solution 2: Internal Apps Similar to the first example, except we go beyond making just chatbots but tools such as report generation and really any sort of internal tool or automations that may incorporate LLM's. For instance, you can have a tool that automatically generates replies to inbound emails based on your client's knowledge base. Or an automation that does the same thing but for replies to Instagram comments. Another example could be a tool that generates a description and screeenshot based on a URL (useful for directory sites, made one for my own :P). Getting into more advanced implementations of LLMs, we can have tools that can generate entire drafts of reports (think 80+ pages), based not only on data from a knowledge base but also the writing style, format, and author voice of previous reports. One good tool to create content generation panels for your clients would be MindStudio. You can train LLM's via prompt engineering in a structured way with your own data to essentially fine tune them for whatever text you need it to generate. Furthermore, it has a GUI where you can dictate the entire AI flow. You can also upload data sources via multiple formats, including PDF, CSV, and Docx. For automations that require interactions between multiple apps, I recommend the OG zapier/make.com if you want a no-code solution. For instance, for the automatic email reply generator, I can have a trigger such that when an email is received, a custom AI reply is generated by MyAskAI, and finally a draft is created in my email client. Or, for an automation where I can create a social media posts on multiple platforms based on a RSS feed (news feed), I can implement this directly in Zapier with their native GPT action (see screenshot) As for more complex LLM flows that may require multiple layers of LLMs, data sources, and APIs working together to generate a single response i.e. a long form 100 page report, I would recommend tools such as Stack AI or Flowise (open-source alternative) to build these solutions out. Essentially, you get most of the functions and features of Python packages such as Langchain and LlamaIndex in a GUI. See screenshot for an example of a flow How the hell are you supposed to find clients? With all that being said, none of this matters if you can't find anyone to sell to. You will have to do cold sales, one way or the other, especially if you are brand new to the game. And what better way to sell your AI services than with AI itself? If we want to integrate AI into the cold outreach process, first we must identify what it's good at doing, and that's obviously writing a bunch of text, in a short amount of time. Similar to the solutions that an AAA can build for its clients, we can take advantage of the same principles in our own sales processes. How to do outreach Once you've identified your niche and their pain points/opportunities for automation, you want to craft a compelling message in which you can send via cold email and cold calls to get prospects booked on demos/consultations. I won't get into too much detail in terms of exactly how to write emails or calling scripts, as there are millions of resources to help with this, but I will tell you a few key points you want to keep in mind when doing outreach for your AAA. First, you want to keep in mind that many businesses are still hesitant about AI and may not understand what it really is or how it can benefit their operations. However, we can take advantage of how mass media has been reporting on AI this past year- at the very least people are AWARE that sooner or later they may have to implement AI into their businesses to stay competitive. We want to frame our message in a way that introduces generative AI as a technology that can have a direct, tangible, and positive impact on their business. Although it may be hard to quantify, I like to include estimates of man-hours saved or costs saved at least in my final proposals to prospects. Times are TOUGH right now, and money is expensive, so you need to have a compelling reason for businesses to get on board. Once you've gotten your messaging down, you will want to create a list of prospects to contact. Tools you can use to find prospects include Apollo.io, reply.io, zoominfo (expensive af), and Linkedin Sales Navigator. What specific job titles, etc. to target will depend on your niche but for smaller companies this will tend to be the owner. For white collar niches, i.e. law, the professional that will be directly benefiting from the tool (i.e. partners) may be better to contact. And for larger organizations you may want to target business improvement and digital transformation leads/directors- these are the people directly in charge of projects like what you may be proposing. Okay- so you have your message, and your list, and now all it comes down to is getting the good word out. I won't be going into the details of how to send these out, a quick Google search will give you hundreds of resources for cold outreach methods. However, personalization is key and beyond simple dynamic variables you want to make sure you can either personalize your email campaigns directly with AI (SmartWriter.ai is an example of a tool that can do this), or at the very least have the ability to import email messages programmatically. Alternatively, ask ChatGPT to make you a Python Script that can take in a list of emails, scrape info based on their linkedin URL or website, and all pass this onto a GPT prompt that specifies your messaging to generate an email. From there, send away. How tf do I close? Once you've got some prospects booked in on your meetings, you will need to close deals with them to turn them into clients. Call #1: Consultation Tying back to when I mentioned you want to take a consultant-first appraoch, you will want to listen closely to their goals and needs and understand their pain points. This would be the first call, and typically I would provide a high level overview of different solutions we could build to tacke these. It really helps to have a presentation available, so you can graphically demonstrate key points and key technologies. I like to use Plus AI for this, it's basically a Google Slides add-on that can generate slide decks for you. I copy and paste my default company messaging, add some key points for the presentation, and it comes out with pretty decent slides. Call #2: Demo The second call would involve a demo of one of these solutions, and typically I'll quickly prototype it with boilerplate code I already have, otherwise I'll cook something up in a no-code tool. If you have a niche where one type of solution is commonly demanded, it helps to have a general demo set up to be able to handle a larger volume of calls, so you aren't burning yourself out. I'll also elaborate on how the final product would look like in comparison to the demo. Call #3 and Beyond: Once the initial consultation and demo is complete, you will want to alleviate any remaining concerns from your prospects and work with them to reach a final work proposal. It's crucial you lay out exactly what you will be building (in writing) and ensure the prospect understands this. Furthermore, be clear and transparent with timelines and communication methods for the project. In terms of pricing, you want to take this from a value-based approach. The same solution may be worth a lot more to client A than client B. Furthermore, you can create "add-ons" such as monthly maintenance/upgrade packages, training sessions for employeees, and so forth, separate from the initial setup fee you would charge. How you can incorporate AI into marketing your businesses Beyond cold sales, I highly recommend creating a funnel to capture warm leads. For instance, I do this currently with my AI tools directory, which links directly to my AI agency and has consistent branding throughout. Warm leads are much more likely to close (and honestly, much nicer to deal with). However, even without an AI-related website, at the very least you will want to create a presence on social media and the web in general. As with any agency, you will want basic a professional presence. A professional virtual address helps, in addition to a Google Business Profile (GBP) and TrustPilot. a GBP (especially for local SEO) and Trustpilot page also helps improve the looks of your search results immensely. For GBP, I recommend using ProfilePro, which is a chrome extension you can use to automate SEO work for your GBP. Aside from SEO optimzied business descriptions based on your business, it can handle Q/A answers, responses, updates, and service descriptions based on local keywords. Privacy and Legal Concerns of the AAA Model Aside from typical concerns for agencies relating to service contracts, there are a few issues (especially when using no-code tools) that will need to be addressed to run a successful AAA. Most of these surround privacy concerns when working with proprietary data. In your terms with your client, you will want to clearly define hosting providers and any third party tools you will be using to build their solution, and a DPA with these third parties listed as subprocessors if necessary. In addition, you will want to implement best practices like redacting private information from data being used for building solutions. In terms of addressing concerns directly from clients, it helps if you host your solutions on their own servers (not possible with AI tools), and address the fact only ChatGPT queries in the web app, not OpenAI API calls, will be used to train OpenAI's models (as reported by mainstream media). The key here is to be open and transparent with your clients about ALL the tools you are using, where there data will be going, and make sure to get this all in writing. have fun, and keep an open mind Before I finish this post, I just want to reiterate the fact that this is NOT an easy way to make money. Running an AI agency will require hours and hours of dedication and work, and constantly rearranging your schedule to meet prospect and client needs. However, if you are looking for a new business to run, and have a knack for understanding business operations and are genuinely interested in the pracitcal applications of generative AI, then I say go for it. The time is ticking before AAA becomes the new dropshipping or SMMA, and I've a firm believer that those who set foot first and establish themselves in this field will come out top. And remember, while 100 thousand people may read this post, only 2 may actually take initiative and start.

Jinxed - $0 month after bragging about my first $10k month here. (PROGRESS UPDATE)
reddit
LLM Vibe Score0
Human Vibe Score1
swagamoneyThis week

Jinxed - $0 month after bragging about my first $10k month here. (PROGRESS UPDATE)

A month ago I made a post in this sub about my first $10k month. It went viral. And guess what - I didn't make another dollar since. Honestly, I shouldn't have made any money that first month also. Because I didn't have an offer. If you're familiar with Alex Hormozi you know that the offer is what makes or breaks a business. And I simply didn't have it. I managed to close my first clients just because I rode the AI hype train and managed to capture a couple of CEOs who were riding it too. Took whatever I could get for installment without thinking about the future. (It also helped that I wasn't bullshitting and had a legit enterprise-grade custom GPT framework ready). But that's not a business strategy at all. You can't base your business solely off hype. So the last month was dedicated to crafting a proper offer. No selling involved. Purely discovery chats with as many people as possible. The viral post helped because I connected with some badass people I wouldn't have reached otherwise. Even managed to add a new team member from Reddit. But most importantly, we now have the offer: Enterprise-grade AI assistant trained on your data for a fraction of the market cost. Basically a custom GPT for companies that want a secure assistant "trained" on their data but are not willing to spend millions on OpenAI's Custom Models or hundreds of thousands on Enterprise ChatGPT. (OpenAI's introduction of exclusive business GPTs for $2-3M is an incredibly good leverage for this offer). Also got rid of the big installment fee and switched to a $1k/month starting price for attractiveness and simplicity for companies (that covers their Azure fees also). The key offer points here are: Data security (as there are cheap, but not enterprise-grade tools like PDF.ai) Good price (as not all businesses can afford to pay 6 figure premiums for their data security) So the lesson here (I suppose) is that it's okay to take a step back sometimes. Reevaluate your direction. It's not worth sprinting when you're running in circles. P.S. finally made a website https://jongri.tech

We made $325k in 2023 from AI products, starting from 0, with no-code, no funding and no audience
reddit
LLM Vibe Score0
Human Vibe Score1
hopefully_usefulThis week

We made $325k in 2023 from AI products, starting from 0, with no-code, no funding and no audience

I met my co-founder in late 2022 after an introduction from a mutual friend to talk about how to find contract Product Management roles. I was sporadically contracting at start-up at the time and he had just come out of another start-up that was wiped out by the pandemic. We hit it off, talking about ideas, sharing what other indie-hackers were doing, and given GPT-3’s prominence at the time, we started throwing around ideas about things we could build with it, if nothing else, just to learn. I should caveat, neither of us were AI experts when starting out, everything we learned has been through Twitter and blogs, my background is as an accountant, and his a consultant. Here’s how it went since then: &#x200B; Nov 2022 (+$50) \- We built a simple tool in around a week using GPT-3 fine-tuning and a no-code tool (Bubble) that helped UK university students write their personal statements for their applications \- We set some Google Ads going and managed to make a few sales (\~$50) in the first week \- OpenAI were still approving applications at the time and said this went against their “ethics” so we had to take it down &#x200B; Dec 2022 (+$200) \- We couldn’t stop coming up with ideas related to AI fine-tuning, but realised it was almost impossible to decide which to pursue \- We needed a deadline to force us so we signed up for the Ben’s Bites hackathon in late December \- In a week, we built and launched a no-code fine-tuning platform, allowing people to create fine-tuned models by dragging and dropping an Excel file onto it \- We launched it on Product Hunt, having no idea how to price it, and somehow managed to get \~2,000 visitors on the site and make 2 sales at $99 &#x200B; Jan 2023 (+$3,000) \- We doubled down on the fine-tuning idea and managed to get up to \~$300 MRR, plus a bunch of one-time sales and a few paid calls to help people get the most out of their models \- We quickly realised that people didn’t want to curate models themselves, they just wanted to dump data and get magic out \- That was when we saw people building “Talk with x book/podcast” on Twitter as side projects and realised that was the missing piece, we needed to turn it into a tool \- We started working on the new product in late January &#x200B; Feb 2023 (+$9,000) \- We started pre-selling access to an MVP for the new product, which allowed people to “chat with their data/content”, we got $5,000 in pre-sales, more than we made from the previous product in total \- By mid-February, after 3 weeks of building we were able to launch and immediately managed to get traction, getting to $1k MRR in < 1 week, building on the hype of ChatGPT and AI (we were very lucky here) &#x200B; Mar - Jul 2023 (+$98,000) \- We worked all the waking hours to keep up with customer demand, bugs, OpenAI issues \- We built integrations for a bunch of services like Slack, Teams, Wordpress etc, added tons of new functionality and continue talking to customers every day \- We managed to grow to $17k MRR (just about enough to cover our living expenses and costs in London) through building in public on Twitter, newsletters and AI directories (and a million other little things) \- We sold our fine-tuning platform for \~$20k and our university project for \~$3k on Acquire &#x200B; Aug 2023 (+$100,000) \- We did some custom development work based on our own product for a customer that proved pretty lucrative &#x200B; Sep - Oct 2023 (+$62,000) \- After 8 months of building constantly, we started digging more seriously into our usage and saw subscriptions plateauing \- We talked to and analysed all our paying users to identify the main use cases and found 75% were for SaaS customer support \- We took the leap to completely rebuild a version of our product around this use case, our biggest to date (especially given most features with no-code took us <1 day) &#x200B; Nov - Dec 2023 (+$53,000) \- We picked up some small custom development work that utilised our own tech \- We’re sitting at around $22k MRR now with a few bigger clients signed up and coming soon \- After 2 months of building and talking to users, we managed to finish our “v2” of our product, focussed squarely on SaaS customer support and launched it today. &#x200B; We have no idea what the response will be to this new version, but we’re pretty happy with it, but couldn’t have planned anything that happened to us in 2023 so who knows what will come of 2024, we just know that we are going to be learning a ton more. &#x200B; Overall, it is probably the most I have had to think in my life - other jobs you can zone out from time to time or rely on someone else if you aren’t feeling it - not when you are doing this, case and point, I am writing this with a banging head-cold right now, but wanted to get this done. A few more things we have learned along the way - context switching is unreal, as is keeping up with, learning and reacting to AI. There isn’t a moment of the day I am not thinking about what we do next. But while in some way we now have hundreds of bosses (our customers) I still haven’t felt this free and can’t imagine ever going back to work for someone else. Next year we’re really hoping to figure out some repeatable distribution channels and personally, I want to get a lot better at creating content/writing, this is a first step! Hope this helps someone else reading this to just try starting something and see what happens.

The delicate balance of building an online community business
reddit
LLM Vibe Score0
Human Vibe Score0.895
matthewbarbyThis week

The delicate balance of building an online community business

Hey /r/Entrepreneur 👋 Just under two years ago I launched an online community business called Traffic Think Tank with two other co-founders, Nick Eubanks and Ian Howells. As a Traffic Think Tank customer you (currently) pay $119 a month to get access to our online community, which is run through Slack. The community is focused on helping you learn various aspects of marketing, with a particular focus on search engine optimization (SEO). Alongside access to the Slack community, we publish new educational video content from outside experts every week that all customers have access to. At the time of writing, Traffic Think Tank has around 650 members spanning across 17 of the 24 different global time zones. I was on a business trip over in Sydney recently, and during my time there I met up with some of our Australia-based community members. During dinner I was asked by several of them how the idea for Traffic Think Tank came about and what steps we took to validate that the idea was worth pursuing.  This is what I told them… How it all began It all started with a personal need. Nick, an already successful entrepreneur and owner of a marketing agency, had tested out an early version Traffic Think Tank in early 2017. He offered real-time consulting for around ten customers that he ran from Slack. He would publish some educational videos and offer his advice on projects that the members were running. The initial test went well, but it was tough to maintain on his own and he had to charge a fairly high price to make it worth his time. That’s when he spoke to me and Ian about turning this idea into something much bigger. Both Ian and I offered something slightly different to Nick. We’ve both spent time in senior positions at marketing agencies, but currently hold senior director positions in 2,000+ public employee companies (HubSpot and LendingTree). Alongside this, as a trio we could really ramp up the quality and quantity of content within the community, spread out the administrative workload and just generally have more resources to throw at getting this thing off the ground. Admittedly, Nick was much more optimistic about the potential of Traffic Think Tank – something I’m very thankful for now – whereas Ian and I were in the camp of “you’re out of your mind if you think hundreds of people are going to pay us to be a part of a Slack channel”. To validate the idea at scale, we decided that we’d get an initial MVP of the community up and running with a goal of reaching 100 paying customers in the first six months. If we achieved that, we’d validated that it was a viable business and we would continue to pursue it. If not, we’d kill it. We spent the next month building out the initial tech stack that enabled us to accept payments, do basic user management to the Slack channel, and get a one-page website up and running with information on what Traffic Think Tank was all about.  After this was ready, we doubled down on getting some initial content created for members – I mean, we couldn’t have people just land in an empty Slack channel, could we? We created around ten initial videos, 20 or so articles and then some long threads full of useful information within the Slack channel so that members would have some content to pour into right from the beginning.  Then, it was time to go live. The first 100 customers Fortunately, both Nick and I had built a somewhat substantial following in the SEO space over the previous 5-10 years, so we at least had a large email list to tap into (a total of around 40,000 people). We queued up some launch emails, set an initial price of $99 per month and pressed send. [\[LINK\] The launch email I sent to my subscribers announcing Traffic Think Tank](https://mailchi.mp/matthewbarby/future-of-marketing-1128181) What we didn’t expect was to sell all of the initial 100 membership spots in the first 72 hours. “Shit. What do we do now? Are we ready for this many people? Are we providing them with enough value? What if something breaks in our tech stack? What if they don’t like the content? What if everyone hates Slack?” All of these were thoughts running through my head. This brings me to the first great decision we made: we closed down new membership intake for 3 months so that we could focus completely on adding value to the first cohort of users. The right thing at the right time SEO is somewhat of a dark art to many people that are trying to learn about it for the first time. There’s hundreds of thousands (possibly millions) of articles and videos online that talk about how to do SEO.  Some of it’s good advice; a lot of it is very bad advice.  Add to this that the barrier to entry of claiming to be an “expert” in SEO is practically non-existent and you have a recipe for disaster. This is why, for a long time, individuals involved in SEO have flocked in their masses to online communities for information and to bounce ideas off of others in the space. Forums like SEObook, Black Hat World, WickedFire, Inbound.org, /r/BigSEO, and many more have, at one time, been called home by many SEOs.  In recent times, these communities have either been closed down or just simply haven’t adapted to the changing needs of the community – one of those needs being real-time feedback on real-world problems.  The other big need that we all spotted and personally had was the ability to openly share the things that are working – and the things that aren’t – in SEO within a private forum. Not everyone wanted to share their secret sauce with the world. One of the main reasons we chose Slack as the platform to run our community on was the fact that it solved these two core needs. It gave the ability to communicate in real-time across multiple devices, and all of the information shared within it was outside of the public domain. The other problem that plagued a lot of these early communities was spam. Most of them were web-based forums that were free to access. That meant they became a breeding ground for people trying to either sell their services or promote their own content – neither of which is conducive to building a thriving community. This was our main motivation for charging a monthly fee to access Traffic Think Tank. We spent a lot of time thinking through pricing. It needed to be enough money that people would be motivated to really make use of their membership and act in a way that’s beneficial to the community, but not too much money that it became cost prohibitive to the people that would benefit from it the most. Considering that most of our members would typically spend between $200-800 per month on SEO software, $99 initially felt like the perfect balance. Growing pains The first three months of running the community went by without any major hiccups. Members were incredibly patient with us, gave us great feedback and were incredibly helpful and accommodating to other members. Messages were being posted every day, with Nick, Ian and myself seeding most of the engagement at this stage.  With everything going smoothly, we decided that it was time to open the doors to another intake of new members. At this point we’d accumulated a backlog of people on our waiting list, so we knew that simply opening our doors would result in another large intake. Adding more members to a community has a direct impact on the value that each member receives. For Traffic Think Tank in particular, the value for members comes from three areas: The ability to have your questions answered by me, Nick and Ian, as well as other members of the community. The access to a large library of exclusive content. The ability to build connections with the wider community. In the early stages of membership growth, there was a big emphasis on the first of those three points. We didn’t have an enormous content library, nor did we have a particularly large community of members, so a lot of the value came from getting a lot of one-to-one time with the community founders. [\[IMAGE\] Screenshot of engagement within the Traffic Think Tank Slack community](https://cdn.shortpixel.ai/client/qglossy,retimg,w_1322/https://www.matthewbarby.com/wp-content/uploads/2019/08/Community-Engagement-in-Traffic-Think-Tank.png) The good thing about having 100 members was that it was just about feasible to give each and every member some one-to-one time within the month, which really helped us to deliver those moments of delight that the community needed early on. Two-and-a-half months after we launched Traffic Think Tank, we opened the doors to another 250 people, taking our total number of members to 350. This is where we experienced our first growing pains.  Our original members had become used to being able to drop us direct messages and expect an almost instant response, but this wasn’t feasible anymore. There were too many people, and we needed to create a shift in behavior. We needed more value to come from the community engaging with one another or we’d never be able to scale beyond this level. We started to really pay attention to engagement metrics; how many people were logging in every day, and of those, how many were actually posting messages within public channels.  We asked members that were logging in a lot but weren’t posting (the “lurkers”) why that was the case. We also asked the members that engaged in the community the most what motivated them to post regularly. We learned a lot from doing this. We found that the large majority of highly-engaged members had much more experience in SEO, whereas most of the “lurkers” were beginners. This meant that most of the information being shared in the community was very advanced, with a lot of feedback from the beginners in the group being that they “didn’t want to ask a stupid question”.  As managers of the community, we needed to facilitate conversations that catered to all of our members, not just those at a certain level of skill. To tackle this problem, we created a number of new channels that had a much deeper focus on beginner topics so novice members had a safe place to ask questions without judgment.  We also started running live video Q&As each month where we’d answer questions submitted by the community. This gave our members one-on-one time with me, Nick and Ian, but spread the value of these conversations across the whole community rather than them being hidden within private messages. As a result of these changes, we found that the more experienced members in the community were really enjoying sharing their knowledge with those with less experience. The number of replies within each question thread was really starting to increase, and the community started to shift away from just being a bunch of threads created by me, Nick and Ian to a thriving forum of diverse topics compiled by a diverse set of individuals. This is what we’d always wanted. A true community. It was starting to happen. [\[IMAGE\] Chart showing community engagement vs individual member value](https://cdn.shortpixel.ai/client/qglossy,retimg,w_1602/https://www.matthewbarby.com/wp-content/uploads/2019/08/Community-Engagement-Balance-Graph.jpg) At the same time, we started to realize that we’ll eventually reach a tipping point where there’ll be too much content for us to manage and our members to engage with. When we reach this point, the community will be tough to follow and the quality of any given post will go down. Not only that, but the community will become increasingly difficult to moderate. We’re not there yet, but we recognize that this will come, and we’ll have to adjust our model again. Advocating advocacy As we started to feel more comfortable about the value that members were receiving, we made the decision to indefinitely open for new members. At the same time, we increased the price of membership (from $99 a month to $119) in a bid to strike the right balance between profitability as a business and to slow down the rate at which we were reaching the tipping point of community size. We also made the decision to repay all of our early adopters by grandfathering them in to the original pricing – and committing to always do this in the future. Despite the price increase, we saw a continued flow of new members come into the community. The craziest part about this was that we were doing practically no marketing activities to encourage new members– this was all coming from word of mouth. Our members were getting enough value from the community that they were recommending it to their friends, colleagues and business partners.  The scale at which this was happening really took us by surprise and it told us one thing very clearly: delivering more value to members resulted in more value being delivered to the business. This is a wonderful dynamic to have because it perfectly aligns the incentives on both sides. We’d said from the start that we wouldn’t sacrifice value to members for more revenue – this is something that all three of us felt very strongly about. First and foremost, we wanted to create a community that delivered value to its members and was run in a way that aligned with our values as people. If we could find a way to stimulate brand advocacy, while also tightening the bonds between all of our individual community members, we’d be boosting both customer retention and customer acquisition in the same motion. This became our next big focus. [\[TWEET\] Adam, one of our members wore his Traffic Think Tank t-shirt in the Sahara desert](https://twitter.com/AdamGSteele/status/1130892481099382784) We started with some simple things: We shipped out Traffic Think Tank branded T-shirts to all new members. We’d call out each of the individuals that would submit questions to our live Q&A sessions and thank them live on air. We set up a new channel that was dedicated to sharing a quick introduction to who you are, what you do and where you’re based for all new members. We’d created a jobs channel and a marketplace for selling, buying and trading services with other members. Our monthly “blind dates” calls were started where you’d be randomly grouped with 3-4 other community members so that you could hop on a call to get to know each other better. The Traffic Think Tank In Real Life (IRL)* channel was born, which enabled members to facilitate in-person meetups with each other. In particular, we saw that as members started to meet in person or via calls the community itself was feeling more and more like a family. It became much closer knit and some members started to build up a really positive reputation for being particularly helpful to other members, or for having really strong knowledge in a specific area. [\[TWEET\] Dinner with some of the Traffic Think Tank members in Brighton, UK](https://twitter.com/matthewbarby/status/1117175584080134149) Nick, Ian and I would go out of our way to try and meet with members in real life wherever we could. I was taken aback by how appreciative people were for us doing this, and it also served as an invaluable way to gain honest feedback from members. There was another trend that we’d observed that we didn’t really expect to happen. More and more members were doing business with each another. We’ve had people find new jobs through the community, sell businesses to other members, launch joint ventures together and bring members in as consultants to their business. This has probably been the most rewarding thing to watch, and it was clear that the deeper relationships that our members were forming were resulting in an increased level of trust to work with each other. We wanted to harness this and take it to a new level. This brought us to arguably the best decision we’ve made so far running Traffic Think Tank… we were going to run a big live event for our members. I have no idea what I’m doing It’s the first week of January 2019 and we’re less than three weeks away from Traffic Think Tank LIVE, our first ever in-person event hosting 150 people, most of which are Traffic Think Tank members. It's like an ongoing nightmare I can’t wake up from. That was Nick’s response in our private admin channel to myself and Ian when I asked if they were finding the run-up to the event as stressful as I was. I think that all three of us were riding on such a high from how the community was growing that we felt like we could do anything. Running an event? How hard can it be? Well, turns out it’s really hard. We had seven different speakers flying over from around the world to speak at the event, there was a pre- and after event party, and we’d planned a charity dinner where we would take ten attendees (picked at random via a raffle) out for a fancy meal. Oh, and Nick, Ian and I were hosting a live Q&A session on stage. It wasn’t until precisely 48 hours before the event that we’d realized we didn’t have any microphones, nor had a large amount of the swag we’d ordered arrived. Plus, a giant storm had hit Philly causing a TON of flight cancellations. Perfect. Just perfect. This was honestly the tip of the iceberg. We hadn’t thought about who was going to run the registration desk, who would be taking photos during the event and who would actually field questions from the audience while all three of us sat on stage for our live Q&A panel. Turns out that the answer to all of those questions were my wife, Laura, and Nick’s wife, Kelley. Thankfully, they were on hand to save our asses. The weeks running up to the event were honestly some of the most stressful of my life. We sold around 50% of our ticket allocation within the final two weeks before the event. All of the event organizers told us this would happen, but did we believe them? Hell no!  Imagine having two weeks until the big day and as it stood half of the room would be completely empty. I was ready to fly most of my extended family over just to make it look remotely busy. [\[IMAGE\] One of our speakers, Ryan Stewart, presenting at Traffic Think Tank LIVE](https://cdn.shortpixel.ai/client/qglossy,retimg,w_1920/https://www.matthewbarby.com/wp-content/uploads/2019/08/Traffic-Think-Tank-LIVE-Ryan-Presenting.jpg) Thankfully, if all came together. We managed to acquire some microphones, the swag arrived on the morning of the event, all of our speakers were able to make it on time and the weather just about held up so that our entire allocation of ticket holders was able to make it to the event. We pooled together and I’m proud to say that the event was a huge success. While we made a substantial financial loss on the event itself, January saw a huge spike in new members, which more than recouped our losses. Not only that, but we got to hang out with a load of our members all day while they said really nice things about the thing we’d built. It was both exhausting and incredibly rewarding. Bring on Traffic Think Tank LIVE 2020! (This time we’re hiring an event manager...)   The road ahead Fast forward to today (August 2019) and Traffic Think Tank has over 650 members. The biggest challenges that we’re tackling right now include making sure the most interesting conversations and best content surfaces to the top of the community, making Slack more searchable (this is ultimately one of its flaws as a platform) and giving members a quicker way to find the exclusive content that we create. You’ll notice there’s a pretty clear theme here. In the past 30 days, 4,566 messages were posted in public channels inside Traffic Think Tank. If you add on any messages posted inside private direct messages, this number rises to 21,612. That’s a lot of messages. To solve these challenges and enable further scale in the future, we’ve invested a bunch of cash and our time into building out a full learning management system (LMS) that all members will get access to alongside the Slack community. The LMS will be a web-based portal that houses all of the video content we produce. It will also  provide an account admin section where users can update or change their billing information (they have to email us to do this right now, which isn’t ideal), a list of membership perks and discounts with our partners, and a list of links to some of the best threads within Slack – when clicked, these will drop you directly into Slack. [\[IMAGE\] Designs for the new learning management system (LMS)](https://cdn.shortpixel.ai/client/qglossy,retimg,w_2378/https://www.matthewbarby.com/wp-content/uploads/2019/08/Traffic-Think-Tank-LMS.png) It’s not been easy, but we’re 95% of the way through this and I’m certain that it will have a hugely positive impact on the experience for our members. Alongside this we hired a community manager, Liz, who supports with any questions that our members have, coordinates with external experts to arrange webinars for the community, helps with new member onboarding, and has tightened up some of our processes around billing and general accounts admin. This was a great decision. Finally, we’ve started planning next year’s live event, which we plan to more than double in size to 350 attendees, and we decided to pick a slightly warmer location in Miami this time out. Stay tuned for me to have a complete meltdown 3 weeks from the event. Final thoughts When I look back on the journey we’ve had so far building Traffic Think Tank, there’s one very important piece to this puzzle that’s made all of this work that I’ve failed to mention so far: co-founder alignment. Building a community is a balancing act that relies heavily on those in charge being completely aligned. Nick, Ian and I completely trust each other and more importantly, are philosophically aligned on how we want to run and grow the community. If we didn’t have this, the friction between us could tear apart the entire community. Picking the right people to work with is important in any company, but when your business is literally about bringing people together, there’s no margin for error here.  While I’m sure there will be many more challenges ahead, knowing that we all trust each other to make decisions that fall in line with each of our core values makes these challenges dramatically easier to overcome. Finally, I’d like to thank all of our members for making the community what it is today – it’d be nothing without you and I promise that we’ll never take that for granted. &#x200B; I originally posted this on my blog here. Welcoming all of your thoughts, comments, questions and I'll do my best to answer them :)

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security

Uzair Javaid, a Ph.D. with a passion for data privacy, co-founded Betterdata to tackle one of AI's most pressing challenges: protecting privacy while enabling innovation. Recently, Betterdata secured a lucrative contract with the US Department of Homeland Security, 1 of only 4 companies worldwide to do so and the only one in Asia. Here's how he did it: The Story So what's your story? I grew up in Peshawar, Pakistan, excelling in coding despite studying electrical engineering. Inspired by my professors, I set my sights on studying abroad and eventually earned a Ph.D. scholarship at NUS Singapore, specializing in data security and privacy. During my research, I ethically hacked Ethereum and published 15 papers—three times the requirement. While wrapping up my Ph.D., I explored startup ideas and joined Entrepreneur First, where I met Kevin Yee. With his expertise in generative models and mine in privacy, we founded Betterdata. Now, nearly three years in, we’ve secured a major contract with the U.S. Department of Homeland Security—one of only four companies globally and the only one from Asia. The Startup In a nutshell, what does your startup do? Betterdata is a startup that uses AI and synthetic data generation to address two major challenges: data privacy and the scarcity of high-quality data for training AI models. By leveraging generative models and privacy-enhancing technologies, Betterdata enables businesses, such as banks, to use customer data without breaching privacy regulations. The platform trains AI on real data, learns its patterns, and generates synthetic data that mimics the real thing without containing any personal or sensitive information. This allows companies to innovate and develop AI solutions safely and ethically, all while tackling the growing need for diverse, high-quality data in AI development. How did you conduct ideation and validation for your startup? The initial idea for Betterdata came from personal experience. During my Ph.D., I ethically hacked Ethereum’s blockchain, exposing flaws in encryption-based data sharing. This led me to explore AI-driven deep synthesis technology—similar to deepfakes but for structured data privacy. With GDPR impacting 28M+ businesses, I saw a massive opportunity to help enterprises securely share data while staying compliant. To validate the idea, I spoke to 50 potential customers—a number that strikes the right balance. Some say 100, but that’s impractical for early-stage founders. At 50, patterns emerge: if 3 out of 10 mention the same problem, and this repeats across 50, you have 10–15 strong signals, making it a solid foundation for an MVP. Instead of outbound sales, which I dislike, we used three key methods: Account-Based Marketing (ABM)—targeting technically savvy users with solutions for niche problems, like scaling synthetic data for banks. Targeted Content Marketing—regular customer conversations shaped our thought leadership and outreach. Raising Awareness Through Partnerships—collaborating with NUS, Singapore’s PDPC, and Plug and Play to build credibility and educate the market. These strategies attracted serious customers willing to pay, guiding Betterdata’s product development and market fit. How did you approach the initial building and ongoing product development? In the early stages, we built synthetic data generation algorithms and a basic UI for proof-of-concept, using open-source datasets to engage with banks. We quickly learned that banks wouldn't share actual customer data due to privacy concerns, so we had to conduct on-site installations and gather feedback to refine our MVP. Through continuous consultation with customers, we discovered real enterprise data posed challenges, such as missing values, which led us to adapt our prototype accordingly. This iterative approach of listening to customer feedback and observing their usage allowed us to improve our product, enhance UX, and address unmet needs while building trust and loyalty. Working closely with our customers also gives us a data advantage. Our solution’s effectiveness depends on customer data, which we can't fully access, but bridging this knowledge gap gives us a competitive edge. The more customers we test on, the more our algorithms adapt to diverse use cases, making it harder for competitors to replicate our insights. My approach to iteration is simple: focus solely on customer feedback and ignore external noise like trends or advice. The key question for the team is: which customer is asking for this feature or solution? As long as there's a clear answer, we move forward. External influences, such as AI hype, often bring more confusion than clarity. True long-term success comes from solving real customer problems, not chasing trends. Customers may not always know exactly what they want, but they understand their problems. Our job is to identify these problems and solve them in innovative ways. While customers may suggest specific features, we stay focused on solving the core issue rather than just fulfilling their exact requests. The idea aligns with the quote often attributed to Henry Ford: "If I asked people what they wanted, they would have said faster horses." The key is understanding their problems, not just taking requests at face value. How do you assess product-market fit? To assess product-market fit, we track two key metrics: Customers' Willingness to Pay: We measure both the quantity and quality of meetings with potential customers. A high number of meetings with key decision-makers signals genuine interest. At Betterdata, we focused on getting meetings with people in banks and large enterprises to gauge our product's resonance with the target market. How Much Customers Are Willing to Pay: We monitor the price customers are willing to pay, especially in the early stages. For us, large enterprises, like banks, were willing to pay a premium for our synthetic data platform due to the growing need for privacy tech. This feedback guided our product refinement and scaling strategy. By focusing on these metrics, we refined our product and positioned it for scaling. What is your business model? We employ a structured, phase-driven approach for out business model, as a B2B startup. I initially struggled with focusing on the core value proposition in sales, often becoming overly educational. Eventually, we developed a product roadmap with models that allowed us to match customer needs to specific offerings and justify our pricing. Our pricing structure includes project-based pilots and annual contracts for successful deployments. At Betterdata, our customer engagement unfolds across three phases: Phase 1: Trial and Benchmarking \- We start with outreach and use open-source datasets to showcase results, offering customers a trial period to evaluate the solution. Phase 2: Pilot or PoC \- After positive trial results, we conduct a PoC or pilot using the customer’s private data, with the understanding that successful pilots lead to an annual contract. Phase 3: Multi-Year Contracts \- Following a successful pilot, we transition to long-term commercial contracts, focusing on multi-year agreements to ensure stability and ongoing partnerships. How do you do marketing for your brand? We take a non-conventional approach to marketing, focusing on answering one key question: Which customers are willing to pay, and how much? This drives our messaging to show how our solution meets their needs. Our strategy centers around two main components: Building a network of lead magnets \- These are influential figures like senior advisors, thought leaders, and strategic partners. Engaging with institutions like IMDA, SUTD, and investors like Plug and Play helps us gain access to the right people and foster warm introductions, which shorten our sales cycle and ensure we’re reaching the right audience. Thought leadership \- We build our brand through customer traction, technology evidence, and regulatory guidelines. This helps us establish credibility in the market and position ourselves as trusted leaders in our field. This holistic approach has enabled us to navigate diverse market conditions in Asia and grow our B2B relationships. By focusing on these areas, we drive business growth and establish strong trust with stakeholders. What's your advice for fundraising? Here are my key takeaways for other founders when it comes to fundraising: Fundraise When You Don’t Need To We closed our seed round in April 2023, a time when we weren't actively raising. Founders should always be in fundraising mode, even when they're not immediately in need of capital. Don’t wait until you have only a few months of runway left. Keep the pipeline open and build relationships. When the timing is right, execution becomes much easier. For us, our investment came through a combination of referrals and inbound interest. Even our lead investor initially rejected us, but after re-engaging, things eventually fell into place. It’s crucial to stay humble, treat everyone with respect, and maintain those relationships for when the time is right. Be Mindful of How You Present Information When fundraising, how you present information matters a lot. We created a comprehensive, easily digestible investment memo, hosted on Notion, which included everything an investor might need—problem, solution, market, team, risks, opportunities, and data. The goal was for investors to be able to get the full picture within 30 minutes without chasing down extra details. We also focused on making our financial model clear and meaningful, even though a 5-year forecast might be overkill at the seed stage. The key was clarity and conciseness, and making it as easy as possible for investors to understand the opportunity. I learned that brevity and simplicity are often the best ways to make a memorable impact. For the pitch itself, keep it simple and focus on 4 things: problem, solution, team, and market. If you can summarize each of these clearly and concisely, you’ll have a compelling pitch. Later on, you can expand into market segments, traction, and other metrics, but for seed-stage, focus on those four areas, and make sure you’re strong in at least three of them. If you do, you'll have a compelling case. How do you run things day-to-day? i.e what's your operational workflow and team structure? Here's an overview of our team structure and process: Internally: Our team is divided into two main areas: backend (internal team) and frontend (market-facing team). There's no formal hierarchy within the backend team. We all operate as equals, defining our goals based on what needs to be developed, assigning tasks, and meeting weekly to share updates and review progress. The focus is on full ownership of tasks and accountability for getting things done. I also contribute to product development, identifying challenges and clearing obstacles to help the team move forward. Backend Team: We approach tasks based on the scope defined by customers, with no blame or hierarchy. It's like a sports team—sometimes someone excels, and other times they struggle, but we support each other and move forward together. Everyone has the creative freedom to work in the way that suits them best, but we establish regular meetings and check-ins to ensure alignment and progress. Frontend Team: For the market-facing side, we implement a hierarchy because the market expects this structure. If I present myself as "CEO," it signals authority and credibility. This distinction affects how we communicate with the market and how we build our brand. The frontend team is split into four main areas: Business Product (Software Engineering) Machine Learning Engineering R&D The C-suite sits at the top, followed by team leads, and then the executors. We distill market expectations into actionable tasks, ensuring that everyone is clear on their role and responsibilities. Process: We start by receiving market expectations and defining tasks based on them. Tasks are assigned to relevant teams, and execution happens with no communication barriers between team members. This ensures seamless collaboration and focused execution. The main goal is always effectiveness—getting things done efficiently while maintaining flexibility in how individuals approach their work. In both teams, there's an emphasis on accountability, collaboration, and clear communication, but the structure varies according to the nature of the work and external expectations.

Raised $450k for my startup, here are the lessons I've learned along the way
reddit
LLM Vibe Score0
Human Vibe Score1
marin_smiljanicThis week

Raised $450k for my startup, here are the lessons I've learned along the way

2021 has been a pretty amazing year for Omnisearch. Having started initial work on Omnisearch at the end of 2020, we entered the new year with a working MVP yet no revenue, no significant partnerships, and no funding. Fast forward to the end of 2021, and we now have fantastic revenue growth, a partnership with a public company, and a far more powerful, complete and polished product. But one milestone really changed Omnisearch’s trajectory: our $450,000 USD pre-seed round by GoAhead Ventures. In this post I want to share the story of how it came about and offer a couple of takeaways to keep in mind when preparing for fundraising. &#x200B; The story Contrary to most advice, my co-founder Matej and I didn’t allocate a specific time to switch to “fundraising mode” but rather talked to investors on an ongoing basis. It was a bit of a distraction from working on the product, but on the positive side we were able to constantly get feedback on the idea, pitch, go-to-market strategy and hiring, as well as hearing investors’ major concerns sooner rather than later. That being said, our six-month long fundraising efforts weren’t yielding results - we talked to about twenty investors, mostly angels or smaller funds, with no success. The feedback was generally of the “too early for us” variety (since we were still pre-revenue), with additional questions about our go-to-market strategy and ideal customer persona. The introduction to our eventual investors, California-based GoAhead Ventures, came through a friend who had pitched them previously. We wrote a simple blurb and sent our pitch deck. We then went through GoAhead’s hyper-efficient screening process, consisting of a 30-minute call, a recorded three-minute pitch, and filling out a simple Google doc. Throughout the whole process, the GoAhead team left an awesome impression thanks to their knowledge of enterprise software and their responsiveness. They ended up investing and the whole deal was closed within two weeks, which is super fast even by Silicon Valley standards. While our fundraising experience is a single data point and your case might be different, here are the key takeaways from our journey. &#x200B; Perseverance wins: Like I said above, we talked to about twenty investors before we closed our round. Getting a series of “no”s sucks, but we took the feedback seriously and tried to prepare better for questions that caught us off guard. But we persevered, keeping in mind that from a bird’s eye perspective it’s an amazing time to be building startups and raising funds. Focus on traction: Sounds pretty obvious, right? The truth is, though, that even a small amount of revenue is infinitely better than none at all. One of the major differences between our eventual successful investor pitch and the earlier ones was that we had actual paying customers, though our MRR was low. This allows you to talk about customers in the present tense, showing there’s actual demand for your product and making the use cases more tangible. And ideally, highlight a couple of customer testimonials to boost your credibility. Have a demo ready: In Omnisearch’s case, the demo was oftentimes the best received part of the pitch or call. We’d show investors the live demo, and for bonus points even asked them to choose a video from YouTube and then try searching through it. This always had a “wow” effect on prospective investors and made the subsequent conversation more exciting and positive. Accelerators: Accelerators like Y Combinator or Techstars can add enormous value to a startup, especially in the early stages. And while it’s a great idea to apply, don’t rely on them too heavily. Applications happen only a few times a year, and you should have a foolproof fundraising plan in case you don’t get in. In our case, we just constantly looked for investors who were interested in our space (defined as enterprise SaaS more broadly), using LinkedIn, AngelList, and intros from our own network. Practice the pitch ad nauseam: Pitching is tough to get right even for seasoned pros, so it pays to practice as often as possible. We took every opportunity to perfect the pitch: attending meetups and giving the thirty-second elevator pitch to other attendees over beer and pizza, participating in startup competitions, going to conferences and exhibiting at our own booth, attending pre-accelerator programs, and pitching to friends who are in the startup world. Show an understanding of the competition: Frankly, this was one of the strongest parts of our pitch and investor conversations. If you’re in a similar space to ours, Gartner Magic Quadrants and Forrester Waves are an awesome resource, as well as sites like AlternativeTo or Capterra and G2. By thoroughly studying these resources we gained a great understanding of the industry landscape and were able to articulate our differentiation more clearly and succinctly. Presenting this visually in a coordinate system or a feature grid is, from our experience, even more effective. Remember it’s just the beginning! Getting your first round of funding is just the beginning of the journey, so it’s important to avoid euphoria and get back to building and selling the product as soon as possible. While securing funding enables you to scale the team, and is a particular relief if the founders had worked without a salary, the end goal is still to build a big, profitable, and overall awesome startup.

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies)
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies)

AI Palette is an AI-driven platform that helps food and beverage companies predict emerging product trends. I had the opportunity recently to sit down with the founder to get his advice on building an AI-first startup, which he'll be going through in this post. About AI Palette: Co-founders: >!2 (Somsubhra GanChoudhuri, Himanshu Upreti)!!100+!!$12.7M USD!!AI-powered predictive analytics for the CPG (Consumer Packaged Goods) industry!!Signed first paying customer in the first year!!65+ global brands, including Cargill, Diageo, Ajinomoto, Symrise, Mondelez, and L’Oréal, use AI Palette!!Every new product launched has secured a paying client within months!!Expanded into Beauty & Personal Care (BPC), onboarding one of India’s largest BPC companies within weeks!!Launched multiple new product lines in the last two years, creating a unified suite for brand innovation!Identify the pain points in your industry for ideas* When I was working in the flavour and fragrance industry, I noticed a major issue CPG companies faced: launching a product took at least one to two years. For instance, if a company decided today to launch a new juice, it wouldn’t hit the market until 2027. This long timeline made it difficult to stay relevant and on top of trends. Another big problem I noticed was that companies relied heavily on market research to determine what products to launch. While this might work for current consumer preferences, it was highly inefficient since the product wouldn’t actually reach the market for several years. By the time the product launched, the consumer trends had already shifted, making that research outdated. That’s where AI can play a crucial role. Instead of looking at what consumers like today, we realised that companies should use AI to predict what they will want next. This allows businesses to create products that are ahead of the curve. Right now, the failure rate for new product launches is alarmingly high, with 8 out of 10 products failing. By leveraging AI, companies can avoid wasting resources on products that won’t succeed, leading to better, more successful launches. Start by talking to as many industry experts as possible to identify the real problems When we first had the idea for AI Palette, it was just a hunch, a gut feeling—we had no idea whether people would actually pay for it. To validate the idea, we reached out to as many people as we could within the industry. Since our focus area was all about consumer insights, we spoke to professionals in the CPG sector, particularly those in the insights departments of CPG companies. Through these early conversations, we began to see a common pattern emerge and identified the exact problem we wanted to solve. Don’t tell people what you’re building—listen to their frustrations and challenges first. Going into these early customer conversations, our goal was to listen and understand their challenges without telling them what we were trying to build. This is crucial as it ensures that you can gather as much data about the problem to truly understand it and that you aren't biasing their answers by showing your solution. This process helped us in two key ways: First, it validated that there was a real problem in the industry through the number of people who spoke about experiencing the same problem. Second, it allowed us to understand the exact scale and depth of the problem—e.g., how much money companies were spending on consumer research, what kind of tools they were currently using, etc. Narrow down your focus to a small, actionable area to solve initially. Once we were certain that there was a clear problem worth solving, we didn’t try to tackle everything at once. As a small team of two people, we started by focusing on a specific area of the problem—something big enough to matter but small enough for us to handle. Then, we approached customers with a potential solution and asked them for feedback. We learnt that our solution seemed promising, but we wanted to validate it further. If customers are willing to pay you for the solution, it’s a strong validation signal for market demand. One of our early customer interviewees even asked us to deliver the solution, which we did manually at first. We used machine learning models to analyse the data and presented the results in a slide deck. They paid us for the work, which was a critical moment. It meant we had something with real potential, and we had customers willing to pay us before we had even built the full product. This was the key validation that we needed. By the time we were ready to build the product, we had already gathered crucial insights from our early customers. We understood the specific information they wanted and how they wanted the results to be presented. This input was invaluable in shaping the development of our final product. Building & Product Development Start with a simple concept/design to validate with customers before building When we realised the problem and solution, we began by designing the product, but not by jumping straight into coding. Instead, we created wireframes and user interfaces using tools like InVision and Figma. This allowed us to visually represent the product without the need for backend or frontend development at first. The goal was to showcase how the product would look and feel, helping potential customers understand its value before we even started building. We showed these designs to potential customers and asked for feedback. Would they want to buy this product? Would they pay for it? We didn’t dive into actual development until we found a customer willing to pay a significant amount for the solution. This approach helped us ensure we were on the right track and didn’t waste time or resources building something customers didn’t actually want. Deliver your solution using a manual consulting approach before developing an automated product Initially, we solved problems for customers in a more "consulting" manner, delivering insights manually. Recall how I mentioned that when one of our early customer interviewees asked us to deliver the solution, we initially did it manually by using machine learning models to analyse the data and presenting the results to them in a slide deck. This works for the initial stages of validating your solution, as you don't want to invest too much time into building a full-blown MVP before understanding the exact features and functionalities that your users want. However, after confirming that customers were willing to pay for what we provided, we moved forward with actual product development. This shift from a manual service to product development was key to scaling in a sustainable manner, as our building was guided by real-world feedback and insights rather than intuition. Let ongoing customer feedback drive iteration and the product roadmap Once we built the first version of the product, it was basic, solving only one problem. But as we worked closely with customers, they requested additional features and functionalities to make it more useful. As a result, we continued to evolve the product to handle more complex use cases, gradually developing new modules based on customer feedback. Product development is a continuous process. Our early customers pushed us to expand features and modules, from solving just 20% of their problems to tackling 50–60% of their needs. These demands shaped our product roadmap and guided the development of new features, ultimately resulting in a more complete solution. Revenue and user numbers are key metrics for assessing product-market fit. However, critical mass varies across industries Product-market fit (PMF) can often be gauged by looking at the size of your revenue and the number of customers you're serving. Once you've reached a certain critical mass of customers, you can usually tell that you're starting to hit product-market fit. However, this critical mass varies by industry and the type of customers you're targeting. For example, if you're building an app for a broad consumer market, you may need thousands of users. But for enterprise software, product-market fit may be reached with just a few dozen key customers. Compare customer engagement and retention with other available solutions on the market for product-market fit Revenue and the number of customers alone isn't always enough to determine if you're reaching product-market fit. The type of customer and the use case for your product also matter. The level of engagement with your product—how much time users are spending on the platform—is also an important metric to track. The more time they spend, the more likely it is that your product is meeting a crucial need. Another way to evaluate product-market fit is by assessing retention, i.e whether users are returning to your platform and relying on it consistently, as compared to other solutions available. That's another key indication that your solution is gaining traction in the market. Business Model & Monetisation Prioritise scalability Initially, we started with a consulting-type model where we tailor-made specific solutions for each customer use-case we encountered and delivered the CPG insights manually, but we soon realized that this wasn't scalable. The problem with consulting is that you need to do the same work repeatedly for every new project, which requires a large team to handle the workload. That is not how you sustain a high-growth startup. To solve this, we focused on building a product that would address the most common problems faced by our customers. Once built, this product could be sold to thousands of customers without significant overheads, making the business scalable. With this in mind, we decided on a SaaS (Software as a Service) business model. The benefit of SaaS is that once you create the software, you can sell it to many customers without adding extra overhead. This results in a business with higher margins, where the same product can serve many customers simultaneously, making it much more efficient than the consulting model. Adopt a predictable, simplistic business model for efficiency. Look to industry practices for guidance When it came to monetisation, we considered the needs of our CPG customers, who I knew from experience were already accustomed to paying annual subscriptions for sales databases and other software services. We decided to adopt the same model and charge our customers an annual upfront fee. This model worked well for our target market, aligning with industry standards and ensuring stable, recurring revenue. Moreover, our target CPG customers were already used to this business model and didn't have to choose from a huge variety of payment options, making closing sales a straightforward and efficient process. Marketing & Sales Educate the market to position yourself as a thought leader When we started, AI was not widely understood, especially in the CPG industry. We had to create awareness around both AI and its potential value. Our strategy focused on educating potential users and customers about AI, its relevance, and why they should invest in it. This education was crucial to the success of our marketing efforts. To establish credibility, we adopted a thought leadership approach. We wrote blogs on the importance of AI and how it could solve problems for CPG companies. We also participated in events and conferences to demonstrate our expertise in applying AI to the industry. This helped us build our brand and reputation as leaders in the AI space for CPG, and word-of-mouth spread as customers recognized us as the go-to company for AI solutions. It’s tempting for startups to offer products for free in the hopes of gaining early traction with customers, but this approach doesn't work in the long run. Free offerings don’t establish the value of your product, and customers may not take them seriously. You should always charge for pilots, even if the fee is minimal, to ensure that the customer is serious about potentially working with you, and that they are committed and engaged with the product. Pilots/POCs/Demos should aim to give a "flavour" of what you can deliver A paid pilot/POC trial also gives you the opportunity to provide a “flavour” of what your product can deliver, helping to build confidence and trust with the client. It allows customers to experience a detailed preview of what your product can do, which builds anticipation and desire for the full functionality. During this phase, ensure your product is built to give them a taste of the value you can provide, which sets the stage for a broader, more impactful adoption down the line. Fundraising & Financial Management Leverage PR to generate inbound interest from VCs When it comes to fundraising, our approach was fairly traditional—we reached out to VCs and used connections from existing investors to make introductions. However, looking back, one thing that really helped us build momentum during our fundraising process was getting featured in Tech in Asia. This wasn’t planned; it just so happened that Tech in Asia was doing a series on AI startups in Southeast Asia and they reached out to us for an article. During the interview, they asked if we were fundraising, and we mentioned that we were. As a result, several VCs we hadn’t yet contacted reached out to us. This inbound interest was incredibly valuable, and we found it far more effective than our outbound efforts. So, if you can, try to generate some PR attention—it can help create inbound interest from VCs, and that interest is typically much stronger and more promising than any outbound strategies because they've gone out of their way to reach out to you. Be well-prepared and deliberate about fundraising. Keep trying and don't lose heart When pitching to VCs, it’s crucial to be thoroughly prepared, as you typically only get one shot at making an impression. If you mess up, it’s unlikely they’ll give you a second chance. You need to have key metrics at your fingertips, especially if you're running a SaaS company. Be ready to answer questions like: What’s your retention rate? What are your projections for the year? How much will you close? What’s your average contract value? These numbers should be at the top of your mind. Additionally, fundraising should be treated as a structured process, not something you do on the side while juggling other tasks. When you start, create a clear plan: identify 20 VCs to reach out to each week. By planning ahead, you’ll maintain momentum and speed up the process. Fundraising can be exhausting and disheartening, especially when you face multiple rejections. Remember, you just need one investor to say yes to make it all worthwhile. When using funds, prioritise profitability and grow only when necessary. Don't rely on funding to survive. In the past, the common advice for startups was to raise money, burn through it quickly, and use it to boost revenue numbers, even if that meant operating at a loss. The idea was that profitability wasn’t the main focus, and the goal was to show rapid growth for the next funding round. However, times have changed, especially with the shift from “funding summer” to “funding winter.” My advice now is to aim for profitability as soon as possible and grow only when it's truly needed. For example, it’s tempting to hire a large team when you have substantial funds in the bank, but ask yourself: Do you really need 10 new hires, or could you get by with just four? Growing too quickly can lead to unnecessary expenses, so focus on reaching profitability as soon as possible, rather than just inflating your team or burn rate. The key takeaway is to spend your funds wisely and only when absolutely necessary to reach profitability. You want to avoid becoming dependent on future VC investments to keep your company afloat. Instead, prioritize reaching break-even as quickly as you can, so you're not reliant on external funding to survive in the long run. Team-Building & Leadership Look for complementary skill sets in co-founders When choosing a co-founder, it’s important to find someone with a complementary skill set, not just someone you’re close to. For example, I come from a business and commercial background, so I needed someone with technical expertise. That’s when I found my co-founder, Himanshu, who had experience in machine learning and AI. He was a great match because his technical knowledge complemented my business skills, and together we formed a strong team. It might seem natural to choose your best friend as your co-founder, but this can often lead to conflict. Chances are, you and your best friend share similar interests, skills, and backgrounds, which doesn’t bring diversity to the table. If both of you come from the same industry or have the same strengths, you may end up butting heads on how things should be done. Having diverse skill sets helps avoid this and fosters a more collaborative working relationship. Himanshu (left) and Somsubhra (right) co-founded AI Palette in 2018 Define roles clearly to prevent co-founder conflict To avoid conflict, it’s essential that your roles as co-founders are clearly defined from the beginning. If your co-founder and you have distinct responsibilities, there is no room for overlap or disagreement. This ensures that both of you can work without stepping on each other's toes, and there’s mutual respect for each other’s expertise. This is another reason as to why it helps to have a co-founder with a complementary skillset to yours. Not only is having similar industry backgrounds and skillsets not particularly useful when building out your startup, it's also more likely to lead to conflicts since you both have similar subject expertise. On the other hand, if your co-founder is an expert in something that you're not, you're less likely to argue with them about their decisions regarding that aspect of the business and vice versa when it comes to your decisions. Look for employees who are driven by your mission, not salary For early-stage startups, the first hires are crucial. These employees need to be highly motivated and excited about the mission. Since the salary will likely be low and the work demanding, they must be driven by something beyond just the paycheck. The right employees are the swash-buckling pirates and romantics, i.e those who are genuinely passionate about the startup’s vision and want to be part of something impactful beyond material gains. When employees are motivated by the mission, they are more likely to stick around and help take the startup to greater heights. A litmus test for hiring: Would you be excited to work with them on a Sunday? One of the most important rounds in the hiring process is the culture fit round. This is where you assess whether a candidate shares the same values as you and your team. A key question to ask yourself is: "Would I be excited to work with this person on a Sunday?" If there’s any doubt about your answer, it’s likely not a good fit. The idea is that you want employees who align with the company's culture and values and who you would enjoy collaborating with even outside of regular work hours. How we structure the team at AI Palette We have three broad functions in our organization. The first two are the big ones: Technical Team – This is the core of our product and technology. This team is responsible for product development and incorporating customer feedback into improving the technology Commercial Team – This includes sales, marketing, customer service, account managers, and so on, handling everything related to business growth and customer relations. General and Administrative Team – This smaller team supports functions like finance, HR, and administration. As with almost all businesses, we have teams that address the two core tasks of building (technical team) and selling (commercial team), but given the size we're at now, having the administrative team helps smoothen operations. Set broad goals but let your teams decide on execution What I've done is recruit highly skilled people who don't need me to micromanage them on a day-to-day basis. They're experts in their roles, and as Steve Jobs said, when you hire the right person, you don't have to tell them what to do—they understand the purpose and tell you what to do. So, my job as the CEO is to set the broader goals for them, review the plans they have to achieve those goals, and periodically check in on progress. For example, if our broad goal is to meet a certain revenue target, I break it down across teams: For the sales team, I’ll look at how they plan to hit that target—how many customers they need to sell to, how many salespeople they need, and what tactics and strategies they plan to use. For the technical team, I’ll evaluate our product offerings—whether they think we need to build new products to attract more customers, and whether they think it's scalable for the number of customers we plan to serve. This way, the entire organization's tasks are cascaded in alignment with our overarching goals, with me setting the direction and leaving the details of execution to the skilled team members that I hire.

[CASE STUDY] From 217/m to $2,836/m in 9 months - Sold for $59,000; I grow and monetise web traffic of 5, 6, 7 figures USD valued passive income content sites [AMA]
reddit
LLM Vibe Score0
Human Vibe Score1
jamesackerman1234This week

[CASE STUDY] From 217/m to $2,836/m in 9 months - Sold for $59,000; I grow and monetise web traffic of 5, 6, 7 figures USD valued passive income content sites [AMA]

Hello Everyone (VERY LONG CASE STUDY AHEAD) - 355% return in 9 months Note: I own a 7-figures USD valued portfolio of 41+ content sites that generates 5-6 figures USD a month in passive income. This is my first time posting in this sub and my goal is to NOT share generic advice but precise numbers, data and highly refined processes so you can also get started with this business yourself or if you already have an existing business, drive huge traffic to it and scale it substantially (get more customers). I will use a case study to explain the whole process. As most of us are entrepreneurs here, explaining an actual project would be more meaningful. In this case study I used AI assisted content to grow an existing site from $217/m to $2,836/m in 9 months (NO BACKLINKS) and sold it for $59,000. ROI of 3 months: 355% Previous case studies (before I give an overview of the model) Amazon Affiliate Content Site: $371/m to $19,263/m in 14 MONTHS - $900K CASE STUDY \[AMA\] Affiliate Website from $267/m to $21,853/m in 19 months (CASE STUDY - Amazon?) \[AMA\] Amazon Affiliate Website from $0 to $7,786/month in 11 months Amazon Affiliate Site from $118/m to $3,103/m in 8 MONTHS (SOLD it for $62,000+) Note: You can check pinned posts on my profile. Do go through the comments as well as a lot of questions are answered in those. However, if you still have any questions, feel free to reach out. This is an \[AMA\]. Quick Overview of the Model Approach: High traffic, niche specific, informative content websites that monetise its traffic through highly automated methods like display ads and affiliate. The same model can be applied to existing businesses to drive traffic and get customers. Main idea: Make passive income in a highly automated way Easy to understand analogy You have real estate (here you have digital asset like a website) You get rental income (here you get ads and affiliate income with no physical hassle, in case you have a business like service, product etc. then you can get customers for that too but if not, it's alright) Real estate has value (this digital asset also has value that can be appreciated with less effort) Real estate can be sold (this can be sold too but faster) IMPORTANT NOTE: Search traffic is the BEST way to reach HUGE target audience and it's important when it comes to scaling. This essentially means that you can either monetise that via affiliate, display etc. or if you have a business then you can reach a bigger audience to scale. Overview of this website's valuation (then and now: Oct. 2022 and June 2023) Oct 2022: $217/m Valuation: $5,750.5 (26.5x) - set it the same as the multiple it was sold for June 2023: $2,836/m Traffic and revenue trend: growing fast Last 3 months avg: $2,223 Valuation now: $59,000 (26.5x) Description: The domain was registered in 2016, it grew and then the project was left unattended. I decided to grow it again using properly planned AI assisted content. Backlink profile: 500+ Referring domains (Ahrefs). Backlinks mean the sites linking back to you. This is important when it comes to ranking. Summary of Results of This Website - Before and After Note: If the terms seem technical, do not worry. I will explain them in detail later. Still if you have any questions. Feel free to comment or reach out. |Metric|Oct 2022|June 2023|Difference|Comments| |:-|:-|:-|:-|:-| |Articles|314|804|\+490|AI Assisted content published in 3 months| |Traffic|9,394|31,972|\+22,578|Organic| |Revenue|$217|$2,836|\+$2,619|Multiple sources| |RPM (revenue/1000 web traffic)|23.09|$88.7|\+$65.61|Result of Conversion rate optimisation (CRO). You make changes to the site for better conversions| |EEAT (expertise, experience, authority and trust of website)|2 main authors|8 authors|6|Tables, video ads and 11 other fixations| |CRO|Nothing|Tables, video ads |Tables, video ads and 11 other fixations || &#x200B; Month by Month Growth |Month|Revenue|Steps| |:-|:-|:-| |Sept 2022|NA|Content Plan| |Oct 2022|$217|Content Production| |Nov 2022|$243|Content production + EEAT authors| |Dec 2022|$320|Content production + EEAT authors| |Jan 2023|$400|Monitoring| |Feb 2023|$223|Content production + EEAT authors| |Mar 2023|$2,128|CRO & Fixations| |April 2023|$1,609|CRO & Fixations| |May 2023|$2,223|Content production + EEAT authors| |June 2023|$2,836|CRO and Fixations| |Total|$10,199|| &#x200B; What will I share Content plan and Website structure Content Writing Content Uploading, formatting and onsite SEO Faster indexing Conversion rate optimisation Guest Posting EEAT (Experience, Expertise, Authority, Trust) Costing ROI The plans moving forward with these sites &#x200B; Website Structure and Content Plan This is probably the most important important part of the whole process. The team spends around a month just to get this right. It's like defining the direction of the project. Description: Complete blueprint of the site's structure in terms of organisation of categories, subcategories and sorting of articles in each one of them. It also includes the essential pages. The sorted articles target main keyword, relevant entities and similar keywords. This has to be highly data driven and we look at over 100 variables just to get it right. It's like beating Google's algorithm to ensure you have a blueprint for a site that will rank. It needs to be done right. If there is a mistake, then even if you do everything right - it's not going to work out and after 8-16 months you will realise that everything went to waste. Process For this project, we had a niche selected already so we didn't need to do a lot of research pertaining to that. We also knew the topic since the website was already getting good traffic on that. We just validated from Ahrefs, SEMRUSH and manual analysis if it would be worth it to move forward with that topic. &#x200B; Find entities related to the topic: We used Ahrefs and InLinks to get an idea about the related entities (topics) to create a proper topical relevance. In order to be certain and have a better idea, we used ChatGPT to find relevant entities as well \> Ahrefs (tool): Enter main keyword in keywords explorer. Check the left pain for popular topics \> Inlinks (tool): Enter the main keyword, check the entity maps \> ChatGPT (tool): Ask it to list down the most important and relevant entities in order of their priority Based on this info, you can map out the most relevant topics that are semantically associated to your main topic Sorting the entities in topics (categories) and subtopics (subcategories): Based on the information above, cluster them properly. The most relevant ones must be grouped together. Each group must be sorted into its relevant category. \> Example: Site about cycling. \> Categories/entities: bicycles, gear and equipment, techniques, safety, routes etc. \> The subcategories/subentities for let's say "techniques" would be: Bike handling, pedaling, drafting etc. Extract keywords for each subcategory/subentity: You can do this using Ahrefs or Semrush. Each keyword would be an article. Ensure that you target the similar keywords in one article. For example: how to ride a bicycle and how can I ride a bicycle will be targeted by one article. Make the more important keyword in terms of volume and difficulty as the main keyword and the other one(s) as secondary Define main focus vs secondary focus: Out of all these categories/entities - there will be one that you would want to dominate in every way. So, focus on just that in the start. This will be your main focus. Try to answer ALL the questions pertaining to that. You can extract the questions using Ahrefs. \> Ahrefs > keywords explorer \> enter keyword \> Questions \> Download the list and cluster the similar ones. This will populate your main focus category/entity and will drive most of the traffic. Now, you need to write in other categories/subentities as well. This is not just important, but crucial to complete the topical map loop. In simple words, if you do this Google sees you as a comprehensive source on the topic - otherwise, it ignores you and you don't get ranked Define the URLs End result: List of all the entities and sub-entities about the main site topic in the form of categories and subcategories respectively. A complete list of ALL the questions about the main focus and at around 10 questions for each one of the subcategories/subentities that are the secondary focus Content Writing So, now that there's a plan. Content needs to be produced. Pick out a keyword (which is going to be a question) and... Answer the question Write about 5 relevant entities Answer 10 relevant questions Write a conclusion Keep the format the same for all the articles. Content Uploading, formatting and onsite SEO Ensure the following is taken care of: H1 Permalink H2s H3s Lists Tables Meta description Socials description Featured image 2 images in text \\Schema Relevant YouTube video (if there is) Note: There are other pointers link internal linking in a semantically relevant way but this should be good to start with. Faster Indexing Indexing means Google has read your page. Ranking only after this step has been done. Otherwise, you can't rank if Google hasn't read the page. Naturally, this is a slow process. But, we expedite it in multiple ways. You can use RankMath to quickly index the content. Since, there are a lot of bulk pages you need a reliable method. Now, this method isn't perfect. But, it's better than most. Use Google Indexing API and developers tools to get indexed. Rank Math plugin is used. I don't want to bore you and write the process here. But, a simple Google search can help you set everything up. Additionally, whenever you post something - there will be an option to INDEX NOW. Just press that and it would be indexed quite fast. Conversion rate optimisation Once you get traffic, try adding tables right after the introduction of an article. These tables would feature a relevant product on Amazon. This step alone increased our earnings significantly. Even though the content is informational and NOT review. This still worked like a charm. Try checking out the top pages every single day in Google analytics and add the table to each one of them. Moreover, we used EZOIC video ads as well. That increased the RPM significantly as well. Both of these steps are highly recommended. Overall, we implemented over 11 fixations but these two contribute the most towards increasing the RPM so I would suggest you stick to these two in the start. Guest Posting We made additional income by selling links on the site as well. However, we were VERY careful about who we offered a backlink to. We didn't entertain any objectionable links. Moreover, we didn't actively reach out to anyone. We had a professional email clearly stated on the website and a particularly designated page for "editorial guidelines" A lot of people reached out to us because of that. As a matter of fact, the guy who bought the website is in the link selling business and plans to use the site primarily for selling links. According to him, he can easily make $4000+ from that alone. Just by replying to the prospects who reached out to us. We didn't allow a lot of people to be published on the site due to strict quality control. However, the new owner is willing to be lenient and cash it out. EEAT (Experience, Expertise, Authority, Trust) This is an important ranking factor. You need to prove on the site that your site has authors that are experienced, have expertise, authority and trust. A lot of people were reaching out to publish on our site and among them were a few established authors as well. We let them publish on our site for free, added them on our official team, connected their socials and shared them on all our socials. In return, we wanted them to write 3 articles each for us and share everything on all the social profiles. You can refer to the tables I shared above to check out the months it was implemented. We added a total of 6 writers (credible authors). Their articles were featured on the homepage and so were their profiles. Costing Well, we already had the site and the backlinks on it. Referring domains (backlinks) were already 500+. We just needed to focus on smart content and content. Here is the summary of the costs involved. Articles: 490 Avg word count per article: 1500 Total words: 735,000 (approximately) Cost per word: 2 cents (includes research, entities, production, quality assurance, uploading, formatting, adding images, featured image, alt texts, onsite SEO, publishing/scheduling etc.) Total: $14,700 ROI (Return on investment) Earning: Oct 22 - June 23 Earnings: $10,199 Sold for: $59,000 Total: $69,199 Expenses: Content: $14,700 Misc (hosting and others): $500 Total: $15,200 ROI over a 9 months period: 355.25% The plans moving forward This website was a part of a research and development experiment we did. With AI, we wanted to test new waters and transition more towards automation. Ideally, we want to use ChatGPT or some other API to produce these articles and bulk publish on the site. The costs with this approach are going to be much lower and the ROI is much more impressive. It's not the the 7-figures projects I created earlier (as you may have checked the older case studies on my profile), but it's highly scalable. We plan to refine this model even further, test more and automate everything completely to bring down our costs significantly. Once we have a model, we are going to scale it to 100s of sites. The process of my existing 7-figures websites portfolio was quite similar. I tested out a few sites, refined the model and scaled it to over 41 sites. Now, the fundamentals are the same however, we are using AI in a smarter way to do the same but at a lower cost, with a smaller team and much better returns. The best thing in my opinion is to run numerous experiments now. Our experimentation was slowed down a lot in the past since we couldn't write using AI but now it's much faster. The costs are 3-6 times lower so when it used to take $50-100k to start, grow and sell a site. Now you can pump 3-6 more sites for the same budget. This is a good news for existing business owners as well who want to grow their brand. Anyway, I am excited to see the results of more sites. In the meantime, if you have any questions - feel free to let me know. Best of luck for everything. Feel free to ask questions. I'd be happy to help. This is an AMA.

Thoughts on FasterCapital VC?
reddit
LLM Vibe Score0
Human Vibe Score1
Momof3rascalsThis week

Thoughts on FasterCapital VC?

TLDR: I pitched to FasterCapital and got an "offer". Trying to figure out if this is a legitimate opportunity or a waste of my time. I'm not familiar with VCs and hadn't considered actually getting an investor on board with my plan. I sent my pitch deck to FasterCapital, honestly not expecting a response. It was my first pitch deck and a complete long shot. I ended up getting a response, they asked me for clarification on a few things. Than I get this email about what they are offering here's the main part We specialize in warm introductions to angel investors, VCs, and HNWIs, ensuring you connect with the right investors through personalized recommendations—not ineffective mass email campaigns. Cold outreach, such as LinkedIn messages, rarely succeeds, as investors receive hundreds of such requests and disregard them. To raise money, you need a strong partner like ourselves who has a wide network and direct connection with those angel investors built throughout 10 years. You can see some of the reviews of the startups we have helped attached and reviews on independent sites. Based on our experience and the matching that we have done already on our own AI system and for raising $55M-$65M in 5 years, a suitable package in your case is $50k - $64k and the chances of raising money is %87 - %93, but you were accepted in the exceptional rising star offer, where you pay half of that amount as an advance which is $25k-$32k and the other half ONLY when we raise you the first $1M. Other startups in our standard offers pays double that amount. First, I don't understand all of it, except for the "where you pay half of that amount as an advance which is $25k-$32k" I am no where near being able to come close to that, mostly because if I had that much, I wouldn't apply to a VC. I responded and politely told her that was not something our company could financially do right now. Than this email Thanks for your kind reply. We are flexible on paying this amount into monthly installments. We offer money back guarantee if we didn't raise the capital in 6 months from signing. This is how much we are confident with our approach of warm introductions. Raising the first amount of money and getting the first investor onboard is the most challenging part. You need time to build trust and network of investors. You need to have a good partner to help you. Please note that the down payment is for raising at least $55M over five years as we are interested in long-term partnership to raise multiple rounds because we make money through the commission. Companies take only commission or success fee are doing cold introductions and mass emails and this approach has low chances of success when it comes to raising capital. It is about the chances of success. You can talk to these companies and ask them about their success rate. Mass emails campaign has zero chances of success.  We have helped more than 742 startups raise more than $2.2B. Our network includes 155,000 angel investors and more than 50K funding institutions (VCs, HNI, family offices..etc). We have been in this business for more than 10 years. We have more than 92% success rate in our program so far. So if you are familiar with VC, Is this an actual opportunity. I have a tendency to jump or dive head first into things. As much as I want to get excited because this would be the jumpstart to most of my goals and ambitions. I'm not familiar with VCs. I have bootstrapped all my ventures so far.

Follow Along as I Flip this Website - Case Study
reddit
LLM Vibe Score0
Human Vibe Score1
jshogren10This week

Follow Along as I Flip this Website - Case Study

I am starting a new case study where I will be documenting my attempt to flip a website that I just purchased from Flippa. However, unlike most case studies where people hide certain parts and details from the public I will instead be sharing everything. That means you will know the exact URL of the site that I purchased and I will share everything with you all as I progress.I know that case studies are lot more interesting and you can learn better when you can see real examples of what I am talking about. Enough of the chatting, let's jump straight into this new case study and I will explain what this is all about. Before you get into the case study I want to give you the option of reading this one my website where all of the images can be seen within the post and it is easier to read. I also want to say that I have nothing to sell you or anything close to it. So if you want to read it there you can do so here ##Introductory Video I have put together a video that talks about many of the things that I cover in this article. So if you would rather watch a video you can watch that here - https://www.youtube.com/watch?v=EE3SxtNnqts However, I go into more detail in the actual article FYI. Also, I plan on using Youtube very frequently in this case study so be on the lookout for new videos.There is going to be a video that will accompany every single case study post because I like having it being presented in two different mediums. ##The Website I Just Bought Around a week ago I made a new website purchase from Flippa and you can view the website's Flippa listing here - https://flippa.com/6439965-hvactraining101-com Screenshot of the Homepage - http://imgur.com/T6Iv1QN I paid $1,250 for the site and you will soon see that I got a really good deal. As you might be able to tell from the URL, this site is focused around training and education for becoming a HVAC technician. This is a lucrative niche to be in and Adsense pays very well. I do not have control of the site yet due to the transfer process not being completed. However, I am hoping within a few days everything will be finalized and I will take full control of the site. In the meantime, I figured it would be a good time to put together the introduction post for this new case study! ##Why I Bought this Website Now that you have a general idea of the website that I purchased, I now want to explain the reasoning behind the purchase. There are 3 major reasons for this purchase and I will explain each one of them below. GREAT Price As I mentioned earlier, I bought this website for $1,250. However, that doesn't mean a whole lot unless you know how much the site is making each month. Screenshot of the earnings for the last 12 months - http://imgur.com/NptxCHy Average Monthly Profits: 3 Month = $126 6 Month = $128 12 Month = $229.50 Let's use the 6 month average of $128/month as our baseline average. Since it is making on average $128/month and it was sold for $1,250 then that means I bought this site at a multiple of 9.76x! Most sites in today's market go for 20x-30x multiples. As you can see, I got a great deal on this site. Although the great price was the biggest reason for me buying this site there are other factors that persuaded me as well. You need to remember that just because you can get a website for a good price it doesn't mean it is a good deal. There are other factors that you need to look at as well. Extremely Under Optimized This site is currently being monetized mainly by Adsense and a very small amount from Quinstreet. From my experience with testing and optimizing Adsense layouts for my site in my Website Investing case study I know the common ad layouts that work best for maximizing Adsense revenue. With that being said, I can quickly determine if a website is being under optimized in terms of the ad layout. One of the first things I did when analyzing this site was examine the ad layout it was using. Screenshot of the website with the ad layout the previous owner was using - http://imgur.com/wqleLVA There is only ONE ad per page being used, that's it. Google allows up to 6 total ads to be used per page and you can imagine how much money is being left on the table because of this. I am estimating that I can probably double the earnings for the site practically overnight once I add more ads to the site. Adding more ads in combination with my favorite Adsense plugin, AmpedSense, I will be able to easily boost the earnings for this site quickly. It is also worth mentioning how lucrative this niche is and how much advertisers are willing to spend on a per click basis. The average CPC for the top keywords this site is currently ranking for in Google - http://imgur.com/ifxiy8B Look at those average CPC numbers, they are insanely high! I could be making up to $25 per click for some of those keywords, which is so absurd to me. Combine these extremely high CPC with the fact that the site currently only has one ad per page and you can start to understand just how under optimized this site truly is. I also plan on utilizing other ad networks such as Quinstreet and Campus Explorer more as well. These two networks are targeted at the education niche which works very well with my site. I will be testing to see if these convert better than normal Adsense ads. Goldmine of Untapped Keywords One of the biggest opportunities I see for growing this site is to target local keywords related to HVAC training. As of right now, the site has only scratched the surface when it comes to trying to rank for state/city keywords. Currently there are only two pages on the entire website which go after local keywords, those two pages target Texas and Florida HVAC search terms. These two pages are two of the more popular pages in terms of total amount of traffic. See the screenshot of the Google Analytics - http://imgur.com/NB0xJ4G Two out of the top five most popular pages for the entire website are focused on local search terms. However, these are the ONLY two pages that target local search terms on the whole site! There are 48 other states, although there may not be search volume for all states, and countless cities that are not being targeted. Why do I think this is such a good opportunity? For a few reasons: Local keywords are a lot easier to rank for in Google than more general keywords This site has been able to rank for two states successfully already and it proves it is possible Traffic going to these local pages is WAY more targeted and will convert at a much higher rate, which means more commissions for me There are so many more states and cities that get a good amount of searches that I can target To give you an idea of the type of keywords these local pages rank for, you can see the top keywords that the Florida page is ranking for in Google: Top ranking keywords for the Florida page - http://imgur.com/j7uKzl2 As you can see these keywords don't get a ton of searches each month, but ranking 1st for a keyword getting 90 searches a month is better than being ranked 10th for a keyword getting 1,000 searches a month. I have started to do some keyword research for other states and I am liking what I am finding so far. Keywords that I have found which I will be targeting with future articles - http://imgur.com/8CCCCWU I will go into more detail about my keyword research in future articles, but I wanted to give you an idea of what my strategy will be! I also wanted to share why I am super excited about the future potential to grow this site by targeting local keywords. ##Risks Yes, there are many good things about this website, but there are always risks involved no matter what the investment is. The same thing goes for this site. Below are some of the risks that I currently see. HTML Site This website is a HTML site and I will need to transfer it to Wordpress ASAP. I have been doing some research on this process and it shouldn't be too hard to get this over to Wordpress. In doing so it will make adding content, managing the back end and just about everything else easier. Also, I am hoping that when I transfer it to Wordpress that it will become more optimized for Google which will increase keyword rankings. Declining Earnings Looking at the last 12 months of earnings you will notice a drop off from last year till now. Earnings from the last 12 months - http://imgur.com/WsotZsj In May of 2015 it looks like the site earned right around $500, which is much higher than the $128 that it is earning now. However, the last 7 or so months have been consistent which is a good sign. Even though the earnings are much lower now then they were a year ago it is good to know that this site has the potential to earn $500/month because it has done it before. Slightly Declining Traffic In the last 12 months the site's traffic has declined, however, it looks like it is picking back up. Traffic from the last 12 months - http://imgur.com/aiYZW9W The decline is nothing serious, but there is a drop on traffic. Let's take a look at the complete history of this site's traffic so we can get a better idea of what is going on here: Complete traffic history - http://imgur.com/tYmboVn The above screenshot is from 2012 all the way up to right now. In the grand scheme of things you can see that the traffic is still doing well and it looks like it is on the upswing now. Those three risks mentioned above are the three biggest risks with this site at this point. It is always good to note the risks and do everything you can to prevent them from causing a problem. ##My Growth Strategy Whenever I purchase a new site I always create an outline or plan on how I will grow the site. Right now, I have some basic ideas on how I will grow this site, but as I go on I will continue to change and optimize my strategies to be more effective. Below I have outlined my current plans to grow: Add more Adsense Ads The very first thing I will do once I get control of the site is add more ads per page. I am predicting that by just adding a few more ads per page I will be able to more than likely double the earnings. I will touch on exactly how I will be optimizing the ad layouts in future posts. Test other Ad Networks I will be doing a lot of testing and experimenting when it comes to the ad networks. I plan on trying out Adsense, Media.net, Quinstreet, Campus Explorer and finding the combination of those 4 which produces the most revenue. The Adsense and Media.net ads will perform well on the more general pages while Quinstreet and Campus Explorer ads will be geared towards the local search terms. There will probably be other ad networks I will try out but these are the four which I will be using right away. If you are aware of any other ad networks out there which are geared towards the education niche please let me know in the comments below! Target Local Keywords with new Content I have already touched on this, but I will starting to produce content targeting these local keywords ASAP. The sooner I add the content to the site the sooner it will start to rank and bring in traffic. I will not be writing my own content and instead I will be outsourcing all of it via Upwork. I will show you all how I go about outsourcing content production and you can see my process for doing that. ##Goals for this Website My goal for the website is to have it valued at $10,000+ within 12 months. Let's break down this larger goal into smaller chunks which will make achieving it easier and more attainable. Earnings - $500/month To get the site valued at $10,000 the site will need to be making $500/month using a 20x monthly multiple. Right now, the site is making around $130/month so it has a ways to before it reaches the $500 a month mark. However, after doing some Adsense optimization I think we could push the earnings to around $300/month without much work. From there, it will come down to trying to bring in more traffic! Traffic - 5,000 Visitors per Month Why 5,000 visitors? Because that is how much traffic it is going to take to get to the $500/month goal. Let me explain how I came to this conclusion: The average RPM for this site is currently $50, which means for every 1,000 page views the site earns $50. After I optimize the Adsense layout for the site and add more ads per page I think I will be able to double the RPM to $100. Using the RPM of $100 the site will need to have 5,000 monthly visitors to earn $500. So 5,000 monthly visitors is the traffic goal I have set and aiming for! The site is currently getting around 3,000 visitors per month so I will need to add an extra 2,000 visitors to get to this goal. ##Want to Follow this Case Study? I will be using Youtube a lot in this case study so make sure to follow my Youtube channel here - www.youtube.com/c/joshshogren Other than that, I think that is going to bring us to the end of the introductory post for this new case study. I hope that you enjoyed reading and that you are excited to follow along! If you have any suggestions to make this case study better PLEASE let me know in the comment below. I want to make this case study the best one I have done yet. Talk to you all in the comment section.

Is there any point in building a product with AI anymore?
reddit
LLM Vibe Score0
Human Vibe Score1
jottrledThis week

Is there any point in building a product with AI anymore?

Everyone and their grandmother are building AI products. So it begs the question. Is the AI market now too saturated? Have all the AI apps been thought of? Of course not don't be silly, There is still a significant opportunity to create AI products and become profitable. But let's play devils advocate here. Let's say you're a developer and you want to build your first SaaS product. Now, imagine a world where all AI products were already thought of (a scary thought). What would you do? Would you move onto something else? See, when people think of an idea for a SaaS product what often happens is they do a quick Google search and tend to think "oh crap, it already exists, better move on". Maybe that's the right thing to do...but then again maybe it's not. Before you run for the hills, make sure to check the SEO potential for your idea. If your idea has the potential to rank high on Google and there are already hundreds/thousands of people looking for it then you can take that as all the validation you need to start building it. Here are 3 AI ideas that all have good SEO potential. Each idea has keywords that you can target with a difficulty level 500. This means it's easy to rank high in Google for them and they have a high number of people searching for them each month. AI Accounting Software A Saas product that uses AI to analyze bank transactions, invoices, and receipts to automatically match them and reconcile accounts in real-time, reducing manual work and errors. It would also offer predictive insights, suggesting optimal payment times or highlighting potential cash flow issues based on historical data. Could potentially be integrated with popular accounting software like QuickBooks or Xero. SEO Potential Keyword: ai accounting Keyword Difficulty: 9 Average Search Volume 2900 AI Human Resources Software AI Human Resources Software An AI-driven candidate screening and onboarding platform for small to medium-sized businesses. The tool would use AI to automatically filter job applications based on predefined criteria, rank candidates, and even conduct initial interview assessments using natural language processing. It could also manage onboarding tasks by automating the distribution of paperwork, training schedules, and team introductions. SEO Potential Keyword: ai human resources Keyword Difficulty: 17 Average Search Volume 2900 AI Nutrition Tool A Micro SaaS which creates personalized meal planning and nutrition analysis. The platform would use AI to create tailored meal plans based on users' dietary goals, preferences, allergies, and health data (such as activity level or medical conditions). It could analyze food labels, suggest healthier alternatives, and track nutrient intake in real time, helping users maintain balanced diets. SEO Potential Keyword: ai nutrition Keyword Difficulty: 3 Average Search Volume 720 I created a tool (check the first comment) to find ideas like this.

In 2018, I started an AI chatbot company...today, we have over 4000 paying customers and ChatGPT is changing EVERYTHING
reddit
LLM Vibe Score0
Human Vibe Score1
Millionaire_This week

In 2018, I started an AI chatbot company...today, we have over 4000 paying customers and ChatGPT is changing EVERYTHING

Intro: 5 years ago, my co-founders and I ventured into the space of AI chatbots and started our first truly successful company. Never in a million years did I see myself in this business and we truly stumbled upon the opportunity by chance. Prior to that, we ran a successful lead generation business and questioned whether a simple ai chat product would increase our online conversions. Of the 3 co-founders, I was skeptical that it would, but the data was clear that we had something that really worked. We built a really simple MVP version of the product and gave it to some of our top lead buyers who saw even better conversion improvements on their own websites. In just a matter of weeks, a new business opportunity was born and a major pivot away from our lead generation business started. Our growth story: Startup growth is really interesting and in most cases, founders aren't really educated on what a typical growth curve looks like. While we hear about "hockey stick" growth curves, it's really atypical to actually see or experience this. From my experience, growth curves take place in a "stair curve". For example, you can scrap your way to a $100k run rate without much process or tracking. You can even get to $1 million ARR being super disorganized. As you start going beyond $1M ARR, things start to break and growth can flatten out while you put new processes and systems in place. Eventually you'll get to $2M or 3M with your new strategy and then things start breaking again. I've seen the process repeat itself and as you increase your ARR, the processes and systems become more difficult to work through...mainly because more people get involved and the product becomes more complex. When you do end up cracking the code in each step, the growth accelerates faster and faster before things start to break down and flatten out again. Without getting too much into the numbers, here were some of our initial levers for growth: Our first "stair" step was to leverage our existing customer base from our prior lead generation business. Having prior business relationships and a proven track record made it really simple to have conversations with people who already trusted us to try something new that we had to offer. Stair #2 was to build out a partner channel. Since our chat product involved a web developer or agency installing the chat on client sites, we partnered with these developers and agencies to leverage their already existing customer bases. We essentially piggy-backed off of their relationships and gave them a cut of the revenue. We built an internal partner tracking portal which took 6+ months, but it was well worth it. Stair #3 was our most expensive step, biggest headache, but added the most revenue. After COVID, we had and SDR/Account Executive sales team of roughly 30 people. It added revenue fast, but the payback periods were 12+ months so we had to cut back on this strategy after exhausting our universe of clients. Stair #4 involves a variety of paid advertisement strategies with product changes and the introduction of new onboarding features. We're in the middle of this stair and hope it's multiple years before things breakdown again. Don't give up I know it sounds really cliché, but the #1 indicator of success is doing the really boring stuff day in and day out and making incremental improvements. As the weeks, months, and years pass by, you will slowly gain domain expertise and start to see the gaps in the market that can set you apart from your competition. It's so hard for founders to stay focused and not get distracted so I would say it's equally as important to have co-founders who hold each other accountable on what your collective goals are. How GPT is changing everything I could write pages and pages about how GPT is going to change how the world operates, but I'll keep it specific to our business and chatbots. In 2021, we built an industry specific AI model that did a great job of classifying intents which allowed us to train future actions during a chat. It was a great advancement in our customer's industry at the time. With GPT integrated into our system, that training process that would take an employee hours to do, can be done in 5 minutes. The model is also cheaper than our own and more accurate. Because of these training improvements, we have been able to conduct research that is allowing us to leverage GPT models like no one else in the industry. This is both in the realm of chat and also training during onboarding. I really want to refrain from sharing our company, but if you are interested in seeing a model trained for your specific company or website, just PM me your link and I'll send you a free testing link with a model fully trained for your site to play around with. Where we are headed and the dangers of AI The level of advancement in AI is not terribly dangerous in its current state. I'm sure you've heard it before, but those who leverage the technology today will be the ones who get ahead. In the coming years, AI will inevitably replace a large percentage of human labor. This will be great for overall value creation and productivity for the world, but the argument that humans have always adapted and new jobs will be created is sadly not going to be as relevant in this case. As the possibility of AGI becomes a reality in the coming years or decades, productivity through AI will be off the charts. There is a major risk that human innovation and creative thinking will be completely stalled...human potential as we know it will be capped off and there will need to be major economic reform for displaced workers. This may not happen in the next 5 or 10 years, but you would be naïve not to believe the world we live in today will not be completely different in 20 to 30 years. Using AI to create deepfakes, fake voice agents, scam the unsuspecting, or exploit technical vulnerabilities are just a few other examples I could write about, but don't want to go into to much detail for obvious reasons. Concluding If you found the post interesting or you have any questions, please don't hesitate to ask. I'll do my best to answer whatever questions come from this! &#x200B; \*EDIT: Wasn't expecting this sort of response. I posted this right before I went to sleep so I'll get to responding soon.

We made $325k in 2023 from AI products, starting from 0, with no-code, no funding and no audience
reddit
LLM Vibe Score0
Human Vibe Score1
hopefully_usefulThis week

We made $325k in 2023 from AI products, starting from 0, with no-code, no funding and no audience

I met my co-founder in late 2022 after an introduction from a mutual friend to talk about how to find contract Product Management roles. I was sporadically contracting at start-up at the time and he had just come out of another start-up that was wiped out by the pandemic. We hit it off, talking about ideas, sharing what other indie-hackers were doing, and given GPT-3’s prominence at the time, we started throwing around ideas about things we could build with it, if nothing else, just to learn. I should caveat, neither of us were AI experts when starting out, everything we learned has been through Twitter and blogs, my background is as an accountant, and his a consultant. Here’s how it went since then: &#x200B; Nov 2022 (+$50) \- We built a simple tool in around a week using GPT-3 fine-tuning and a no-code tool (Bubble) that helped UK university students write their personal statements for their applications \- We set some Google Ads going and managed to make a few sales (\~$50) in the first week \- OpenAI were still approving applications at the time and said this went against their “ethics” so we had to take it down &#x200B; Dec 2022 (+$200) \- We couldn’t stop coming up with ideas related to AI fine-tuning, but realised it was almost impossible to decide which to pursue \- We needed a deadline to force us so we signed up for the Ben’s Bites hackathon in late December \- In a week, we built and launched a no-code fine-tuning platform, allowing people to create fine-tuned models by dragging and dropping an Excel file onto it \- We launched it on Product Hunt, having no idea how to price it, and somehow managed to get \~2,000 visitors on the site and make 2 sales at $99 &#x200B; Jan 2023 (+$3,000) \- We doubled down on the fine-tuning idea and managed to get up to \~$300 MRR, plus a bunch of one-time sales and a few paid calls to help people get the most out of their models \- We quickly realised that people didn’t want to curate models themselves, they just wanted to dump data and get magic out \- That was when we saw people building “Talk with x book/podcast” on Twitter as side projects and realised that was the missing piece, we needed to turn it into a tool \- We started working on the new product in late January &#x200B; Feb 2023 (+$9,000) \- We started pre-selling access to an MVP for the new product, which allowed people to “chat with their data/content”, we got $5,000 in pre-sales, more than we made from the previous product in total \- By mid-February, after 3 weeks of building we were able to launch and immediately managed to get traction, getting to $1k MRR in < 1 week, building on the hype of ChatGPT and AI (we were very lucky here) &#x200B; Mar - Jul 2023 (+$98,000) \- We worked all the waking hours to keep up with customer demand, bugs, OpenAI issues \- We built integrations for a bunch of services like Slack, Teams, Wordpress etc, added tons of new functionality and continue talking to customers every day \- We managed to grow to $17k MRR (just about enough to cover our living expenses and costs in London) through building in public on Twitter, newsletters and AI directories (and a million other little things) \- We sold our fine-tuning platform for \~$20k and our university project for \~$3k on Acquire &#x200B; Aug 2023 (+$100,000) \- We did some custom development work based on our own product for a customer that proved pretty lucrative &#x200B; Sep - Oct 2023 (+$62,000) \- After 8 months of building constantly, we started digging more seriously into our usage and saw subscriptions plateauing \- We talked to and analysed all our paying users to identify the main use cases and found 75% were for SaaS customer support \- We took the leap to completely rebuild a version of our product around this use case, our biggest to date (especially given most features with no-code took us <1 day) &#x200B; Nov - Dec 2023 (+$53,000) \- We picked up some small custom development work that utilised our own tech \- We’re sitting at around $22k MRR now with a few bigger clients signed up and coming soon \- After 2 months of building and talking to users, we managed to finish our “v2” of our product, focussed squarely on SaaS customer support and launched it today. &#x200B; We have no idea what the response will be to this new version, but we’re pretty happy with it, but couldn’t have planned anything that happened to us in 2023 so who knows what will come of 2024, we just know that we are going to be learning a ton more. &#x200B; Overall, it is probably the most I have had to think in my life - other jobs you can zone out from time to time or rely on someone else if you aren’t feeling it - not when you are doing this, case and point, I am writing this with a banging head-cold right now, but wanted to get this done. A few more things we have learned along the way - context switching is unreal, as is keeping up with, learning and reacting to AI. There isn’t a moment of the day I am not thinking about what we do next. But while in some way we now have hundreds of bosses (our customers) I still haven’t felt this free and can’t imagine ever going back to work for someone else. Next year we’re really hoping to figure out some repeatable distribution channels and personally, I want to get a lot better at creating content/writing, this is a first step! Hope this helps someone else reading this to just try starting something and see what happens.

Started a content marketing agency 6 years ago - $0 to $5,974,324 (2023 update)
reddit
LLM Vibe Score0
Human Vibe Score1
mr_t_forhireThis week

Started a content marketing agency 6 years ago - $0 to $5,974,324 (2023 update)

Hey friends, My name is Tyler and for the past 6 years, I’ve been documenting my experience building a content marketing agency called Optimist. Year 1 - 0 to $500k ARR Year 2 - $500k to $1MM ARR Year 3 - $1MM ARR to $1.5MM(ish) ARR Year 4 - $3,333,686 Revenue Year 5 - $4,539,659 Revenue How Optimist Works First, an overview/recap of the Optimist business model: We operate as a “collective” of full time/professional freelancers Everyone aside from me is a contractor Entirely remote/distributed team Each freelancer earns $65-85/hour Clients pay us a flat monthly fee for full-service content marketing (research, strategy, writing, editing, design/photography, reporting and analytics, targeted linkbuilding, and more) We recently introduced hourly engagements for clients who fit our model but have some existing in-house support Packages range in price from $10-20k/mo We offer profit share to everyone on our core team as a way to give everyone ownership in the company In 2022, we posted $1,434,665 in revenue. It was our highest revenue year to date and brings our lifetime total to $5,974,324. Here’s our monthly revenue from January 2017 to December of 2022. But, like every year, it was a mix of ups and downs. Here’s my dispatch for 2023. — Running a business is like spilling a drink. It starts as a small and simple thing. But, if you don’t clean it up, the spill will spread and grow — taking up more space, seeping into every crack. There’s always something you could be doing. Marketing you could be working on. Pitches you could be making. Networking you could be doing. Client work you could help with. It can be all-consuming. And it will be — if you don’t clean up the spill. I realized this year that I had no containment for the spill that I created. Running an agency was spilling over into nearly every moment of my life. When I wasn’t working, I was thinking about work. When I wasn’t thinking about work, I was dreaming about it. Over the years, I’ve shared about a lot of my personal feelings and experience as an entrepreneur. And I also discussed my reckoning with the limitations of running the business we’ve built. My acceptance that it was an airplane but not a rocket. And my plan to try to compartmentalize the agency to make room in my life for other things — new business ideas, new revenue streams, and maybe some non-income-producing activity. 🤷 What I found in 2022 was that the business wasn’t quite ready for me to make that move. It was still sucking up too much of my time and attention. There were still too many gaps to fill and I was the one who was often filling them. So what do you do? Ultimately you have two choices on the table anytime you run a business and it’s not going the way you want it: Walk away Turn the ship — slowly For a huge number of reasons (personal, professional, financial, etc), walking away from Optimist was not really even an option or the right move for me. But it did feel like things needed to change. I needed to keep turning the ship to get it to the place where it fit into my life — instead of my life fitting around the business. This means 2022 was a year of transition for the agency. (Again?) Refocusing on Profit Some money is better than no money. Right? Oddly, this was one of the questions I found myself asking in 2022. Over the years, we’ve been fortunate to have many clients who have stuck with us a long time. In some cases, we’ve had clients work with us for 2, 3, or even 4 years. (That’s over half of our existence!) But, things have gotten more expensive — we’ve all felt it. We’ve had to increase pay to remain competitive for top talent. Software costs have gone up. It’s eaten into our margin. Because of our increasing costs and evolving scope, many of our best, most loyal clients were our least profitable. In fact, many were barely profitable — if at all. We’ve tried to combat that by increasing rates on new, incoming clients to reflect our new costs and try to make up for shrinking margin on long-term clients. But we didn’t have a good strategy in place for updating pricing for current clients. And it bit us in the ass. Subsidizing lower-profit, long-term clients with new, higher-margin clients ultimately didn’t work out. Our margins continued to dwindle and some months we were barely breaking even while posting six-figures of monthly revenue. 2022 was our highest revenue year but one of our least profitable. It only left one option. We had to raise rates on some of our long-term clients. But, of course, raising rates on a great, long-term client can be delicate. You’ve built a relationship with these people over the years and you’re setting yourself up for an ultimatum — are you more valuable to the client or is the client more valuable to you? Who will blink first? We offered all of these clients the opportunity to move to updated pricing. Unfortunately, some of them weren’t on board. Again, we had 2 options: Keep them at a low/no profit rate Let them churn It seems intuitive that having a low-profit client is better than having no client. But we’ve learned an important lesson many times over the years. Our business doesn’t scale infinitely and we can only handle so many clients at a time. That means that low-profit clients are actually costing us money in some cases. Say our average client generates $2,500 per month in profit — $30,000 per year. If one of our clients is only generating $500/mo in profit, working with them means missing out on bringing on a more profitable client (assuming our team is currently at capacity). Instead of $30,000/year, we’re only making $6,000. Keeping that client costs us $24,000. That’s called opportunity cost. So it’s clear: We had to let these clients churn. We decided to churn about 25% of our existing clients. On paper, the math made sense. And we had a pretty consistent flow of new opportunities coming our way. At the time, it felt like a no-brainer decision. And I felt confident that we could quickly replace these low-profit clients with higher-margin ones. I was wrong. Eating Shit Right after we initiated proactively churning some of our clients, other clients — ones we planned to keep — gave us notice that they were planning to end the engagement. Ouch. Fuck. We went from a 25% planned drop in revenue to a nearly 40% cliff staring us right in the face. Then things got even worse. Around Q3 of this year, talk of recession and layoffs really started to intensify. We work primarily with tech companies and startups. And these were the areas most heavily impacted by the economic news. Venture funding was drying up. Our leads started to slow down. This put us in a tough position. Looking back now, I think it’s clear that I made the wrong decision. We went about this process in the wrong way. The reality sinks in when you consider the imbalance between losing a client and gaining a client. It takes 30 days for someone to fire us. It’s a light switch. But it could take 1-3 months to qualify, close, and onboard a new client. We have lots of upfront work, research, and planning that goes into the process. We have to learn a new brand voice, tone, and style. It’s a marathon. So, for every client we “trade”, there’s a lapse in revenue and work. This means that, in retrospect, I would probably have made this transition using some kind of staggered schedule rather than a cut-and-dry approach. We could have gradually off-boarded clients when we had more definitive work to replace them. I was too confident. But that’s a lesson I had to learn the hard way. Rebuilding & Resetting Most of the voluntary and involuntary churn happened toward the end of 2022. So we’re still dealing with the fall out. Right now, it feels like a period of rebuilding. We didn’t quite lose 50% of our revenue, but we definitely saw a big hit heading into 2023. To be transparent: It sucks. It feels like a gigantic mistake that I made which set us back significantly from our previous high point. I acted rashly and it cost us a lot of money — at least on the surface. But I remind myself of the situation we were in previously. Nearly twice the revenue but struggling to maintain profitability. Would it have been better to try to slowly fix that situation and battle through months of loss or barely-break-even profits? Or was ripping off the bandaid the right move after all? I’m an optimist. (Heh, heh) Plus, I know that spiraling over past decisions won’t change them or help me move forward. So I’m choosing to look at this as an opportunity — to rebuild, reset, and refocus the company. I get to take all of the tough lessons I’ve learned over the last 6 years and apply them to build the company in a way that better aligns with our new and current goals. It’s not quite a fresh, clean start, but by parting ways with some of our oldest clients, we’ve eliminated some of the “debt” that’s accumulated over the years. We get a chance to fully realize the new positioning that we rolled out last year. Many of those long-term clients who churned had a scope of work or engagement structure that didn’t fit with our new positioning and focus. So, by losing them, we’re able to completely close up shop on the SOWs that no longer align with the future version of Optimist. Our smaller roster of clients is a better fit for that future. My job is to protect that positioning by ensuring that while we’re rebuilding our new roster of clients we don’t get desperate. We maintain the qualifications we set out for future clients and only take on work that fits. How’s that for seeing the upside? Some other upside from the situation is that we got an opportunity to ask for candid feedback from clients who were leaving. We asked for insight about their decision, what factors they considered, how they perceived us, and the value of our work. Some of the reasons clients left were obvious and possibly unavoidable. Things like budget cuts, insourcing, and uncertainty about the economy all played at least some part of these decisions. But, reading between the lines, where was one key insight that really struck me. It’s one of those, “oh, yeah — duh — I already knew that,” things that can be difficult to learn and easy to forget…. We’re in the Relationship Business (Plan Accordingly) For all of our focus on things like rankings, keywords, content, conversions, and a buffet of relevant metrics, it can be easy to lose the forest for the trees. Yes, the work itself matters. Yes, the outcomes — the metrics — matter. But sometimes the relationship matters more. When you’re running an agency, you can live or die by someone just liking you. Admittedly, this feels totally unfair. It opens up all kinds of dilemmas, frustration, opportunity for bias and prejudice, and other general messiness. But it’s the real world. If a client doesn’t enjoy working with us — even if for purely personal reasons — they could easily have the power to end of engagement, regardless of how well we did our actual job. We found some evidence of this in the offboarding conversations we had with clients. In some cases, we had clients who we had driven triple- and quadruple-digital growth. Our work was clearly moving the needle and generating positive ROI and we had the data to prove it. But they decided to “take things in another direction” regardless. And when we asked about why they made the decision, it was clear that it was more about the working relationship than anything we could have improved about the service itself. The inverse is also often true. Our best clients have lasting relationships with our team. The work is important — and they want results. But even if things aren’t quite going according to plan, they’re patient and quick to forgive. Those relationships feel solid — unshakeable. Many of these folks move onto new roles or new companies and quickly look for an opportunity to work with us again. On both sides, relationships are often more important than the work itself. We’ve already established that we’re not building a business that will scale in a massive way. Optimist will always be a small, boutique service firm. We don’t need 100 new leads per month We need a small, steady roster of clients who are a great fit for the work we do and the value we create. We want them to stick around. We want to be their long-term partner. I’m not built for churn-and-burn agency life. And neither is the business. When I look at things through this lens, I realize how much I can cut from our overall business strategy. We don’t need an ultra-sophisticated, multi-channel marketing strategy. We just need strong relationships — enough of them to make our business work. There are a few key things we can take away from this as a matter of business strategy: Put most of our effort into building and strengthening relationships with our existing clients Be intentional about establishing a strong relationship with new clients as part of onboarding Focus on relationships as the main driver of future business development Embracing Reality: Theory vs Practice Okay, so with the big learnings out the way, I want to pivot into another key lesson from 2022. It’s the importance of understanding theory vs practice — specifically when it comes to thinking about time, work, and life. It all started when I was considering how to best structure my days and weeks around running Optimist, my other ventures, and my life goals outside of work. Over the years, I’ve dabbled in many different ways to block time and find focus — to compartmentalize all of the things that are spinning and need my attention. As I mapped this out, I realized that I often tried to spread myself too thin throughout the week. Not just that I was trying to do too much but that I was spreading that work into too many small chunks rather than carving out time for focus. In theory, 5 hours is 5 hours. If you have 5 hours of work to get done, you just fit into your schedule whenever you have an open time slot. In reality, a single 5-hour block of work is 10x more productive and satisfying than 10, 30-minute blocks of work spread out across the week. In part, this is because of context switching. Turning your focus from one thing to another thing takes time. Achieving flow and focus takes time. And the more you jump from one project to another, the more time you “lose” to switching. This is insightful for me both in the context of work and planning my day, but also thinking about my life outside of Optimist. One of my personal goals is to put a finite limit on my work time and give myself more freedom. I can structure that in many different ways. Is it better to work 5 days a week but log off 1 hour early each day? Or should I try to fit more hours into each workday so I can take a full day off? Of course, it’s the latter. Both because of the cost of context switching and spreading work into more, smaller chunks — but also because of the remainder that I end up with when I’m done working. A single extra hour in my day probably means nothing. Maybe I can binge-watch one more episode of a new show or do a few extra chores around the house. But it doesn’t significantly improve my life or help me find greater balance. Most things I want to do outside of work can’t fit into a single extra hour. A full day off from work unlocks many more options. I can take the day to go hiking or biking. I can spend the day with my wife, planning or playing a game. Or I can push it up against the weekend and take a 3-day trip. It gives me more of the freedom and balance that I ultimately want. So this has become a guiding principle for how I structure my schedule. I want to: Minimize context switching Maximize focused time for work and for non-work The idea of embracing reality also bleeds into some of the shifts in business strategy that I mentioned above. In theory, any time spent on marketing will have a positive impact on the company. In reality, focusing more on relationships than blasting tweets into the ether is much more likely to drive the kind of growth and stability that we’re seeking. As I think about 2023, I think this is a recurring theme. It manifests in many ways. Companies are making budget cuts and tough decisions about focus and strategy. Most of us are looking for ways to rein in the excess and have greater impact with a bit less time and money. We can’t do everything. We can’t even do most things. So our #1 priority should be to understand the reality of our time and our effort to make the most of every moment (in both work and leisure). That means thinking deeply about our strengths and our limitations. Being practical, even if it feels like sacrifice. Update on Other Businesses Finally, I want to close up by sharing a bit about my ventures outside of Optimist. I shared last year how I planned to shift some of my (finite) time and attention to new ventures and opportunities. And, while I didn’t get to devote as much as I hoped to these new pursuits, they weren’t totally in vain. I made progress across the board on all of the items I laid out in my post. Here’s what happened: Juice: The first Optimist spin-out agency At the end of 2021, we launched our first new service business based on demand from Optimist clients. Focused entirely on building links for SEO, we called the agency Juice. Overall, we made strong progress toward turning this into a legitimate standalone business in 2022. Relying mostly on existing Optimist clients and a few word-of-mouth opportunities (no other marketing), we built a team and set up a decent workflow and operations. There’s still many kinks and challenges that we’re working through on this front. All told, Juice posted almost $100,000 in revenue in our first full year. Monetizing the community I started 2022 with a focus on figuring out how to monetize our free community, Top of the Funnel. Originally, my plan was to sell sponsorships as the main revenue driver. And that option is still on the table. But, this year, I pivoted to selling paid content and subscriptions. We launched a paid tier for content and SEO entrepreneurs where I share more of my lessons, workflows, and ideas for building and running a freelance or agency business. It’s gained some initial traction — we reached \~$1,000 MRR from paid subscriptions. In total, our community revenue for 2022 was about $2,500. In 2023, I’m hoping to turn this into a $30,000 - $50,000 revenue opportunity. Right now, we’re on track for \~$15,000. Agency partnerships and referrals In 2022, we also got more serious about referring leads to other agencies. Any opportunity that was not a fit for Optimist or we didn’t have capacity to take on, we’d try to connect with another partner. Transparently, we struggled to operationalize this as effectively as I would have liked. In part, this was driven by my lack of focus here. With the other challenges throughout the year, I wasn’t able to dedicate as much time as I’d like to setting goals and putting workflows into place. But it wasn’t a total bust. We referred out several dozen potential clients to partner agencies. Of those, a handful ended up converting into sales — and referral commission. In total, we generated about $10,000 in revenue from referrals. I still see this as a huge opportunity for us to unlock in 2023. Affiliate websites Lastly, I mentioned spending some time on my new and existing affiliate sites as another big business opportunity in 2022. This ultimately fell to the bottom of my list and didn’t get nearly the attention I wanted. But I did get a chance to spend a few weeks throughout the year building this income stream. For 2022, I generated just under $2,000 in revenue from affiliate content. My wife has graciously agreed to dedicate some of her time and talent to these projects. So, for 2023, I think this will become a bit of a family venture. I’m hoping to build a solid and consistent workflow, expand the team, and develop a more solid business strategy. Postscript — AI, SEO, OMG As I’m writing this, much of my world is in upheaval. If you’re not in this space (and/or have possibly been living under a rock), the release of ChatGPT in late 2022 has sparked an arms race between Google, Bing, OpenAI, and many other players. The short overview: AI is likely to fundamentally change the way internet search works. This has huge impact on almost all of the work that I do and the businesses that I run. Much of our focus is on SEO and understanding the current Google algorithm, how to generate traffic for clients, and how to drive traffic to our sites and projects. That may all change — very rapidly. This means we’re standing at a very interesting point in time. On the one hand, it’s scary as hell. There’s a non-zero chance that this will fundamentally shift — possibly upturn — our core business model at Optimist. It could dramatically change how we work and/or reduce demand for our core services. No bueno. But it’s also an opportunity (there’s the optimist in me, again). I certainly see a world where we can become leaders in this new frontier. We can pivot, adjust, and capitalize on a now-unknown version of SEO that’s focused on understanding and optimizing for AI-as-search. With that, we may also be able to help others — say, those in our community? — also navigate this tumultuous time. See? It’s an opportunity. I wish I had the answers right now. But, it’s still a time of uncertainty. I just know that there’s a lot of change happening and I want to be in front of it rather than trying to play catch up. Wish me luck. — Alright friends — that's my update for 2023! I’ve always appreciated sharing these updates with the Reddit community, getting feedback, being asked tough questions, and even battling it out with some of my haters (hey!! 👋) As usual, I’m going to pop in throughout the next few days to respond to comments or answer questions. Feel free to share thoughts, ideas, and brutal takedowns in the comments. If you're interested in following the Optimist journey and the other projects I'm working on in 2023, you can follow me on Twitter. Cheers, Tyler P.S. - If you're running or launching a freelance or agency business and looking for help figuring it out, please DM me. Our subscription community, Middle of the Funnel, was created to provide feedback, lessons, and resources for other entrepreneurs in this space.

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company
reddit
LLM Vibe Score0
Human Vibe Score0.778
wutangsamThis week

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company

I’ve learned so much over the years from this subreddit. I thought I’d return the favour and share some of my own learnings. In November 2020 my best friend and I had an idea. “What if we could find out which stocks the Internet is talking about?” This formed the origins of Ticker Nerd. 9 months later we sold Ticker Nerd to Finder (an Australian fintech company valued at around $500m). In this post, I am going to lay out how we got there. How we came up with the idea First off, like other posts have covered - you don’t NEED a revolutionary or original idea to build a business. There are tonnes of “boring” businesses making over 7 figures a year e.g. law firms, marketing agencies, real estate companies etc. If you’re looking for an exact formula to come up with a great business idea I’m sorry, but it doesn’t exist. Finding new business opportunities is more of an art than a science. Although, there are ways you can make it easier to find inspiration. Below are the same resources I use for inspiration. I rarely ever come up with ideas without first searching one of the resources below for inspiration: Starter Story Twitter Startup Ideas My First Million Trends by the Hustle Trends VC To show how you how messy, random and unpredictable it can be to find an idea - let me explain how my co-founder and I came up with the idea for Ticker Nerd: We discovered a new product on Twitter called Exploding Topics. It was a newsletter that uses a bunch of software and algorithms to find trends that are growing quickly before they hit the mainstream. I had recently listened to a podcast episode from My First Million where they spoke about Motley Fool making hundreds of millions from their investment newsletters. We asked ourselves what if we could build a SaaS platform similar to Exploding Topics but it focused on stocks? We built a quick landing page using Carrd + Gumroad that explained what our new idea will do and included a payment option to get early access for $49. We called it Exploding Stock (lol). We shared it around a bunch of Facebook groups and subreddits. We made $1,000 in pre-sales within a couple days. My co-founder and I can’t code so we had to find a developer to build our idea. We interviewed a bunch of potential candidates. Meanwhile, I was trawling through Wall Street Bets and found a bunch of free tools that did roughly what we wanted to build. Instead of building another SaaS tool that did the same thing as these free tools we decided to pivot from our original idea. Our new idea = a paid newsletter that sends a weekly report that summarises 2 of the best stocks that are growing in interest on the Internet. We emailed everyone who pre-ordered access, telling them about the change and offered a full refund if they wanted. tl;dr: We essentially combined two existing businesses (Exploding Topics and Motley Fool) and made it way better. We validated the idea by finding out if people will actually pay money for it BEFORE we decided to build it. The idea we started out with changed over time. How to work out if your idea will actually make money It’s easy to get hung up on designing the logo or choosing the perfect domain name for your new idea. At this stage none of that matters. The most important thing is working out if people will pay money for it. This is where validation comes in. We usually validate ideas using Carrd. It lets you build a simple one page site without having to code. The Ticker Nerd site was actually built using a Carrd template. Here’s how you can do it yourself (at a high level): Create a Carrd pro account (yes it's a $49 one off payment but you’ll get way more value out of it). Buy a cheap template and send it to your Carrd account. You can build your own template but this will save you a lot of time. Once the template reaches your Carrd account, duplicate it. Leave the original so it can be duplicated for other ideas. Jump onto Canva (free) and create a logo using the free logos provided. Import your logo. Add copy to the page that explains your idea. Use the AIDA formula. Sign up to Gumroad (free) and create a pre-sale campaign. Create a discounted lifetime subscription or version of the product. This will be used pre-sales. Add the copy from the site into the pre-sale campaign on Gumroad. Add a ‘widget’ to Carrd and connect it to Gumroad using the existing easy integration feature. Purchase a domain name. Connect it to Carrd. Test the site works. Share your website Now the site is ready you can start promoting it in various places to see how the market reacts. An easy method is to find relevant subreddits using Anvaka (Github tool) or Subreddit Stats. The Anvaka tool provides a spider map of all the connected subreddits that users are active in. The highlighted ones are most relevant. You can post a thread in these subreddits that offer value or can generate discussion. For example: ‘I’m creating a tool that can write all your copy, would anyone actually use this?’ ‘What does everything think of using AI to get our copy written faster?’ ‘It’s time to scratch my own itch, I’m creating a tool that writes marketing copy using GPT-3. What are the biggest problems you face writing marketing copy? I’ll build a solution for it’ Reddit is pretty brutal these days so make sure the post is genuine and only drop your link in the comments or in the post if it seems natural. If people are interested they’ll ask for the link. Another great place to post is r/entrepreuerridealong and r/business_ideas. These subreddits expect people to share their ideas and you’ll likely make some sales straight off the bat. I also suggest posting in some Facebook groups (related to your idea) as well just for good measure. Assess the results If people are paying you for early access you can assume that it’s worth building your idea. The beauty of posting your idea on Reddit or in Facebook groups is you’ll quickly learn why people love/hate your idea. This can help you decide how to tweak the idea or if you should drop it and move on to the next one. How we got our first 100 customers (for free) By validating Ticker Nerd using subreddits and Facebook groups this gave us our first paying customers. But we knew this wouldn’t be sustainable. We sat down and brainstormed every organic strategy we could use to get traction as quickly as possible. The winner: a Product Hunt launch. A successful Product Hunt launch isn’t easy. You need: Someone that has a solid reputation and audience to “hunt” your product (essentially an endorsement). An aged Product Hunt account - you can’t post any products if your account is less than a week old. To be following relevant Product Hunt members - since they get notified when you launch a new product if they’re following you. Relationships with other builders and makers on Product Hunt that also have a solid reputation and following. Although, if you can pull it off you can get your idea in front of tens of thousands of people actively looking for new products. Over the next few weeks, I worked with my co-founder on connecting with different founders, indie hackers and entrepreneurs mainly via Twitter. We explained to them our plans for the Product Hunt launch and managed to get a small army of people ready to upvote our product on launch day. We were both nervous on the day of the launch. We told ourselves to have zero expectations. The worst that could happen was no one signed up and we were in the same position as we’re in now. Luckily, within a couple of hours Ticker Nerd was on the homepage of Product Hunt and in the top 10. The results were instant. After 24 hours we had around 200 people enter their payment details to sign up for our free trial. These signups were equal to around $5,800 in monthly recurring revenue. \-- I hope this post was useful! Drop any questions you have below and I’ll do my best to respond :)

Innovating marketing strategies: ads crafted with AI show a 2x boost in views and an 4x rise in likes
reddit
LLM Vibe Score0
Human Vibe Score0
Bryan_JostlingThis week

Innovating marketing strategies: ads crafted with AI show a 2x boost in views and an 4x rise in likes

Hey there! Recently, me and my friends conducted a new comprehensive study comparing the effectiveness of standard video ads to AI-generated content on TikTok. Our goal was to gain insights into which type of content garnered more attention in terms of views, both organic and paid, as well as engagement rate. Study background: Imalent, known for its innovation in portable lighting and powerful flashlights, decided to use an AI video generator for creating TikTok ads. For this study, we selected three top-performing ads created by designers and three ads generated with Creatify AI. Here's a breakdown of our findings: Organic Views: The results showed that AI-powered videos outperformed standard ones by 8x in organic views: videos produced by human designers got 24K organic views, while those generated by artificial intelligence got 189K views. Paid Views: AI videos attracted twice as many views as regular ads for the same budget. Traditional ads got 115K views and AI-generated ones got 259K views. Engagement: Perhaps the most astonishing aspect of our study was the engagement metrics. AI-generated content received 7 times more saves, 4 times more likes, and twice as many comments compared to standard videos. Have you considered using AI ads to promote your brand? Share your insights and experiences below! For more data and screenshots, visit the full study here.

Secret behind Airbnb's Billion-Dollar Empire? Spamming Craigslist
reddit
LLM Vibe Score0
Human Vibe Score1
deadcoder0904This week

Secret behind Airbnb's Billion-Dollar Empire? Spamming Craigslist

Silicon Valley wants you to believe that their unicorn startups succeeded doing things legally. But that couldn't be far from truth. For starters, Airbnb used multiple Gmail accounts to spam Craigslist. "They posted unrealistically (fake) cheap rentals of beautiful apartments in places where normal rent should be 10x more. Once people replied, they auto-responded that the unit has been rented, but they should be looking for another unit on AirBnB." The Game of Blackhat is a cat-and-mouse game. You need a lot of guardrails to protect yourself from people using your Social Site by spamming their products. Craigslist is a team of 30 people. There's stuff AI can automate now with such a small team but back then, it wasn't possible. Airbnb used Craigslist as its playground to spam Craigslist visitors to grow their supply-side. In a 2-sided marketplace, growing both supply and demand is very important. And both must grow at the same time for the marketplace to work. A Blackhat Marketer created a new test site to get vacation rental owners to sign-up so that he can test his Airbnb theory. He grabbed their real email-addresses (not Craigslist anonymous addresses) via Craigslist by specifically targeting those who were advertising their vacation rentals on Craigslist. He skipped over the other categories that were directly related to AirBnB's business model because they didn't fit with the test site he built. Once he got 1000+ sign-ups, he then took it upon himself to post it to the advertising section on Craigslist. The email said this: I am emailing you because you have one of the nicest listings on Craigslist in Idaho and I want to recommend you feature it (for free) on one of the largest Idaho housing sites on the web, Airbnb. The site already has 3,000,000 pages views a month. Check it out here to list now: airbnb(dot)com Sarah Surpisingly, all emails were by ladies. He did the same in Week 2 and Week 3 to test if it wasn't a one-time thing. Surely, it wasn't a fluke. After posting 4 ads on Craigslist in 3 weeks, he received 5 identical emails from 2 ladies who were raving fans of AirBnB and spent their days emailing Craigslist advertisers. This is one of the greatest blackhat strategies used in the real world to build a billion-dollar marketplace by growing the supply-side with pure blackhat. These strategies are not mentioned in Press Interviews, Media, or any Founder stories but this is probably the most important piece of the puzzle. Without it, Airbnb probably wouldn't have survived. "Some very famous investors have alluded to the fact that they look for a dangerous streak in the entrepreneurs they invest in…and while those investors will never come out and tell you what they mean, this kind of thing is probably what they mean." It definitely violates CAN-SPAM act. Some comments from Hacker News: "CAN-SPAM, sending from a fake address (illegal headers). CA has a specific law that pre-empts CAN-SPAM that definitely makes this illegal if sent from CA." But I guess it worked in Airbnb's favour lol as they were never caught or fined until after. "It's commercial email 100%. Probably a fake sender name (illegal), against gmail ToS, against CL ToS and no unsubscribe link and no one even subscribed in the first place. 100% against CAN-SPAM." Thanks for reading. If you'd like to learn more blackhat tactics like this, check this site which is a growth hacking newsletter with real-world blackhat examples. PS: Actual emails & screenshots from the Airbnb x Craigslist spam can be found here.

boring passive site... now 42k monthly visitors and $2540 MRR
reddit
LLM Vibe Score0
Human Vibe Score1
TasAdamsThis week

boring passive site... now 42k monthly visitors and $2540 MRR

people underestimate SEO... It is evergreen... passive... digital real estate. it can do magic... if you are consistent. Especially now with AI you can 2X your traffic growth and automate 85% of the work. For the past 6 months... we've been building an online directory. we just reached $2540 MRR... with SEO only... from a complete zero. I did share this on other subreddits. Maybe this gives ideas to someone. \+ This can be easily replicated if you have a website lol Current metrics: $2540 MRR - businesses pay us to list on the directory + display ads + pay to be featured. 43k monthly visitors - in the past couple of weeks our SEO growth is a hockey stick. DR (Domain Rating) 35 - it took us 2.5 months to get to that. 51 okay-ish quality referring domains (90% of them are do-follow) and 1.6k backlinks. There are probably 3 main pillars I try to focus on: keywords --> which then is the basis for ALL the content pieces we do blogs, landing pages, about us pages, competitor comparisons etc --> we use a DIY excel file to automate content production at scale. backlinks --> boost DR --> one of the main things to boost ranking on google. website health --> this is technical stuff like internal and external linking, schemas, canonical tags, alt texts, load speeds, compressed images, meta descriptions, titles etc --> do this once... and do it GOOD. $0.07 per SEO optimised blog at scale with AI Yep... we've literally built our own SEO blog tool... and it is a Spreadsheet with bunch of app scripts :D NOTE that we add a little bit of human touch to those blogs that are picked up by Google rank top in 25 How it works... is that we paste in bunch of links (other websites, blogs, news articles) and with a click of a button we can get up to 2000 SEO optimised content pieces... from an Excel file... $0.07 per blog. The spreadsheet is integrated with Chat gpt (obviously). We use GPT-4 for meta descriptions, titles, transforming the content from text to html code since it is more powerful, and GPT-4o for content itself because it is cheaper and faster for "general text". The spreadsheet repurposes content. The spreadsheet generates: Meta descriptions and titles FAQs sections - DON'T skip FAQ sections! They are a must for SEO. On Ahrefs... there is a section of questions people are searching about your keyword... that's your FAQs It can find contextual youtube videos (links to those videos) - to show google that our content is not "just text" thus higher quality. Screenshots and images of the original source (the website link we inputed) I then download a csv version of the excel and import it into our Webflow. The csv file column names match our webflow CMS field names. tbh... we didn't even know that it can be done with a spreadsheet. We "tried" building it because every other tool we were using is (1) expensive from $0.59 per SEO content piece (2) they didn't provide the scale we wanted (3) we wanted more control over the output. Focus on DR 35+ backlinks... easier We bought backlinks only once... rest of the backlinks was a manual work from us. Bunch of free listing databases (about 65% of our backlinks) You can comment on open forums with your link to get a backlink (be careful tho) Post a blog on Medium com --> DR 94 backlink (takes time to Index) If you pay for Notion you can get a DR 94 backlink from Notion If you use Beehiiv you can get a DR 86 backlink from Beehiiv Google product stacking (Google sites, Google notes etc) --> backlink from almighty Google itself A lot of work goes into backlinks because they are THAT important. I have tried bunch of "black hat" strategies as well... but note that all of these strategies won't work if you don't index the primary source from where your backlink is coming from. BIG search volume and low KD Key things I'm looking for in keywords: I use Ahrefs Keyword research tool... it is literally free BIG search volume - 2k+ is oaky-ish for a single keyword EASY to rank - KD (keyword difficulty) below 15 Look for long tail keywords (these are golden nuggets since they have a VERY clear search intent) - "how to edit..." "how to change..." "how to delete..." "how to paint..." I hope you got the idea. on Ahrefs you can use "\" to get BIG volume long tail keywords... like this "my keyword\". Ahrefs then populates the "\" with the tail. Check SERP (Search Engine Result Page) for your keywords - it shows current top 10 pages for those KWs. Check their content. Can you improve it? Have they missed anything? Keyword gap from your competitors - shows EASY keywords that your competitors have missed and also shows what keywords overlap with you. Also one cool thing... if you don't type any keywords on Ahrefs and press "Enter"... you can browse all the keywords out there... it is magical. Once we have the keywords, we run our spreadsheet. And that's pretty much it. I hope that you can get some ideas from this little silly project. Also... if you have any questions about this... I might share the SEO blog automation excel file/help if people are interested...

AI Will Make You Extremely Rich or Kill Your Business in 2024
reddit
LLM Vibe Score0
Human Vibe Score1
AntsyNursery58This week

AI Will Make You Extremely Rich or Kill Your Business in 2024

Preface: I'm a solo-founder in the AI space and previously worked as an ML scientist; the new advancements in AI that I'm seeing are going to impact everyone here. It doesn't matter if you're just starting out, or a bootstrapped brick and mortar founder, or even a VC backed hard tech founder. Last year was when the seeds were laid, and this is the year we'll see them bloom. There will be an onslaught of advancements that take place that are borderline inconceivable due to the nature of exponential progress. This will change every single vertical. I'm making this post because I think AI execution strategy will make or break businesses. Dramatically. Over $50B was put into AI startups in 2023 alone. This figure excludes the hundreds of billions poured into AI from enterprises. So, let's follow the money: &#x200B; 1) AI enterprise software. There's a lot to unpack here and this is what I’m currently working on. AI enterprise software will encompass everything from hyper personalized email outbound to AI cold calls to AI that A/B tests ads on synthetic data to vertical specific software. The impact of the former is relatively self explanatory, so I'll focus on the latter. To illustrate vertical specific AI software, I'll use a simple example in the legal space. Lawyers typically have to comb through thousands of pages of documents. Now, using an LLM + a VDB, an AI can instantly answer all of those questions while surfacing the source and highlighting the specific answer in the contract/document. There are dozens of AI startups for this use case alone. This saves lawyers an immense amount of time and allows them to move faster. Firms that adopt this have a fundamental advantage over law firms that don't adopt this. This was 2023 technology. I'm seeing vertical AI software getting built by my friends in areas from construction, to real estate, to even niche areas like chimney manufacturing. This will exist everywhere. Now, this can be extrapolated much further to be applicable to systems that can do reports and even browse the Internet. This brings me to my next point. &#x200B; 2) AI information aggregation and spread. My gut tells me that this will have a crescendo moment in the future with hardware advancements (Rabbit, Tab, etc.). You won't have to google things because it will be surfaced to you. It's predictive in nature. The people who can get information the fastest will grow their business the fastest. This part is semi-speculative, but due to the nature of LLMs being so expensive to train, I have a strong feeling that large institutions will have access to the \fastest\ and \best\ models that can do this quicker than you and I can. This is why it's important to stay on top. &#x200B; 3) AI content generation This is relevant to running advertisements and any digital marketing aspect of your business. If you can rapidly make content faster than your competitors to put in social media, you will outpace your competitors rapidly. I think most folks are familiar with MidJourney, Stable diffusion, etc. but don't know how to use it. You can generate consistent models for a clothing brand or generate images of a product that you would normally need to hire a professional photographer to take. There's also elevenlabs which is relatively easy to use and can be used to make an MP3 clip as a narration for an ad; this is something I've already done. I'm also still shocked by how many people are unfamiliar with tools like Pika which can do video generation. You could imagine companies having fleets of digital influencers that they control or conjuring up the perfect ad for a specific demographic using a combination of all of the aforementioned tools. &#x200B; In summary, if you feel like I'm being hyperbolic or propagating science fiction fantasies, you're likely already behind. I truly recommend that everyone stays up to date on these advancements as much as possible. If your competitor comes across an AI tool that can increase their ROAS by 5x they can crush you. If your competitor uses a tool that increases the rate at which they receive and aggregate information by 200% (modest estimate) they will crush you. If your competitors have a tool that can reduce their employee size, then they will use it. They'll fire their employees to cut costs and reinvest the money back into their business. It will compound to the point where you're outpaced, and this isn't a level of innovation we've seen since the birth of the industrial revolution. Your customers can get stolen overnight, or you can steal your competition’s customers overnight. TL;DR: This is an opportunity for entrepreneurs to scale faster than they could have possibly imagined, but this also comes with the potential for your company to be obliterated. We've never seen advancements that can have this drastic of an impact this quickly. Adoption will happen fast, and first movers will have a disproportionate and compounding advantage. Watch guides, meet with startups, follow the news, and get rich.

Why Ignoring AI Agents in 2025 Will Kill Your Marketing Strategy
reddit
LLM Vibe Score0
Human Vibe Score1
frankiemuiruriThis week

Why Ignoring AI Agents in 2025 Will Kill Your Marketing Strategy

If you're still focusing solely on grabbing the attention of human beings with your marketing efforts, you're already behind. In 2025, the game will change. Good marketing will demand an in-depth understanding of the AI space, especially the AI Agent space. Why? Your ads and content won’t just be seen by humans anymore. They’ll be analyzed, indexed, and often acted upon by AI agents—automated systems that will be working on behalf of companies and consumers alike. Your New Audience: Humans + AI Agents It’s not just about appealing to people. Companies are employing AI robots to research, negotiate, and make purchasing decisions. These AI agents are fast, thorough, and unrelenting. Unlike humans, they can analyze millions of options in seconds. And if your marketing isn’t optimized for them, you’ll get filtered out before you even reach the human decision-maker. How to Prepare Your Marketing for AI Agents The companies that dominate marketing in 2025 will be the ones that master the art of capturing AI attention. To do this, marketers will need to: Understand the AI agents shaping their industry. Research how AI agents function in your niche. What are they prioritizing? How do they rank options? Create AI-friendly content. Design ads and messaging that are easily understandable and accessible to AI agents. This means clear metadata, structured data, and AI-readable formats. Invest in AI analytics. AI agents leave behind footprints. Tracking and analyzing their behavior is critical. Stay ahead of AI trends. The AI agent space is evolving rapidly. What works today might be obsolete tomorrow. How My Agency Adapted and Thrived in the AI Space At my digital agency, we saw this shift coming and decided to act early. In 2023, we started integrating AI optimization into our marketing strategies. One of our clients—a B2B SaaS company—struggled to get traction because their competitors were drowning them out in Google search rankings and ad platforms. By analyzing the algorithms and behaviors of AI agents in their space, we: Rewrote their website copy with structured data and optimized metadata that was more AI-agent friendly. Created ad campaigns with clear, concise messaging and technical attributes that AI agents could quickly process and index. Implemented predictive analytics to understand what AI agents would prioritize based on past behaviors. The results? Their website traffic doubled in three months, and their lead conversion rate skyrocketed by 40%. Over half of the traffic increase was traced back to AI agents recommending their platform to human users. The Takeaway In 2025, marketing won’t just be about human attention. It’ll be about AI attention—and that requires a completely different mindset. AI agents are not your enemy; they’re your new gatekeepers. Learn to speak their language, and you’ll dominate the marketing game.

My Side Projects: From CEO to 4th Developer (Thanks, AI 🤖)
reddit
LLM Vibe Score0
Human Vibe Score1
tilopediaThis week

My Side Projects: From CEO to 4th Developer (Thanks, AI 🤖)

Hey Reddit 👋, I wanted to share a bit about some side projects I’ve been working on lately. Quick background for context: I’m the CEO of a mid-to-large-scale eCommerce company pulling in €10M+ annually in net turnover. We even built our own internal tracking software that’s now a SaaS (in early review stages on Shopify), competing with platforms like Lifetimely and TrueROAS. But! That’s not really the point of this post — there’s another journey I’ve been on that I’m super excited to share (and maybe get your feedback on!). AI Transformed My Role (and My Ideas List) I’m not a developer by trade — never properly learned how to code, and to be honest, I don’t intend to. But, I’ve always been the kind of guy who jots down ideas in a notes app and dreams about execution. My dev team calls me their “4th developer” (they’re a team of three) because I have solid theoretical knowledge and can kinda read code. And then AI happened. 🛠️ It basically turned my random ideas app into an MVP generation machine. I thought it’d be fun to share one of the apps I’m especially proud of. I am also planning to build this in public and therefore I am planning to post my progress on X and every project will have /stats page where live stats of the app will be available. Tackling My Task Management Problem 🚀 I’ve sucked at task management for YEARS, I still do! I’ve tried literally everything — Sheets, Todoist, Asana, ClickUp, Notion — you name it. I’d start… and then quit after a few weeks - always. What I struggle with the most is delegating tasks. As a CEO, I delegate a ton, and it’s super hard to track everything I’ve handed off to the team. Take this example: A few days ago, I emailed an employee about checking potential collaboration opportunities with a courier company. Just one of 10s of tasks like this I delegate daily. Suddenly, I thought: “Wouldn’t it be AMAZING if just typing out this email automatically created a task for me to track?” 💡 So… I jumped in. With the power of AI and a few intense days of work, I built a task manager that does just that. But of course, I couldn’t stop there. Research & Leveling It Up 📈 I looked at similar tools like TickTick and Todoist, scraped their G2 reviews (totally legally, promise! 😅), and ran them through AI for a deep SWOT analysis. I wanted to understand what their users liked/didn’t like and what gaps my app could fill. Some of the features people said they were missing didn’t align with the vision for my app (keeping it simple and personal), but I found some gold nuggets: Integration with calendars (Google) Reminders Customizable UX (themes) So, I started implementing what made sense and am keeping others on the roadmap for the future. And I’ve even built for that to, it still doesn’t have a name, however the point is you select on how many reviews of a specific app you want to make a SWOT analysis on and it will do it for you. Example for Todoist in comments. But more on that, some other time, maybe other post ... Key Features So Far: Here’s what’s live right now: ✅ Email to Task: Add an email as to, cc, or bcc — and it automatically creates a task with context, due dates, labels, etc. ✅ WhatsApp Reminders: Get nudged to handle your tasks via WhatsApp. ✅ WhatsApp to Task: Send a message like /task buy groceries — bam, it’s added with full context etc.. ✅ Chrome Extension (work-in-progress): Highlight text on any page, right-click, and send it straight to your task list. Next Steps: Build WITH the Community 👥 Right now, the app is 100% free while still in the early stages. But hey, API calls and server costs aren’t cheap, so pricing is something I’ll figure out with you as we grow. For now, my goal is to hit 100 users and iterate from there. My first pricing idea is, without monthly subscription, I don’t want to charge someone for something he didn’t use. So I am planning on charging "per task", what do you think? Here’s what I have planned: 📍 End of Year Goal: 100 users (starting from… 1 🥲). 💸 Revenue Roadmap: When we establish pricing, we’ll talk about that. 🛠️ Milestones: Post on Product Hunt when we hit 100 users. Clean up my self-written spaghetti code (hire a pro dev for review 🙃). Hire a part-time dev once we hit MRR that can cover its costs. You can check how are we doing on thisisatask.me/stats Other Side Projects I’m Working On: Because… what’s life without taking on too much, right? 😂 Full list of things I’m building: Internal HRM: Not public, tried and tested in-house. Android TV App: Syncs with HRM to post announcements to office TVs (streamlined and simple). Stats Tracker App: Connects to our internal software and gives me real-time company insights. Review Analyzer: Scrapes SaaS reviews (e.g., G2) and runs deep analysis via AI. This was originally for my Shopify SaaS but is quickly turning into something standalone. Coming soon! Mobile app game: secret for now. Let’s Build This Together! Would love it if you guys checked out thisisatask.me and gave it a spin! Still super early, super raw, but I’m pumped to hear your thoughts. Also, what’s a must-have task manager feature for you? Anything that frustrates you with current tools? I want to keep evolving this in public, so your feedback is gold. 🌟 Let me know, Reddit! Are you with me? 🙌

How To Learn About AI Agents (A Road Map From Someone Who's Done It)
reddit
LLM Vibe Score0
Human Vibe Score0.882
laddermanUSThis week

How To Learn About AI Agents (A Road Map From Someone Who's Done It)

If you are a newb to AI Agents, welcome, I love newbies and this fledgling industry needs you! You've hear all about AI Agents and you want some of that action right?  You might even feel like this is a watershed moment in tech, remember how it felt when the internet became 'a thing'?  When apps were all the rage?  You missed that boat right?   Well you may have missed that boat, but I can promise you one thing..... THIS BOAT IS BIGGER !  So if you are reading this you are getting in just at the right time.  Let me answer some quick questions before we go much further: Q: Am I too late already to learn about AI agents? A: Heck no, you are literally getting in at the beginning, call yourself and 'early adopter' and pin a badge on your chest! Q: Don't I need a degree or a college education to learn this stuff?  I can only just about work out how my smart TV works! A: NO you do not.  Of course if you have a degree in a computer science area then it does help because you have covered all of the fundamentals in depth... However 100000% you do not need a degree or college education to learn AI Agents.  Q: Where the heck do I even start though?  Its like sooooooo confusing A: You start right here my friend, and yeh I know its confusing, but chill, im going to try and guide you as best i can. Q: Wait i can't code, I can barely write my name, can I still do this? A: The simple answer is YES you can. However it is great to learn some basics of python.  I say his because there are some fabulous nocode tools like n8n that allow you to build agents without having to learn how to code...... Having said that, at the very least understanding the basics is highly preferable. That being said, if you can't be bothered or are totally freaked about by looking at some code, the simple answer is YES YOU CAN DO THIS. Q: I got like no money, can I still learn? A: YES 100% absolutely.  There are free options to learn about AI agents and there are paid options to fast track you.  But defiantly you do not need to spend crap loads of cash on learning this.  So who am I anyway? (lets get some context)  I am an AI Engineer and I own and run my own AI Consultancy business where I design, build and deploy AI agents and AI automations.  I do also run a small academy where I teach this stuff, but I am not self promoting or posting links in this post because im not spamming this group.  If you want links send me a DM or something and I can forward them to you.  Alright so on to the good stuff, you're a newb, you've already read a 100 posts and are now totally confused and every day you consume about 26 hours of youtube videos on AI agents.....I get you, we've all been there.  So here is my 'Worth Its Weight In Gold' road map on what to do: \[1\]  First of all you need learn some fundamental concepts.  Whilst you can defiantly jump right in start building, I strongly recommend you learn some of the basics.  Like HOW to LLMs work, what is a system prompt, what is long term memory, what is Python, who the heck is this guy named Json that everyone goes on about?  Google is your old friend who used to know everything, but you've also got your new buddy who can help you if you want to learn for FREE.  Chat GPT is an awesome resource to create your own mini learning courses to understand the basics. Start with a prompt such as: "I want to learn about AI agents but this dude on reddit said I need to know the fundamentals to this ai tech, write for me a short course on Json so I can learn all about it. Im a beginner so keep the content easy for me to understand. I want to also learn some code so give me code samples and explain it like a 10 year old" If you want some actual structured course material on the fundamentals, like what the Terminal is and how to use it, and how LLMs work, just hit me, Im not going to spam this post with a hundred links. \[2\] Alright so let's assume you got some of the fundamentals down.  Now what? Well now you really have 2 options.  You either start to pick up some proper learning content (short courses) to deep dive further and really learn about agents or you can skip that sh\*t and start building!  Honestly my advice is to seek out some short courses on agents, Hugging Face have an awesome free course on agents and DeepLearningAI also have numerous free courses. Both are really excellent places to start.  If you want a proper list of these with links, let me know.  If you want to jump in because you already know it all, then learn the n8n platform!   And no im not a share holder and n8n are not paying me to say this.  I can code, im an AI Engineer and I use n8n sometimes.   N8N is a nocode platform that gives you a drag and drop interface to build automations and agents.  Its very versatile and you can self host it.  Its also reasonably easy to actually deploy a workflow in the cloud so it can be used by an actual paying customer.  Please understand that i literally get hate mail from devs and experienced AI enthusiasts for recommending no code platforms like n8n.  So im risking my mental wellbeing for you!!!    \[3\] Keep building!   ((WTF THAT'S IT?????))  Yep. the more you build the more you will learn.  Learn by doing my young Jedi learner.  I would call myself pretty experienced in building AI Agents, and I only know a tiny proportion of this tech.  But I learn but building projects and writing about AI Agents.  The more you build the more you will learn.  There are more intermediate courses you can take at this point as well if you really want to deep dive (I was forced to - send help) and I would recommend you do if you like short courses because if you want to do well then you do need to understand not just the underlying tech but also more advanced concepts like Vector Databases and how to implement long term memory.  Where to next? Well if you want to get some recommended links just DM me or leave a comment and I will DM you, as i said im not writing this with the intention of spamming the crap out of the group. So its up to you.  Im also happy to chew the fat if you wanna chat, so hit me up.  I can't always reply immediately because im in a weird time zone, but I promise I will reply if you have any questions. THE LAST WORD (Warning - Im going to motivate the crap out of you now) Please listen to me:  YOU CAN DO THIS.  I don't care what background you have, what education you have, what language you speak or what country you are from..... I believe in you and anyway can do this.  All you need is determination, some motivation to want to learn and a computer (last one is essential really, the other 2 are optional!) But seriously you can do it and its totally worth it.  You are getting in right at the beginning of the gold rush, and yeh I believe that.   AI Agents are going to be HUGE. I believe this will be the new internet gold rush.

Recently hit 6,600,000 monthly organic traffic for a B2C SaaS website. Here's the 40 tips that helped me make that happen.
reddit
LLM Vibe Score0
Human Vibe Score1
DrJigsawThis week

Recently hit 6,600,000 monthly organic traffic for a B2C SaaS website. Here's the 40 tips that helped me make that happen.

Hey guys! So as title says, we recently hit 6,600,000 monthly organic traffic / month for a B2C SaaS website (screenshot. Can't give name publicly, but can show testimonial to a mod). Here's 40 tips that "helped" me make this happen. If you get some value of the post, I write an SEO tip every other day on /r/seogrowth. There's around 10 more tips already up there other than the ones I mention here. If you want to give back for all my walls of text, I'd appreciate a sub <3 Also, there are a bunch of free stuff I mention in the article: content outline, writer guidelines, SEO checklist, and other stuff. Here's the Google Doc with all that! Tip #1. Take SEO With a Grain of Salt A lot of the SEO advice and best practices on the internet are based on 2 things: Personal experiences and case studies of companies that managed to make SEO work for them. Google or John Mueller (Google’s Senior Webmaster Trends Analyst). And, unfortunately, neither of these sources are always accurate. Personal SEO accounts are simply about what worked for specific companies. Sometimes, what worked for others, won’t work for you. For example, you might find a company that managed to rank with zero link-building because their website already had a very strong backlink profile. If you’re starting with a fresh website, chances are, you won’t be able to get the same results. At the same time, information from Google or John Mueller is also not 100% accurate. For example, they’ve said that guest posting is against Google’s guidelines and doesn’t work… But practically, guest posting is a very effective link-building strategy. So the takeaway is this: Take all information you read about SEO with a grain of salt. Analyze the information yourself, and make your conclusions. SEO Tip #2. SEO Takes Time You’ve already heard this one before, but considering how many people keep asking, thought I'd include this anyway. On average, it’s going to take you 6 months to 2 years to get SEO results, depending on the following factors: Your backlink profile. The more quality backlinks you have (or build), the faster you’ll rank. Age of your website. If your website is older (or you purchased an aged website), you can expect your content to rank faster. Amount of content published. The more quality content you publish on your website, the more “authoritative” it is in the eyes of Google, and thus more likely to rank faster. SEO work done on the website. If a lot of your pages are already ranking on Google (page 2-3), it’s easier to get them to page #1 than if you just published the content piece. Local VS global SEO. Ranking locally is (sometimes) easier and faster than ranking globally. That said, some marketing agencies can use “SEO takes time” as an excuse for not driving results. Well, fortunately, there is a way to track SEO results from month #2 - #3 of work. Simply check if your new content pieces/pages are getting more and more impressions on Google Search Console month-to-month. While your content won’t be driving traffic for a while after being published, they’ll still have a growing number of impressions from month #2 or #3 since publication. SEO Tip #3. SEO Might Not Be The Best Channel For You In theory, SEO sounds like the best marketing channel ever. You manage to rank on Google and your marketing seemingly goes on auto-pilot - you’re driving new leads every day from existing content without having to lift a finger… And yet, SEO is not for everyone. Avoid SEO as a marketing channel if: You’re just getting started with your business and need to start driving revenue tomorrow (and not in 1-2 years). If this is you, try Google ads, Facebook ads, or organic marketing. Your target audience is pretty small. If you’re selling enterprise B2B software and have around 2,000 prospects in total worldwide, then it’s simply easier to directly reach out to these prospects. Your product type is brand-new. If customers don’t know your product exists, they probably won’t be Googling it. SEO Tip #4. Traffic Can Be a Vanity Metric I've seen hundreds of websites that drive 6-7 digits of traffic but generate only 200-300 USD per month from those numbers. “What’s the deal?” You might be thinking. “How can you fail to monetize that much traffic?” Well, that brings us to today’s tip: traffic can be a vanity metric. See, not all traffic is created equal. Ranking for “hormone balance supplement” is a lot more valuable than ranking for “Madagascar character names.” The person Googling the first keyword is an adult ready to buy your product. Someone Googling the latter, on the other hand, is a child with zero purchasing power. So, when deciding on which keywords to pursue, always keep in mind the buyer intent behind and don’t go after rankings or traffic just because 6-digit traffic numbers look good. SEO Tip #5. Push Content Fast Whenever you publish a piece of content, you can expect it to rank within 6 months to a year (potentially less if you’re an authority in your niche). So, the faster you publish your content, the faster they’re going to age, and, as such, the faster they’ll rank on Google. On average, I recommend you publish a minimum of 10,000 words of content per month and 20,000 to 30,000 optimally. If you’re not doing link-building for your website, then I’d recommend pushing for even more content. Sometimes, content velocity can compensate for the lack of backlinks. SEO Tip #6. Use Backlink Data to Prioritize Content You might be tempted to go for that juicy, 6-digit traffic cornerstone keyword right from the get-go... But I'd recommend doing the opposite. More often than not, to rank for more competitive, cornerstone keywords, you’ll need to have a ton of supporting content, high-quality backlinks, website authority, and so on. Instead, it’s a lot more reasonable to first focus on the less competitive keywords and then, once you’ve covered those, move on to the rest. Now, as for how to check keyword competitiveness, here are 2 options: Use Mozbar to see the number of backlinks for top-ranking pages, as well as their Domain Authority (DA). If all the pages ranking on page #1 have <5 backlinks and DA of 20 - 40, it’s a good opportunity. Use SEMrush or Ahrefs to sort your keywords by difficulty, and focus on the less difficult keywords first. Now, that said, keep in mind that both of these metrics are third-party, and hence not always accurate. SEO Tip #7. Always Start With Competitive Analysis When doing keyword research, the easiest way to get started is via competitive analysis. Chances are, whatever niche you’re in, there’s a competitor that is doing great with SEO. So, instead of having to do all the work from scratch, run their website through SEMrush or Ahrefs and steal their keyword ideas. But don’t just stop there - once you’ve borrowed keyword ideas from all your competitors, run the seed keywords through a keyword research tool such as UberSuggest or SEMrush Keyword Magic Tool. This should give you dozens of new ideas that your competitors might’ve missed. Finally, don’t just stop at borrowing your competitor’s keyword ideas. You can also borrow some inspiration on: The types of graphics and images you can create to supplement your blog content. The tone and style you can use in your articles. The type of information you can include in specific content pieces. SEO Tip #8. Source a LOT of Writers Content writing is one of those professions that has a very low barrier to entry. Anyone can take a writing course, claim to be a writer, and create an UpWork account… This is why 99% of the writers you’ll have to apply for your gigs are going to be, well, horrible. As such, if you want to produce a lot of content on the reg, you’ll need to source a LOT of writers. Let’s do the math: If, by posting a job ad, you source 100 writers, you’ll see that only 5 of them are a good fit. Out of the 5 writers, 1 has a very high rate, so they drop out. Another doesn’t reply back to your communication, which leaves you with 3 writers. You get the 3 writers to do a trial task, and only one turns out to be a good fit for your team. Now, since the writer is freelance, the best they can do is 4 articles per month for a total of 5,000-words (which, for most niches, ain’t all that much). So, what we’re getting at here is, to hire quality writers, you should source a LOT of them. SEO Tip #9. Create a Process for Filtering Writers If you follow the previous tip, you'll end up with a huge database of hundreds of writers. This creates a whole new problem: You now have a database of 500+ writers waiting for you to sift through them and decide which ones are worth the hire. It would take you 2-3 days of intense work to go through all these writers and vet them yourself. Let’s be real - you don’t have time for that. Here’s what you can do instead: When sourcing writers, always get them to fill in a Google form (instead of DMing or emailing you). In this form, make sure to ask for 3 relevant written samples, a link to the writer’s portfolio page, and the writer’s rate per word. Create a SOP for evaluating writers. The criteria for evaluation should be: Level of English. Does the writer’s sample have any English mistakes? If so, they’re not a good fit. Quality of Samples. Are the samples long-form and engaging content or are they boring 500-word copy-pastes? Technical Knowledge. Has the writer written about a hard-to-explain topic before? Anyone can write about simple topics like traveling—you want to look for someone who knows how to research a new topic and explain it in a simple and easy-to-read way. If someone’s written about how to create a perfect cover letter, they can probably write about traveling, but the opposite isn’t true. Get your VA to evaluate the writer’s samples as per the criteria above and short-list writers that seem competent. If you sourced 500 writers, the end result of this process should be around 50 writers. You or your editor goes through the short-list of 50 writers and invites 5-10 for a (paid) trial task. The trial task is very important - you’ll sometimes find that the samples provided by the writer don’t match their writing level. SEO Tip #10. Use the Right Websites to Find Writers Not sure where to source your writers? Here are some ideas: ProBlogger \- Our #1 choice - a lot of quality writers frequent this website. LinkedIn \- You can headhunt content writers in specific locations. Upwork \- If you post a content gig, most writers are going to be awful. Instead, I recommend headhunting top writers instead. WeWorkRemotely \- Good if you’re looking to make a full-time remote hire. Facebook \- There are a ton of quality Facebook groups for writers. Some of our faves are Cult of Copy Job Board and Content Marketing Lounge. SEO Tip #11. Always Use Content Outlines When giving tasks to your writing team, you need to be very specific about the instructions you give them. Don’t just provide a keyword and tell them to “knock themselves out.” The writer isn’t a SEO expert; chances are, they’re going to mess it up big-time and talk about topics that aren’t related to the keyword you’re targeting. Instead, when giving tasks to writers, do it through content outlines. A content outline, in a nutshell, is a skeleton of the article they’re supposed to write. It includes information on: Target word count (aim for the same or 50% more the word count than that of the competition). Article title. Article structure (which sections should be mentioned and in what order). Related topics of keywords that need to be mentioned in the article. Content outline example in the URL in the post intro. SEO Tip #12. Focus on One Niche at a Time I used to work with this one client that had a SaaS consisting of a mixture of CRM, Accounting Software, and HRS. I had to pick whether we were going to focus on topics for one of these 3 niches or focus on all of them at the same time. I decided to do the former. Here’s why: When evaluating what to rank, Google considers the authority of your website. If you have 60 articles about accounting (most of which link to each other), you’re probably an authority in the niche and are more likely to get good rankings. If you have 20 sales, 20 HR, and 20 accounting articles, though, none of these categories are going to rank as well. It always makes more sense to first focus on a single niche (the one that generates the best ROI for your business), and then move on to the rest. This also makes it easier to hire writers - you hire writers specialized in accounting, instead of having to find writers who can pull off 3 unrelated topics. SEO Tip #13. Just Hire a VA Already It’s 2021 already guys—unless you have a virtual assistant, you’re missing out big-time. Since a lot of SEO tasks are very time-consuming, it really helps to have a VA around to take over. As long as you have solid SOPs in place, you can hire a virtual assistant, train them, and use them to free up your time. Some SEO tasks virtual assistants can help with are: Internal linking. Going through all your blog content and ensuring that they link to each other. Backlink prospecting. Going through hundreds of websites daily to find link opportunities. Uploading content on WordPress and ensuring that the content is optimized well for on-page SEO. SEO Tip #14. Use WordPress (And Make Your Life Easier) Not sure which CMS platform to use? 99% of the time, you’re better off with WordPress. It has a TON of plugins that will make your life easier. Want a drag & drop builder? Use Elementor. It’s cheap, efficient, extremely easy to learn, and comes jam-packed with different plugins and features. Wix, SiteGround, and similar drag & drops are pure meh. SEO Tip #15. Use These Nifty WordPress Plugins There are a lot of really cool WordPress plugins that can make your (SEO) life so much easier. Some of our favorites include: RankMath. A more slick alternative to YoastSEO. Useful for on-page SEO. Smush. App that helps you losslessly compress all images on your website, as well as enables lazy loading. WP Rocket. This plugin helps speed up your website pretty significantly. Elementor. Not a techie? This drag & drop plugin makes it significantly easier to manage your website. WP Forms. Very simple form builder. Akismet Spam Protection. Probably the most popular anti-spam WP plugin. Mammoth Docx. A plugin that uploads your content from a Google doc directly to WordPress. SEO Tip #16. No, Voice Search Is Still Not Relevant Voice search is not and will not be relevant (no matter what sensationalist articles might say). Sure, it does have its application (“Alexa, order me toilet paper please”), but it’s pretty niche and not relevant to most SEOs. After all, you wouldn’t use voice search for bigger purchases (“Alexa, order me a new laptop please”) or informational queries (“Alexa, teach me how to do accounting, thanks”). SEO Tip #17. SEO Is Obviously Not Dead I see these articles every year - “SEO is dead because I failed to make it work.” SEO is not dead and as long as there are people looking up for information/things online, it never will be. And no, SEO is not just for large corporations with huge budgets, either. Some niches are hypercompetitive and require a huge link-building budget (CBD, fitness, VPN, etc.), but they’re more of an exception instead of the rule. SEO Tip #18. Doing Local SEO? Focus on Service Pages If you’re doing local SEO, you’re better off focusing on local service pages than blog content. E.g. if you’re an accounting firm based in Boston, you can make a landing page about /accounting-firm-boston/, /tax-accounting-boston/, /cpa-boston/, and so on. Or alternatively, if you’re a personal injury law firm, you’d want to create pages like /car-accident-law-firm/, /truck-accident-law-firm/, /wrongful-death-law-firm/, and the like. Thing is, you don’t really need to rank on global search terms—you just won’t get leads from there. Even if you ranked on the term “financial accounting,” it wouldn’t really matter for your bottom line that much. SEO Tip #19. Engage With the SEO Community The SEO community is (for the most part) composed of extremely helpful and friendly people. There are a lot of online communities (including this sub) where you can ask for help, tips, case studies, and so on. Some of our faves are: This sub :) SEO Signals Lab (FB Group) Fat Graph Content Ops (FB Group) Proper SEO Group (FB Group) BigSEO Subreddit SEO Tip #20. Test Keywords Before Pursuing Them You can use Google ads to test how profitable any given keyword is before you start trying to rank for it. The process here is: Create a Google Ads account. Pick a keyword you want to test. Create a landing page that corresponds to the search intent behind the keyword. Allocate an appropriate budget. E.g. if you assume a conversion rate of 2%, you’d want to buy 100+ clicks. If the CPC is 2 USD, then the right budget would be 200 USD plus. Run the ads! If you don’t have the budget for this, you can still use the average CPC for the keyword to estimate how well it’s going to convert. If someone is willing to bid 10 USD to rank for a certain keyword, it means that the keyword is most probably generating pretty good revenue/conversions. SEO Tip #21. Test & Improve SEO Headlines Sometimes, you’ll see that you’re ranking in the top 3 positions for your search query, but you’re still not driving that much traffic. “What’s the deal?” you might be asking. Chances are, your headline is not clickable enough. Every 3-4 months, go through your Google Search Console and check for articles that are ranking well but not driving enough traffic. Then, create a Google sheet and include the following data: Targeted keyword Page link CTR (for the last 28 days) Date when you implemented the new title Old title New title New CTR (for the month after the CTR change was implemented) From then on, implement the new headline and track changes in the CTR. If you don’t reach your desired result, you can always test another headline. SEO Tip #22. Longer Content Isn’t Always Better Content You’ve probably heard that long-form content is where it’s at in 2021. Well, this isn’t always the case. Rather, this mostly depends on the keyword you’re targeting. If, for example, you’re targeting the keyword “how to tie a tie,” you don’t need a long-ass 5,000-word mega-guide. In such a case, the reader is looking for something that can be explained in 200-300 words and if your article fails to do this, the reader will bounce off and open a different page. On the other hand, if you’re targeting the keyword “how to write a CV,” you’ll need around 4,000 to 5,000 words to adequately explain the topic and, chances are, you won’t rank with less. SEO Tip #23. SEO is Not All About Written Content More often than not, when people talk about SEO they talk about written blog content creation. It’s very important not to forget, though, that blog content is not end-all-be-all for SEO. Certain keywords do significantly better with video content. For example, if the keyword is “how to do a deadlift,” video content is going to perform significantly better than blog content. Or, if the keyword is “CV template,” you’ll see that a big chunk of the rankings are images of the templates. So, the lesson here is, don’t laser-focus on written content—keep other content mediums in mind, too. SEO Tip #24. Write For Your Audience It’s very important that your content resonates well with your target audience. If, for example, you’re covering the keyword “skateboard tricks,” you can be very casual with your language. Heck, it’s even encouraged! Your readers are Googling the keyword in their free time and are most likely teens or in their early 20s. Meaning, you can use informal language, include pop culture references, and avoid complicated language. Now, on the other hand, if you’re writing about high-level investment advice, your audience probably consists of 40-something suit-and-ties. If you include Rick & Morty references in your article, you'll most likely lose credibility and the Googler, who will go to another website. Some of our best tips on writing for your audience include: Define your audience. Who’s the person you’re writing for? Are they reading the content at work or in their free time? Keep your reader’s level of knowledge in mind. If you’re covering an accounting 101 topic, you want to cover the topic’s basics, as the reader is probably a student. If you’re writing about high-level finance, though, you don’t have to teach the reader what a balance sheet is. More often than not, avoid complicated language. The best practice is to write on a 6th-grade level, as it’s understandable for anyone. Plus, no one wants to read Shakespeare when Googling info online (unless they’re looking for Shakespeare's work, of course). SEO Tip #25. Create Compelling Headlines Want to drive clicks to your articles? You’ll need compelling headlines. Compare the following headline: 101 Productivity Tips \[To Get Things Done in 2021\] With this one: Productivity Tips Guide Which one would you click? Data says it’s the first! To create clickable headlines, I recommend you include the following elements: Keyword. This one’s non-negotiable - you need to include the target keyword in the headline. Numbers. If Buzzfeed taught us anything, it’s that people like to click articles with numbers in their titles. Results. If I read your article, what’s going to be the end result? E.g. “X Resume tips (to land the job)”.* Year (If Relevant). Adding a year to your title shows that the article is recent (which is relevant for some specific topics). E.g. If the keyword is “Marketing Trends,” I want to know marketing trends in 2021, not in 2001. So, adding a year in the title makes the headline more clickable. SEO Tip #26. Make Your Content Visual How good your content looks matters, especially if you're in a competitive niche. Here are some tips on how to make your content as visual as possible: Aim for 2-4 sentences per paragraph. Avoid huge blocks of text. Apply a 60-65% content width to your blog pages. Pick a good-looking font. I’d recommend Montserrat, PT Sans, and Roboto. Alternatively, you can also check out your favorite blogs, see which fonts they’re using, and do the same. Use a reasonable font size. Most top blogs use font sizes ranging from 16 pt to 22 pt. Add images when possible. Avoid stock photos, though. No one wants to see random “office people smiling” scattered around your blog posts. Use content boxes to help convey information better. Content boxes example in the URL in the intro of the post. SEO Tip #27. Ditch the Skyscraper Technique Already Brian Dean’s skyscraper technique is awesome and all, but the following bit really got old: “Hey \[name\], I saw you wrote an article. I, too, wrote an article. Please link to you?” The theory here is, if your content is good, the person will be compelled to link to it. In practice, though, the person really, really doesn’t care. At the end of the day, there’s no real incentive for the person to link to your content. They have to take time out of their day to head over to their website, log in to WordPress, find the article you mentioned, and add a link... Just because some stranger on the internet asked them to. Here’s something that works much better: Instead of fake compliments, be very straightforward about what you can offer them in exchange for that link. Some things you can offer are: A free version of your SaaS. Free product delivered to their doorstep. Backlink exchange. A free backlink from your other website. Sharing their content to your social media following. Money. SEO Tip #28. Get the URL Slug Right for Seasonal Content If you want to rank on a seasonal keyword, there are 2 ways to do this. If you want your article to be evergreen (i.e. you update it every year with new information), then your URL should not contain the year. E.g. your URL would be /saas-trends/, and you simply update the article’s contents+headline each year to keep it timely. If you’re planning on publishing a new trends report annually, though, then you can add a year to the URL. E.g. /saas-trends-2020/ instead of /saas-trends/. SEO Tip #29. AI Content Tools Are a Mixed Bag Lots of people are talking about AI content tools these days. Usually, they’re either saying: “AI content tools are garbage and the output is horrible,” Or: “AI content tools are a game-changer!” So which one is it? The truth is somewhere in-between. In 2021, AI content writing tools are pretty bad. The output you’re going to get is far from something you can publish on your website. That said, some SEOs use such tools to get a very, very rough draft of the article written, and then they do intense surgery on it to make it usable. Should you use AI content writing tools? If you ask me, no - it’s easier to hire a proficient content writer than spend hours salvaging AI-written content. That said, I do believe that such tools are going to get much better years down the line. This one was, clearly, more of a personal opinion than a fact. I’d love to hear YOUR opinion on AI content tools! Are they a fad, or are they the future of content creation? Let me know in the comments. SEO Tip #30. Don’t Overdo it With SEO Tools There are a lot of SEO tools out there for pretty much any SEO function. Keyword research, link-building, on-page, outreach, technical SEO, you name it! If you were to buy most of these tools for your business, you’d easily spend 4-figures on SEO tools per month. Luckily, though, you don’t actually need most of them. At the end of the day, the only must-have SEO tools are: An SEO Suite (Paid). Basically SEMrush or Ahrefs. Both of these tools offer an insane number of features - backlink analysis, keyword research, and a ton of other stuff. Yes, 99 USD a month is expensive for a tool. But then again, if you value your time 20 USD/hour and this tool saves you 6 hours, it's obviously worth it, right? On-Page SEO Tool (Free). RankMath or Yoast. Basically, a tool that's going to help you optimize web pages or blog posts as per SEO best practices. Technical SEO Tool (Freemium). You can use ScreamingFrog to crawl your entire website and find technical SEO problems. There are probably other tools that also do this, but ScreamingFrog is the most popular option. The freemium version of the tool only crawls a limited number of pages (500 URLs, to be exact), so if your website is relatively big, you'll need to pay for the tool. Analytics (Free). Obviously, you'll need Google Analytics (to track website traffic) and Google Search Console (to track organic traffic, specifically) set up on your website. Optionally, you can also use Google Track Manager to better track how your website visitors interact with the site. MozBar (Free). Chrome toolbar that lets you simply track the number of backlinks on Google Search Queries, Domain Authority, and a bunch of other stuff. Website Speed Analysis (Free). You can use Google Page Speed Insights to track how fast your website loads, as well as how mobile-friendly it is. Outreach Tool (Paid). Tool for reaching out to prospects for link-building, guest posting, etc. There are about a dozen good options for this. Personally, I like to use Snov for this. Optimized GMB Profile (Free). Not a tool per se, but if you're a local business, you need to have a well-optimized Google My Business profile. Google Keyword Planner (Free). This gives you the most reliable search volume data of all the tools. So, when doing keyword research, grab the search volume from here. Tool for Storing Keyword Research (Free). You can use Google Sheets or AirTable to store your keyword research and, at the same time, use it as a content calendar. Hemingway App (Free). Helps keep your SEO content easy to read. Spots passive voice, complicated words, etc. Email Finder (Freemium). You can use a tool like Hunter to find the email address of basically anyone on the internet (for link-building or guest posting purposes). Most of the tools that don’t fit into these categories are 100% optional. SEO Tip #31. Hiring an SEO? Here’s How to Vet Them Unless you’re an SEO pro yourself, hiring one is going to be far from easy. There’s a reason there are so many “SEO experts” out there - for the layman, it’s very hard to differentiate between someone who knows their salt and a newbie who took an SEO course, like, last week. Here’s how you can vet both freelance and full-time SEOs: Ask for concrete traffic numbers. The SEO pro should give you the exact numbers on how they’ve grown a website in the past - “100% SEO growth in 1 year” doesn’t mean much if the growth is from 10 monthly traffic to 20. “1,000 to 30,000” traffic, on the other hand, is much better. Ask for client names. While some clients ask their SEOs to sign an NDA and not disclose their collaboration, most don’t. If an SEO can’t name a single client they’ve worked with in the past, that’s a red flag. Make sure they have the right experience. Global and local SEO have very different processes. Make sure that the SEO has experience with the type of SEO you need. Make sure you’re looking for the right candidate. SEO pros can be content writers, link-builders, web developers, or all of the above simultaneously. Make sure you understand which one you need before making the hire. If you’re looking for someone to oversee your content ops, you shouldn’t hire a technical SEO expert. Look for SEO pros in the right places. Conventional job boards are overrated. Post your job ads on SEO communities instead. E.g. this sub, bigseo, SEO Signals Facebook group, etc. SEO Tip #32. Blog Post Not Ranking? Follow This Checklist I wanted to format the post natively for Reddit, but it’s just SO much better on Notion. Tl;dr, the checklist covers every reason your post might not be ranking: Search intent mismatch. Inferior content. Lack of internal linking. Lack of backlinks. And the like. Checklist URL at the intro of the post. SEO Tip #33. Avoid BS Link-Building Tactics The only type of link-building that works is building proper, quality links from websites with a good backlink profile and decent organic traffic. Here’s what DOESN’T work: Blog comment links Forum spam links Drive-by Reddit comment/post links Web 2.0 links Fiverr “100 links for 10 bucks” bs If your “SEO agency” says they’re doing any of the above instead of actually trying to build you links from quality websites, you’re being scammed. SEO Tip #34. Know When to Use 301 and 302 Redirects When doing redirects, it’s very important to know the distinction between these two. 301 is a permanent page redirect and passes on link juice. If you’re killing off a page that has backlinks, it’s better to 301 it to your homepage so that you don’t lose the link juice. If you simply delete a page, it’s going to be a 404, and the backlink juice is lost forever. 302 is a temporary page redirect and doesn’t pass on link juice. If the redirect is temporary, you do a 302. E.g. you want to test how well a new page is going to perform w/ your audience. SEO Tip #35. Social Signals Matter (But Not How You Think) Social signals are NOT a ranking factor. And yet, they can help your content rank on Google’s front page. Wondering what the hell am I talking about? Here’s what’s up: As I said, social signals are not a ranking factor. It’s not something Google takes into consideration to decide whether your article should rank or not. That said, social signals CAN lead to your article ranking better. Let’s say your article goes viral and gets around 20k views within a week. A chunk of these viewers are going to forget your domain/link and they’re going to look up the topic on Google via your chosen keyword + your brand name. The amount of people looking for YOUR keyword and exclusively picking your result over others is going to make Google think that your content is satisfying search intent better than the rest, and thus, reward you with better ranking. SEO Tip #36. Run Remarketing Ads to Lift Organic Traffic Conversions Not satisfied with your conversion rates? You can use Facebook ads to help increase them. Facebook allows you to do something called “remarketing.” This means you can target anyone that visited a certain page (or multiple pages) on your website and serve them ads on Facebook. There are a TON of ways you can take advantage of this. For example, you can target anyone that landed on a high buyer intent page and serve them ads pitching your product or a special offer. Alternatively, you can target people who landed on an educational blog post and offer them something to drive them down the funnel. E.g. free e-book or white paper to teach them more about your product or service. SEO Tip #37. Doing Local SEO? Follow These Tips Local SEO is significantly different from global SEO. Here’s how the two differ (and what you need to do to drive local SEO results): You don’t need to publish content. For 95% of local businesses, you only want to rank for keywords related to your services/products, you don’t actually need to create educational content. You need to focus more on reviews and citation-building. One of Google Maps’ biggest ranking factors is the of reviews your business has. Encourage your customers to leave a review if they enjoyed your product/service through email or real-life communication. You need to create service pages for each location. As a local business, your #1 priority is to rank for keywords around your service. E.g. If you're a personal injury law firm, you want to optimize your homepage for “personal injury law firm” and then create separate pages for each service you provide, e.g. “car accident lawyer,” “motorcycle injury law firm,” etc. Focus on building citations. Being listed on business directories makes your business more trustworthy for Google. BrightLocal is a good service for this. You don’t need to focus as much on link-building. As local SEO is less competitive than global, you don’t have to focus nearly as much on building links. You can, in a lot of cases, rank with the right service pages and citations. SEO Tip #38. Stop Ignoring the Outreach Emails You’re Getting (And Use Them to Build Your Own Links) Got a ton of people emailing you asking for links? You might be tempted to just send them all straight to spam, and I don’t blame you. Outreach messages like “Hey Dr Jigsaw, your article is A+++ amazing! ...can I get a backlink?” can get hella annoying. That said, there IS a better way to deal with these emails: Reply and ask for a link back. Most of the time, people who send such outreach emails are also doing heavy guest posting. So, you can ask for a backlink from a 3rd-party website in exchange for you mentioning their link in your article. Win-win! SEO Tip #39. Doing Internal Linking for a Large Website? This’ll Help Internal linking can get super grueling once you have hundreds of articles on your website. Want to make the process easier? Do this: Pick an article you want to interlink on your website. For the sake of the example, let’s say it’s about “business process improvement.” Go on Google and look up variations of this keyword mentioned on your website. For example: Site:\[yourwebsite\] “improve business process” Site:\[yourwebsite\] “improve process” Site:\[yourwebsite\] “process improvement” The above queries will find you the EXACT articles where these keywords are mentioned. Then, all you have to do is go through them and include the links. SEO Tip #40. Got a Competitor Copying Your Content? File a DMCA Notice Fun fact - if your competitors are copying your website, you can file a DMCA notice with Google. That said, keep in mind that there are consequences for filing a fake notice.

Started a content marketing agency 8 years ago - $0 to $7,863,052 (2025 update)
reddit
LLM Vibe Score0
Human Vibe Score0.882
mr_t_forhireThis week

Started a content marketing agency 8 years ago - $0 to $7,863,052 (2025 update)

Hey friends, My name is Tyler and for the past 8 years, I’ve been documenting my experience building a content marketing agency called Optimist. Year 1 — 0 to $500k ARR Year 2 — $500k to $1MM ARR Year 3 — $1MM ARR to $1.5MM(ish) ARR Year 4 — $3,333,686 Revenue Year 5 — $4,539,659 Revenue Year 6 — $5,974,324 Revenue Year 7 - $6,815,503 Revenue (Edit: Seems like links are banned now. You can check my post history for all of my previous updates with lessons and learnings.) How Optimist Works First, an overview/recap of the Optimist business model: We operate as a “collective” of full time/professional freelancers Everyone aside from me is a contractor Entirely remote/distributed team We pay freelancers a flat fee for most work, working out to roughly $65-100/hour. Clients pay us a flat monthly fee for full-service content marketing (research, strategy, writing, editing, design/photography, reporting and analytics, targeted linkbuilding, and more)\ Packages range in price from \~$10-20k/mo \This is something we are revisiting now* The Financials In 2024, we posted $1,032,035.34 in revenue. This brings our lifetime revenue to $7,863,052. Here’s our monthly revenue from January 2017 to December of 2024. (Edit: Seems like I'm not allowed to link to the chart.) The good news: Revenue is up 23% YoY. EBITDA in Q4 trending up 1-2 points. We hosted our first retreat in 4 years, going to Ireland with about half the team. The bad news: Our revenue is still historically low. At $1MM for the year, we’re down about 33% from our previous years over $1.5MM. Revenue has been rocky. It doesn’t feel like we’ve really “recovered” from the bumps last year. The trend doesn’t really look great. Even though, anecdotally, it feels like we are moving in a good direction. EBITDA is still hovering at around 7%. Would love to get that closer to 20%. (For those who may ask: I’m calculating EBITDA after paying taxes and W2 portion of my income.) — Almost every year, my update starts the same way: This has been a year of growth and change. Both for my business—and me personally. 2024 was no different. I guess that tells you something about entrepreneurship. It’s a lot more like sailing a ship than driving a car. You’re constantly adapting, tides are shifting, and any blip of calm is usually just a moment before the next storm. As with past years, there’s a lot to unpack from the last 12 months. Here we go again. Everything is Burning In the last 2 years, everything has turned upside down in the world of content and SEO. Back in 2020, we made a big decision to re-position the agency. (See post history) We decided to narrow our focus to our most successful, profitable, and consistent segment of clients and re-work our entire operation to focus on serving them. We defined our ICP as: \~Series A ($10mm+ funding) with 6-12 months runway to scale organic as a channel Product-led company with “simple” sales cycle involving fewer stakeholders Demonstrable opportunity to use SEO to drive business growth Our services: Content focused on growing organic search (SEO) Full-service engagements that included research, planning, writing, design, reporting And our engagement structure: Engaged directly with an executive; ownership over strategy and day-to-day execution 1-2 points of contact or stakeholders Strategic partner that drives business growth (not a service vendor who makes content) Most importantly, we decided that we were no longer going to offer a broader range of content that we used to sell. That included everything from thought leadership content to case studies and ebooks. We doubled-down on “SEO content” for product-led SaaS companies. And this worked phenomenally for us. We started bringing on more clients than ever. We developed a lot of internal system and processes that helped us scale and take on more work than we’ve ever had and drive great outcomes for our ideal clients. But in 2023 and 2024, things started going awry. One big change, of course, was the rise of AI. Many companies and executives (and writers) feel that AI can write content just as well as an agency like ours. That made it a lot harder to sell a $10,000 per month engagement when they feel like the bulk of the work could be “done for free.” (Lots of thoughts on this if you want my opinions.) But it wasn’t just that. Google also started tinkering with their algorithm, introducing new features like AI Overviews, and generally changing the rules of the game. This created 3 big shifts in our world: The perceived value of content (especially “SEO content”) dropped dramatically in many people’s minds because of AI’s writing capabilities SEO became less predictable as a source of traffic and revenue It’s harder than ever for startups and smaller companies to rank for valuable keywords (let alone generate any meaningful traffic or revenue from them) The effect? The middle of the content market has hollowed out. People—like us—providing good, human-crafted content aimed on driving SEO growth saw a dramatic decline in demand. We felt it all year. Fewer and fewer leads. The leads we did see usually scoffed at our prices. They were indexing us against the cost of content mills and mass-produced AI articles. It was a time of soul-searching and looking for a way forward. I spent the first half of the year convinced that the only way to survive was to run toward the fire. We have to build our own AI workflows. We have to cut our rates internally. We have to get faster and cheaper to stay competitive with the agencies offering the same number of deliverables for a fraction of our rates. It’s the only way forward. But then I asked myself a question… Is this the game I actually want to play? As an entrepreneur, do I want to run a business where I’m competing mostly on price and efficiency rather than quality and value? Do I want to hop into a race toward cheaper and cheaper content? Do I want to help people chase a dwindling amount of organic traffic that’s shrinking in value? No. That’s not the game I want to play. That’s not a business I want to run. I don’t want to be in the content mill business. So I decided to turn the wheel—again. Repositioning Part II: Electric Boogaloo What do you do when the whole world shifts around you and the things that used to work aren’t working anymore? You pivot. You re-position the business and move in another direction. So that’s what we decided to do. Again. There was only one problem: I honestly wasn’t sure what opportunities existed in the content marketing industry outside of what we were already doing. We lived in a little echo chamber of startups and SEO. It felt like the whole market was on fire and I had fight through the smoke to find an escape hatch. So I started making calls. Good ol’ fashioned market research. I reached out to a few dozen marketing and content leaders at a bunch of different companies. I got on the phone and just asked lots of questions about their content programs, their goals, and their pain points. I wanted to understand what was happening in the market and how we could be valuable. And, luckily, this process really paid off. I learned a lot about the fragmentation happening across content and how views were shifting. I noticed key trends and how our old target market really wasn’t buying what we were selling. Startups and small companies are no longer willing to invest in an agency like ours. If they were doing content and SEO at all, they were focused entirely on using AI to scale output and minimize costs. VC money is still scarce and venture-backed companies are more focused on profitability than pure growth and raising another round. Larger companies (\~500+ employees) are doing more content than ever and drowning in content production. They want to focus on strategy but can barely tread water keeping up with content requests from sales, demand gen, the CEO, and everyone else. Many of the companies still investing in content are looking at channels and formats outside of SEO. Things like thought leadership, data reports, interview-driven content, and more. They see it as a way to stand out from the crowd of “bland SEO content.” Content needs are constantly in flux. They range from data reports and blog posts to product one-pagers. The idea of a fixed-scope retainer is a total mismatch for the needs of most companies. All of this led to the logical conclusion: We were talking to the wrong people about the wrong things\.\ Many companies came to one of two logical conclusions: SEO is a risky bet, so it’s gotta be a moonshot—super-low cost with a possibility for a big upside (i.e., use AI to crank out lots of content. If it works, great. If it doesn’t, then at least we aren’t out much money.) SEO is a risky bet, so we should diversify into other strategies and channels to drive growth (i.e., shift our budget from SEO and keyword-focused content to video, podcasts, thought leadership, social, etc) Unless we were going to lean into AI and dramatically cut our costs and rates, our old buyers weren’t interested. And the segment of the market that needs our help most are looking primarily for production support across a big range of content types. They’re not looking for a team to run a full-blown program focused entirely on SEO. So we had to go back to the drawing board. I’ve written before about our basic approach to repositioning the business. But, ultimately it comes down to identifying our unique strengths as a team and then connecting them to needs in the market. After reviewing the insights from my discussions and taking another hard look at our business and our strengths, I decided on a new direction: Move upmarket: Serve mid-size to enterprise businesses with \~500-5,000 employees instead of startups Focus on content that supports a broader range of business goals instead of solely on SEO and organic growth (e.g., sales, demand gen, brand, etc) Shift back to our broader playbook of content deliverables, including thought leadership, data studies, and more Focus on content execution and production to support an internally-directed content strategy across multiple functions In a way, it’s sort of a reverse-niche move. Rather than zooming in specifically on driving organic growth for startups, we want to be more of an end-to-end content production partner that solves issues of execution and operations for all kinds of content teams. It’s early days, but the response here has been promising. We’ve seen an uptick in leads through Q4. And more companies in our pipeline fit the new ICP. They’re bigger, often have more budget. (But they move more slowly). We should know by the end of the quarter if this maneuver is truly paying off. Hopefully, this will work out. Hopefully our research and strategy are right and we’ll find a soft landing serving a different type of client. If it doesn’t? Then it will be time to make some harder decisions. As I already mentioned, I’m not interested in the race to the bottom of AI content. And if that’s the only game left in town, then it might be time to think hard about a much bigger change. — To be done: Build new content playbooks for expanded deliverables Build new showcase page for expanded deliverables Retooling the Operation It’s easy to say we’re doing something new. It’s a lot harder to actually do it—and do it well. Beyond just changing our positioning, we have to do open-heart surgery on the entire content operation behind the scenes. We need to create new systems that work for a broader range of content types, formats, and goals. Here’s the first rub: All of our workflows are tooled specifically for SEO-focused content. Every template, worksheet, and process that we’ve built and scaled in the last 5 years assumes that the primary goal of every piece of content is SEO. Even something as simple as requiring a target keyword is a blocker in a world where we’re not entirely focused on SEO. This is relatively easy to fix, but it requires several key changes: Update content calendars to make keywords optional Update workflows to determine whether we need an optimization report for each deliverable Next, we need to break down the deliverables into parts rather than a single line item. In our old system, we would plan content as a single row in a Content Calendar spreadsheet. It was a really wide sheet with lots of fields where we’d define the dimensions of each individual article. This was very efficient and simple to follow. But every article had the same overall scope when it came to the workflow. In Asana (our project management tool), all of the steps in the creation were strung together in a single task. We would create a few basic templates for each client, and then each piece would flow through the same steps: Briefing Writing Editing Design etc. If we had anything that didn’t fit into the “standard” workflow, we’d just tag it in the calendar with an unofficial notation \[USING BRACKETS\]. It worked. But it wasn’t ideal. Now we need the steps to be more modular. Imagine, for example, a client asks us to create a mix of deliverables: 1 article with writing + design 1 content brief 1 long-form ebook with an interview + writing + design Each of these would require its own steps and its own workflow. We need to break down the work to accommodate for a wider variety of workflows and variables. This means we need to update the fields and structure of our calendar to accommodate for the new dimensions—while also keeping the planning process simple and manageable. This leads to the next challenge: The number of “products” that we’re offering could be almost infinite. Just looking at the example scope above, you can mix and match all of these different building blocks to create a huge variety of different types of work, each requiring its own workflow. This is part of the reason we pivoted away from this model to focus on a productized, SEO-focused content service back in 2020. Take something as simple as a case study. On the surface, it seems like one deliverable that can be easily scoped and priced, right? Well, unpack what goes into a case study: Is there already source material from the customer or do we need to conduct an interview? How long is it? Is it a short overview case study or a long-form narrative? Does it need images and graphics? How many? Each of these variables opens up 2-3 possibilities. And when you combine them, we end up with something like 10 possible permutations for this single type of deliverable. It gets a bit messy. But not only do we have to figure out how to scope and price all for all of these variables, we also have to figure out how to account for these variables in the execution. We have to specify—for every deliverable—what type it is, how long, which steps are involved and not involved, the timeline for delivery, and all of the other factors. We’re approaching infinite complexity, here. We have to figure out a system that allows for a high level of flexibility to serve the diverse needs of our clients but is also productized enough that we can build workflows, process, and templates to deliver the work. I’ve spent the last few months designing that system. Failed Attempt #1: Ultra-Productization In my first pass, I tried to make it as straight forward as possible. Just sit down, make a list of all of the possible deliverables we could provide and then assign them specific scopes and services. Want a case study? Okay that’ll include an interview, up to 2,000 words of content, and 5 custom graphics. It costs $X. But this solution quickly fell apart when we started testing it against real-world scenarios. What if the client provided the brief instead of us creating one? What if they didn’t want graphics? What if this particular case study really needs to be 3,000 words but all of the others should be 2,000? In order for this system to work, we’d need to individual scope and price all of these permutations of each productized service. Then we’d need to somehow keep track of all of these and make sure that we accurately scope, price, and deliver them across dozens of clients. It’s sort of like a restaurant handling food allergies by creating separate versions of every single dish to account for every individual type of allergy. Most restaurants have figured out that it makes way more sense to have a “standard” and an “allergy-free” version. Then you only need 2 options to cover 100% of the cases. Onto the next option. Failed Attempt #2: Deliverable-Agnostic Services Next, I sat down with my head of Ops, Katy, to try to map it out. We took a big step back and said: Why does the deliverable itself even matter? At the end of the day, what we’re selling is just a few types of work (research, writing, editing, design, etc) that can be packaged up in an infinite number of ways. Rather than try to define deliverables, shouldn’t we leave it open ended for maximum flexibility? From there, we decided to break down everything into ultra-modular building blocks. We started working on this super complex system of modular deliverables where we would have services like writing, design, editing, etc—plus a sliding scale for different scopes like the length of writing or the number of images. In theory, it would allow us to mix and match any combination of services to create custom deliverables for the client. In fact, we wanted the work to be deliverable-agnostic. That way we could mold it to fit any client’s needs and deliver any type of content, regardless of the format or goal. Want a 5,000-word case study with 15 custom graphics? That’ll be $X. Want a 2,000-word blog post with an interview and no visuals? $Y. Just want us to create 10 briefs, you handle the writing, and we do design? It’s $Z. Again, this feels like a reasonable solution. But it quickly spiraled out of amuck. (That’s an Office reference.) For this to work, we need to have incredibly precise scoping process for every single deliverable. Before we can begin work (or even quote a price), we need to know pretty much the exact word count of the final article, for example. In the real world? This almost never happens. The content is as long as the content needs to be. Clients rarely know if the blog post should be 2,000 words or 3,000 words. They just want good content. We have a general ballpark, but we can rarely dial it in within just 1,000 words until we’ve done enough research to create the brief. Plus, from a packaging and pricing perspective, it introduces all kind of weird scenarios where clients will owe exactly $10,321 for this ultra-specific combination of services. We were building an open system that could accommodate any and all types of potential deliverables. On the face that seems great because it makes us incredibly flexible. In reality, the ambiguity actually works against us. It makes it harder for us to communicate to clients clearly about what they’ll get, how much it will cost, and how long it will take. That, of course, also means that it hurts our client relationships. (This actually kind of goes back to my personal learnings, which I’ll mention in a bit. I tend to be a “let’s leave things vague so we don’t have to limit our options” kind of person. But I’m working on fixing this to be more precise, specific, and clear in everything that we do.) Dialing It In: Building a Closed System We were trying to build an open system. We need to build a closed system. We need to force clarity and get specific about what we do, what we don’t do, and how much it all costs. Then we need a system to expand on that closed system—add new types of deliverables, new content playbooks, and new workflows if and when the need arises. With that in mind, we can start by mapping out the key dimensions of any type of deliverable that we would ever want to deliver. These are the universal dimensions that determine the scope, workflow, and price of any deliverable—regardless of the specific type output. Dimensions are: Brief scope Writing + editing scope Design scope Interview scope Revision (rounds) Scope, essentially, just tells us how many words, graphics, interviews, etc are required for the content we’re creating. In our first crack at the system, we got super granular with these scopes. But to help force a more manageable system, we realized that we didn’t need tiny increments for most of this work. Instead, we just need boundaries—you pay $X for up to Y words. We still need some variability around the scope of these articles. Obviously, most clients won’t be willing to pay the same price for a 1,000-word article as a 10,000-word article. But we can be smarter about the realistic break points. We boiled it down to the most common ranges: (Up to) 250 words 1,000 words 3,000 words 6,000 words 10,000 words This gives us a much more manageable number of variables. But we still haven’t exactly closed the system. We need one final dimension: Deliverable type. This tells us what we’re actually building with these building blocks. This is how we’ll put a cap on the potentially infinite number of combinations we could offer. The deliverable type will define what the final product should look like (e.g., blog post, case study, ebook, etc). And it will also give us a way to put standards and expectations around different types of deliverables that we want to offer. Then we can expand on this list of deliverables to offer new services. In the mean time, only the deliverables that we have already defined are, “on the menu,” so to speak. If a client comes to us and asks for something like a podcast summary article (which we don’t currently offer), we’ll have to either say we can’t provide that work or create a new deliverable type and define the dimensions of that specific piece. But here’s the kicker: No matter the deliverable type, it has to still fit within the scopes we’ve already defined. And the pricing will be the same. This means that if you’re looking for our team to write up to 1,000 words of content, it costs the same amount—whether it’s a blog post, an ebook, a LinkedIn post, or anything else. Rather than trying to retool our entire system to offer this new podcast summary article deliverable, we’ll just create the new deliverable type, add it to the list of options, and it’s ready to sell with the pre-defined dimensions we’ve already identified. To do: Update onboarding workflow Update contracts and scope documents Dial in new briefing process Know Thyself For the last year, I’ve been going through personal therapy. (Huge shout out to my wife, Laura, for her support and encouragement throughout the process.) It’s taught me a lot about myself and my tendencies. It’s helped me find some of my weaknesses and think about how I can improve as a person, as a partner, and as an entrepreneur. And it’s forced me to face a lot of hard truths. For example, consider some of the critical decisions I’ve made for my business: Unconventional freelance “collective” model No formal management structure Open-ended retainers with near-infinite flexibility General contracts without defined scope “Take it or leave it” approach to sales and marketing Over the years, I’ve talked about almost everything on this list as a huge advantage. I saw these things as a reflection of how I wanted to do things differently and better than other companies. But now, I see them more as a reflection of my fears and insecurities. Why did I design my business like this? Why do I want so much “flexibility” and why do I want things left open-ended rather than clearly defined? One reason that could clearly explain it: I’m avoidant. If you’re not steeped in the world of therapy, this basically means that my fight or flight response gets turned all the way to “flight.” If I’m unhappy or uncomfortable, my gut reaction is usually to withdraw from the situation. I see commitment and specificity as a prelude to future conflict. And I avoid conflict whenever possible. So I built my business to minimize it. If I don’t have a specific schedule of work that I’m accountable for delivering, then we can fudge the numbers a bit and hope they even out in the end. If I don’t set a specific standard for the length of an article, then I don’t have to let the client know when their request exceeds that limit. Conflict….avoided? Now, that’s not to say that everything I’ve built was wrong or bad. There is a lot of value in having flexibility in your business. For example, I would say that our flexible retainers are, overall, an advantage. Clients have changing needs. Having flexibility to quickly adapt to those needs can be a huge value add. And not everything can be clearly defined upfront (at least not without a massive amount of time and work just to decide how long to write an article). Overly-rigid structures and processes can be just as problematic as loosey-goosey ones. But, on the whole, I realized that my avoidant tendencies and laissez faire approach to management have left a vacuum in many areas. The places where I avoided specificity were often the places where there was the most confusion, uncertainty, and frustration from the team and from clients. People simply didn’t know what to expect or what was expected of them. Ironically, this often creates the conflict I’m trying to avoid. For example, if I don’t give feedback to people on my team, then they feel uneasy about their work. Or they make assumptions about expectations that don’t match what I’m actually expecting. Then the client might get upset, I might get upset, and our team members may be upset. Conflict definitely not avoided. This happens on the client side, too. If we don’t define a specific timeline when something will be delivered, the client might expect it sooner than we can deliver—creating frustration when we don’t meet their expectation. This conflict actually would have been avoided if we set clearer expectations upfront. But we didn’t do that. I didn’t do that. So it’s time to step up and close the gaps. Stepping Up and Closing the Gaps If I’m going to address these gaps and create more clarity and stability, I have to step up. Both personally and professionally. I have to actually face the fear and uncertainty that drives me to be avoidant. And then apply that to my business in meaningful ways that aren’t cop-out ways of kinda-sorta providing structure without really doing it. I’ve gotta be all in. This means: Fill the gaps where I rely on other people to do things that aren’t really their job but I haven’t put someone in place to do it Set and maintain expectations about our internal work processes, policies, and standards Define clear boundaries on things like roles, timelines, budgets, and scopes Now, this isn’t going to happen overnight. And just because I say that I need to step up to close these gaps doesn’t mean that I need to be the one who’s responsible for them (at least not forever). It just means that, as the business leader, I need to make sure the gaps get filled—by me or by someone else who has been specifically charged with owning that part of the operation. So, this is probably my #1 focus over the coming quarter. And it starts by identifying the gaps that exist. Then, step into those gaps myself, pay someone else to fill that role, or figure out how to eliminate the gap another way. This means going all the way back to the most basic decisions in our business. One of the foundational things about Optimist is being a “different kind” of agency. I always wanted to build something that solved for the bureaucracy, hierarchy, and siloed structure of agencies. If a client has feedback, they should be able to talk directly to the person doing the work rather than going through 3 layers of account management and creative directors. So I tried to be clever. I tried to design all kinds of systems and processes that eliminated these middle rungs. (In retrospect, what I was actually doing was designing a system that played into my avoidant tendencies and made it easy to abdicate responsibility for lots of things.) Since we didn’t want to create hierarchy, we never implemented things like Junior and Senior roles. We never hired someone to manage or direct the individual creatives. We didn’t have Directors or VPs. (Hell, we barely had a project manager for the first several years of existence.) This aversion to hierarchy aligned with our values around elevating ownership and collective contribution. I still believe in the value a flat structure. But a flat structure doesn’t eliminate the complexity of a growing business. No one to review writers and give them 1:1 feedback? I guess I’ll just have to do that….when I have some spare time. No Content Director? Okay, well someone needs to manage our content playbooks and roll out new ones. Just add it to my task list. Our flat structure didn’t eliminate the need for these roles. It just eliminated the people to do them. All of those unfilled roles ultimately fell back on me or our ops person, Katy. Of course, this isn’t the first time we’ve recognized this. We’ve known there were growing holes in our business as it’s gotten bigger and more complex. Over the years, we’ve experimented with different ways to solve for it. The Old Solution: Distributed Ops One system we designed was a “distributed ops” framework. Basically, we had one person who was the head of ops (at the time, we considered anything that was non-client-facing to be “ops”). They’d plan and organize all of the various things that needed to happen around Optimist. Then they’d assign out the work to whoever was able to help. We had a whole system for tying this into the our profit share and even gave people “Partner” status based on their contributions to ops. It worked—kinda. One big downfall is that all of the tasks and projects were ad hoc. People would pick up jobs, but they didn’t have much context or expertise to apply. So the output often varied. Since we were trying to maintain a flat structure, there was minimal oversight or management of the work. In other words, we didn’t always get the best results. But, more importantly, we still didn’t close all of the gaps entirely. Because everything was an ad-hoc list of tasks and projects, we never really had the “big picture” view of everything that needed to be done across the business. This also meant we rarely had clarity on what was important, what was trivial, and what was critical. We need a better system. Stop Reinventing the Wheel (And Create a Damn Org Chart) It’s time to get serious about filling the gaps in our business. It can’t be a half-fix or an ad hoc set of projects and tasks. We need clarity on the roles that need to be filled and then fill them. The first step here is to create an org chart. A real one. Map out all of the jobs that need to be done for Optimist to be successful besides just writers and designers. Roles like: Content director Design director SEO manager Reporting Finance Account management Business development Sales Marketing Project management It feels a bit laughable listing all of these roles. Because most are either empty or have my name attached to them. And that’s the problem. I can’t do everything. And all of the empty roles are gaps in our structure—places where people aren’t getting the direction, feedback, or guidance they need to do their best work. Or where things just aren’t being done consistently. Content director, for example, should be responsible for steering the output of our content strategists, writers, and editors. They’re not micromanaging every deliverable. But they give feedback, set overall policy, and help our team identify opportunities to get better. Right now we don’t have anyone in that role. Which means it’s my job—when I have time. Looking at the org chart (a real org chart that I actually built to help with this), it’s plain as day how many roles look like this. Even if we aren’t going to implement a traditional agency structure and a strict hierarchy, we still need to address these gaps. And the only way for that to happen is face the reality and then create a plan to close the gaps. Now that we have a list of theoretical roles, we need to clearly define the responsibilities and boundaries of those roles to make sure they cover everything that actually needs to happen. Then we can begin the process of delegating, assigning, hiring, and otherwise addressing each one. So that’s what I need to do. To be done: Create job descriptions for all of the roles we need to fill Hire Biz Dev role Hire Account Lead role(s) Hire Head of Content Playing Offense As we move into Q1 of 2025 and I reflect on the tumultuous few years we’ve had, one thought keeps running through my head. We need to play offense. Most of the last 1-2 years was reacting to changes that were happening around us. Trying to make sense and chart a new path forward. Reeling. But what I really want—as a person and as an entrepreneur—is to be proactive. I want to think and plan ahead. Figure out where we want to go before we’re forced to change course by something that’s out of our control. So my overarching focus for Q1 is playing offense. Thinking longer term. Getting ahead of the daily deluge and creating space to be more proactive, innovative, and forward thinking. To do: Pilot new content formats Audit and update our own content strategy Improve feedback workflows Build out long-term roadmap for 1-2 years for Optimist Final Note on Follow-Through and Cadence In my reflection this year, one of the things I’ve realized is how helpful these posts are for me. I process by writing. So I actually end up making a lot of decisions and seeing things more clearly each time I sit down to reflect and write my yearly recap. It also gives me a space to hold myself accountable for the things I said I would do. So, I’m doing two things a bit differently from here on out. First: I’m identifying clear action items that I’m holding myself accountable for getting done in the next 3 months (listed in the above sections). In each future update, I’ll do an accounting of what I got done and what wasn’t finished (and why). Second: I’m going to start writing shorter quarterly updates. This will gives me more chances each year to reflect, process, and make decisions. Plus it gives me a shorter feedback loop for the action items that I identified above. (See—playing offense.) — Okay friends, enemies, and frenemies. This is my first update for 2025. Glad to share with y’all. And thanks to everyone who’s read, commented, reached out, and shared their own experiences over the years. We are all the accumulation of our connections and our experiences. As always, I will pop in to respond to comments and answer questions. Feel free to share your thoughts, questions, and general disdain down below. Cheers, Tyler

I run an AI automation agency (AAA). My honest overview and review of this new business model
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

I run an AI automation agency (AAA). My honest overview and review of this new business model

I started an AI tools directory in February, and then branched off that to start an AI automation agency (AAA) in June. So far I've come across a lot of unsustainable "ideas" to make money with AI, but at the same time a few diamonds in the rough that aren't fully tapped into yet- especially the AAA model. Thought I'd share this post to shine light into this new business model and share some ways you could potentially start your own agency, or at the very least know who you are dealing with and how to pick and choose when you (inevitably) get bombarded with cold emails from them down the line. Foreword Running an AAA does NOT involve using AI tools directly to generate and sell content directly. That ship has sailed, and unless you are happy with $5 from Fiverr every month or so, it is not a real business model. Cry me a river but generating generic art with AI and slapping it onto a T-shirt to sell on Etsy won't make you a dime. At the same time, the AAA model will NOT require you to have a deep theoretical knowledge of AI, or any academic degree, as we are more so dealing with the practical applications of generative AI and how we can implement these into different workflows and tech-stacks, rather than building AI models from the ground up. Regardless of all that, common sense and a willingness to learn will help (a shit ton), as with anything. Keep in mind - this WILL involve work and motivation as well. The mindset that AI somehow means everything can be done for you on autopilot is not the right way to approach things. The common theme of businesses I've seen who have successfully implemented AI into their operations is the willingess to work with AI in a way that augments their existing operations, rather than flat out replace a worker or team. And this is exactly the train of thought you need when working with AI as a business model. However, as the field is relatively unsaturated and hype surrounding AI is still fresh for enterprises, right now is the prime time to start something new if generative AI interests you at all. With that being said, I'll be going over three of the most successful AI-adjacent businesses I've seen over this past year, in addition to some tips and resources to point you in the right direction. so.. WTF is an AI Automation Agency? The AI automation agency (or as some YouTubers have coined it, the AAA model) at its core involves creating custom AI solutions for businesses. I have over 1500 AI tools listed in my directory, however the feedback I've received from some enterprise users is that ready-made SaaS tools are too generic to meet their specific needs. Combine this with the fact virtually no smaller companies have the time or skills required to develop custom solutions right off the bat, and you have yourself real demand. I would say in practice, the AAA model is quite similar to Wordpress and even web dev agencies, with the major difference being all solutions you develop will incorporate key aspects of AI AND automation. Which brings me to my second point- JUST AI IS NOT ENOUGH. Rather than reducing the amount of time required to complete certain tasks, I've seen many AI agencies make the mistake of recommending and (trying to) sell solutions that more likely than not increase the workload of their clients. For example, if you were to make an internal tool that has AI answer questions based on their knowledge base, but this knowledge base has to be updated manually, this is creating unnecessary work. As such I think one of the key components of building successful AI solutions is incorporating the new (Generative AI/LLMs) with the old (programmtic automation- think Zapier, APIs, etc.). Finally, for this business model to be successful, ideally you should target a niche in which you have already worked and understand pain points and needs. Not only does this make it much easier to get calls booked with prospects, the solutions you build will have much greater value to your clients (meaning you get paid more). A mistake I've seen many AAA operators make (and I blame this on the "Get Rich Quick" YouTubers) is focusing too much on a specific productized service, rather than really understanding the needs of businesses. The former is much done via a SaaS model, but when going the agency route the only thing that makes sense is building custom solutions. This is why I always take a consultant-first approach. You can only build once you understand what they actually need and how certain solutions may impact their operations, workflows, and bottom-line. Basics of How to Get Started Pick a niche. As I mentioned previously, preferably one that you've worked in before. Niches I know of that are actively being bombarded with cold emails include real estate, e-commerce, auto-dealerships, lawyers, and medical offices. There is a reason for this, but I will tell you straight up this business model works well if you target any white-collar service business (internal tools approach) or high volume businesses (customer facing tools approach). Setup your toolbox. If you wanted to start a pressure washing business, you would need a pressure-washer. This is no different. For those without programming knowledge, I've seen two common ways AAA get setup to build- one is having a network of on-call web developers, whether its personal contacts or simply going to Upwork or any talent sourcing agency. The second is having an arsenal of no-code tools. I'll get to this more in a second, but this works beecause at its core, when we are dealing with the practical applications of AI, the code is quite simple, simply put. Start cold sales. Unless you have a network already, this is not a step you can skip. You've already picked a niche, so all you have to do is find the right message. Keep cold emails short, sweet, but enticing- and it will help a lot if you did step 1 correctly and intimately understand who your audience is. I'll be touching base later about how you can leverage AI yourself to help you with outreach and closing. The beauty of gen AI and the AAA model You don't need to be a seasoned web developer to make this business model work. The large majority of solutions that SME clients want is best done using an API for an LLM for the actual AI aspect. The value we create with the solutions we build comes with the conceptual framework and design that not only does what they need it to but integrates smoothly with their existing tech-stack and workflow. The actual implementation is quite straightforward once you understand the high level design and know which tools you are going to use. To give you a sense, even if you plan to build out these apps yourself (say in Python) the large majority of the nitty gritty technical work has already been done for you, especially if you leverage Python libraries and packages that offer high level abstraction for LLM-related functions. For instance, calling GPT can be as little as a single line of code. (And there are no-code tools where these functions are simply an icon on a GUI). Aside from understanding the capabilities and limitations of these tools and frameworks, the only thing that matters is being able to put them in a way that makes sense for what you want to build. Which is why outsourcing and no-code tools both work in our case. Okay... but how TF am I suppposed to actually build out these solutions? Now the fun part. I highly recommend getting familiar with Langchain and LlamaIndex. Both are Python libraires that help a lot with the high-level LLM abstraction I mentioned previously. The two most important aspects include being able to integrate internal data sources/knowledge bases with LLMs, and have LLMs perform autonomous actions. The two most common methods respectively are RAG and output parsing. RAG (retrieval augmented Generation) If you've ever seen a tool that seemingly "trains" GPT on your own data, and wonder how it all works- well I have an answer from you. At a high level, the user query is first being fed to what's called a vector database to run vector search. Vector search basically lets you do semantic search where you are searching data based on meaning. The vector databases then retrieves the most relevant sections of text as it relates to the user query, and this text gets APPENDED to your GPT prompt to provide extra context to the AI. Further, with prompt engineering, you can limit GPT to only generate an answer if it can be found within this extra context, greatly limiting the chance of hallucination (this is where AI makes random shit up). Aside from vector databases, we can also implement RAG with other data sources and retrieval methods, for example SQL databses (via parsing the outputs of LLM's- more on this later). Autonomous Agents via Output Parsing A common need of clients has been having AI actually perform tasks, rather than simply spitting out text. For example, with autonomous agents, we can have an e-commerce chatbot do the work of a basic customer service rep (i.e. look into orders, refunds, shipping). At a high level, what's going on is that the response of the LLM is being used programmtically to determine which API to call. Keeping on with the e-commerce example, if I wanted a chatbot to check shipping status, I could have a LLM response within my app (not shown to the user) with a prompt that outputs a random hash or string, and programmatically I can determine which API call to make based on this hash/string. And using the same fundamental concept as with RAG, I can append the the API response to a final prompt that would spit out the answer for the user. How No Code Tools Can Fit In (With some example solutions you can build) With that being said, you don't necessarily need to do all of the above by coding yourself, with Python libraries or otherwise. However, I will say that having that high level overview will help IMMENSELY when it comes to using no-code tools to do the actual work for you. Regardless, here are a few common solutions you might build for clients as well as some no-code tools you can use to build them out. Ex. Solution 1: AI Chatbots for SMEs (Small and Medium Enterprises) This involves creating chatbots that handle user queries, lead gen, and so forth with AI, and will use the principles of RAG at heart. After getting the required data from your client (i.e. product catalogues, previous support tickets, FAQ, internal documentation), you upload this into your knowledge base and write a prompt that makes sense for your use case. One no-code tool that does this well is MyAskAI. The beauty of it especially for building external chatbots is the ability to quickly ingest entire websites into your knowledge base via a sitemap, and bulk uploading files. Essentially, they've covered the entire grunt work required to do this manually. Finally, you can create a inline or chat widget on your client's website with a few lines of HTML, or altneratively integrate it with a Slack/Teams chatbot (if you are going for an internal Q&A chatbot approach). Other tools you could use include Botpress and Voiceflow, however these are less for RAG and more for building out complete chatbot flows that may or may not incorporate LLMs. Both apps are essentially GUIs that eliminate the pain and tears and trying to implement complex flows manually, and both natively incoporate AI intents and a knowledge base feature. Ex. Solution 2: Internal Apps Similar to the first example, except we go beyond making just chatbots but tools such as report generation and really any sort of internal tool or automations that may incorporate LLM's. For instance, you can have a tool that automatically generates replies to inbound emails based on your client's knowledge base. Or an automation that does the same thing but for replies to Instagram comments. Another example could be a tool that generates a description and screeenshot based on a URL (useful for directory sites, made one for my own :P). Getting into more advanced implementations of LLMs, we can have tools that can generate entire drafts of reports (think 80+ pages), based not only on data from a knowledge base but also the writing style, format, and author voice of previous reports. One good tool to create content generation panels for your clients would be MindStudio. You can train LLM's via prompt engineering in a structured way with your own data to essentially fine tune them for whatever text you need it to generate. Furthermore, it has a GUI where you can dictate the entire AI flow. You can also upload data sources via multiple formats, including PDF, CSV, and Docx. For automations that require interactions between multiple apps, I recommend the OG zapier/make.com if you want a no-code solution. For instance, for the automatic email reply generator, I can have a trigger such that when an email is received, a custom AI reply is generated by MyAskAI, and finally a draft is created in my email client. Or, for an automation where I can create a social media posts on multiple platforms based on a RSS feed (news feed), I can implement this directly in Zapier with their native GPT action (see screenshot) As for more complex LLM flows that may require multiple layers of LLMs, data sources, and APIs working together to generate a single response i.e. a long form 100 page report, I would recommend tools such as Stack AI or Flowise (open-source alternative) to build these solutions out. Essentially, you get most of the functions and features of Python packages such as Langchain and LlamaIndex in a GUI. See screenshot for an example of a flow How the hell are you supposed to find clients? With all that being said, none of this matters if you can't find anyone to sell to. You will have to do cold sales, one way or the other, especially if you are brand new to the game. And what better way to sell your AI services than with AI itself? If we want to integrate AI into the cold outreach process, first we must identify what it's good at doing, and that's obviously writing a bunch of text, in a short amount of time. Similar to the solutions that an AAA can build for its clients, we can take advantage of the same principles in our own sales processes. How to do outreach Once you've identified your niche and their pain points/opportunities for automation, you want to craft a compelling message in which you can send via cold email and cold calls to get prospects booked on demos/consultations. I won't get into too much detail in terms of exactly how to write emails or calling scripts, as there are millions of resources to help with this, but I will tell you a few key points you want to keep in mind when doing outreach for your AAA. First, you want to keep in mind that many businesses are still hesitant about AI and may not understand what it really is or how it can benefit their operations. However, we can take advantage of how mass media has been reporting on AI this past year- at the very least people are AWARE that sooner or later they may have to implement AI into their businesses to stay competitive. We want to frame our message in a way that introduces generative AI as a technology that can have a direct, tangible, and positive impact on their business. Although it may be hard to quantify, I like to include estimates of man-hours saved or costs saved at least in my final proposals to prospects. Times are TOUGH right now, and money is expensive, so you need to have a compelling reason for businesses to get on board. Once you've gotten your messaging down, you will want to create a list of prospects to contact. Tools you can use to find prospects include Apollo.io, reply.io, zoominfo (expensive af), and Linkedin Sales Navigator. What specific job titles, etc. to target will depend on your niche but for smaller companies this will tend to be the owner. For white collar niches, i.e. law, the professional that will be directly benefiting from the tool (i.e. partners) may be better to contact. And for larger organizations you may want to target business improvement and digital transformation leads/directors- these are the people directly in charge of projects like what you may be proposing. Okay- so you have your message, and your list, and now all it comes down to is getting the good word out. I won't be going into the details of how to send these out, a quick Google search will give you hundreds of resources for cold outreach methods. However, personalization is key and beyond simple dynamic variables you want to make sure you can either personalize your email campaigns directly with AI (SmartWriter.ai is an example of a tool that can do this), or at the very least have the ability to import email messages programmatically. Alternatively, ask ChatGPT to make you a Python Script that can take in a list of emails, scrape info based on their linkedin URL or website, and all pass this onto a GPT prompt that specifies your messaging to generate an email. From there, send away. How tf do I close? Once you've got some prospects booked in on your meetings, you will need to close deals with them to turn them into clients. Call #1: Consultation Tying back to when I mentioned you want to take a consultant-first appraoch, you will want to listen closely to their goals and needs and understand their pain points. This would be the first call, and typically I would provide a high level overview of different solutions we could build to tacke these. It really helps to have a presentation available, so you can graphically demonstrate key points and key technologies. I like to use Plus AI for this, it's basically a Google Slides add-on that can generate slide decks for you. I copy and paste my default company messaging, add some key points for the presentation, and it comes out with pretty decent slides. Call #2: Demo The second call would involve a demo of one of these solutions, and typically I'll quickly prototype it with boilerplate code I already have, otherwise I'll cook something up in a no-code tool. If you have a niche where one type of solution is commonly demanded, it helps to have a general demo set up to be able to handle a larger volume of calls, so you aren't burning yourself out. I'll also elaborate on how the final product would look like in comparison to the demo. Call #3 and Beyond: Once the initial consultation and demo is complete, you will want to alleviate any remaining concerns from your prospects and work with them to reach a final work proposal. It's crucial you lay out exactly what you will be building (in writing) and ensure the prospect understands this. Furthermore, be clear and transparent with timelines and communication methods for the project. In terms of pricing, you want to take this from a value-based approach. The same solution may be worth a lot more to client A than client B. Furthermore, you can create "add-ons" such as monthly maintenance/upgrade packages, training sessions for employeees, and so forth, separate from the initial setup fee you would charge. How you can incorporate AI into marketing your businesses Beyond cold sales, I highly recommend creating a funnel to capture warm leads. For instance, I do this currently with my AI tools directory, which links directly to my AI agency and has consistent branding throughout. Warm leads are much more likely to close (and honestly, much nicer to deal with). However, even without an AI-related website, at the very least you will want to create a presence on social media and the web in general. As with any agency, you will want basic a professional presence. A professional virtual address helps, in addition to a Google Business Profile (GBP) and TrustPilot. a GBP (especially for local SEO) and Trustpilot page also helps improve the looks of your search results immensely. For GBP, I recommend using ProfilePro, which is a chrome extension you can use to automate SEO work for your GBP. Aside from SEO optimzied business descriptions based on your business, it can handle Q/A answers, responses, updates, and service descriptions based on local keywords. Privacy and Legal Concerns of the AAA Model Aside from typical concerns for agencies relating to service contracts, there are a few issues (especially when using no-code tools) that will need to be addressed to run a successful AAA. Most of these surround privacy concerns when working with proprietary data. In your terms with your client, you will want to clearly define hosting providers and any third party tools you will be using to build their solution, and a DPA with these third parties listed as subprocessors if necessary. In addition, you will want to implement best practices like redacting private information from data being used for building solutions. In terms of addressing concerns directly from clients, it helps if you host your solutions on their own servers (not possible with AI tools), and address the fact only ChatGPT queries in the web app, not OpenAI API calls, will be used to train OpenAI's models (as reported by mainstream media). The key here is to be open and transparent with your clients about ALL the tools you are using, where there data will be going, and make sure to get this all in writing. have fun, and keep an open mind Before I finish this post, I just want to reiterate the fact that this is NOT an easy way to make money. Running an AI agency will require hours and hours of dedication and work, and constantly rearranging your schedule to meet prospect and client needs. However, if you are looking for a new business to run, and have a knack for understanding business operations and are genuinely interested in the pracitcal applications of generative AI, then I say go for it. The time is ticking before AAA becomes the new dropshipping or SMMA, and I've a firm believer that those who set foot first and establish themselves in this field will come out top. And remember, while 100 thousand people may read this post, only 2 may actually take initiative and start.

26 Ways to Make Money as a Startup Founder (for coders & noncoders)
reddit
LLM Vibe Score0
Human Vibe Score1
johnrushxThis week

26 Ways to Make Money as a Startup Founder (for coders & noncoders)

I've launched 24 projects (here is the proof johnrush.me). None of my projects is making millions a month, but many of them make over $1k a month, some do over $10k, and few do even more. I'd not recommend anyone to start by trying to build a unicorn. Better start simple. Aim for $2-4k a month first. Once you get there, either scale it or start a new project with large TAM. From my own experience, the 26 Ways to Make Money as a Startup Founder: One-Feature SaaS. Extract a feature from a popular tool and build a micro SaaS around it. Idea: A SaaS that only offers automated email follow-ups. Launchpads. Develop a launch platform for a specific industry. Idea: A launchpad for growth tools. SEO Tools. Create a tool that focuses on a single aspect of SEO. Idea: A tool that generates alt texts for images. Productized Services. Offer standardized services that are repeatable. Idea: design, coding or social media management. Marketplace Platforms. Create a platform that connects buyers and sellers, earning transaction fees. Idea: An online marketplace for domains. Membership Sites. A subscription-based site with exclusive content. Idea: A founder 0-to-1 site. White Labeling. A product that other businesses can rebrand as their own. Idea: A white-labeled website builder. Selling Data. Provide anonymized data insights to companies. Idea: Selling user behavior data. Affiliate Marketing. Promote products/services and earn commissions on sales. Idea: Recommending hosting services on a tech blog. Selling Leads. Generate and sell business leads. Idea: Selling leads who raised a fresh seed round. Niche Social Networks. Create a paid community around a specific interest. Idea: A network for SEO experts. Sell Domains. Buy and sell domain names for profit. Virtual Products. Sell digital products like templates or graphics. Idea: Website themes for nextjs or boilerplates. On-Demand Services. Build a platform for gigs like delivery or tutoring. Idea: An app for freelance tutors. Niche Job Boards. Start a job board focused on a specific industry. Idea: A job board for remote tech jobs. Crowdsourced Content. Create a user-generated content platform and monetize through ads. Idea: Site to share startup hacks. Buy and Flip Businesses. Purchase underperforming businesses, improve them, and sell for profit. Idea: Acquiring a low-traffic blog, optimizing it, and selling. AI-Powered agents. Develop AI tools that solve specific business problems. Idea: An AI tool that automates customer support. Microservices. Offer small, specialized tools, sdks or APIs. Idea: An api for currency conversion. Influencer Platforms. Create a platform connecting influencers with brands. Idea: Connect AI influencers with AI founders. Niche Directories. Build a paid directory for a specific industry. Idea: A directory of developers who can train models. E-Learning Platforms. Build a platform for educators to sell courses. Idea: A site where AI experts sell AI courses. Virtual assistants. Hire them and sell on subscription. No-Code Tools. Create tools that allow non-technical users to build things. Idea: A no-code website builder for bakeries. Labor arbitrage. Idea: Connect support agents from Portugal with US clients and charge commission.

B2B Marketers: What’s Your #1 Tip for Selling SaaS to Other Businesses? (Building a Tool for Shopify SEO)
reddit
LLM Vibe Score0
Human Vibe Score1
iammanmanthemanThis week

B2B Marketers: What’s Your #1 Tip for Selling SaaS to Other Businesses? (Building a Tool for Shopify SEO)

Hi everyone! I’m part of a team building an AI-powered SEO tool specifically for Shopify stores (think automated technical fixes, predictive keyword optimization, etc.). We’re in the early stages and want to learn from seasoned B2B marketers: We’re struggling with: How to position a technical SaaS product to non-technical Shopify merchants. Cutting through the noise in a crowded SEO tools market. Building trust quickly with time-strapped business owners. Questions for B2B Marketing Pros: What’s the biggest mistake you made when marketing a SaaS product to businesses? What’s one underrated tactic that’s worked wonders for B2B lead gen? How do you prove ROI to skeptical buyers? (Especially for something abstract like SEO.) What’s your go-to channel for cold outreach that doesn’t feel spammy? What’s a hidden psychological trigger that works in B2B sales? What’s the best way to leverage case studies/testimonials when you’re just starting out? What’s one thing most founders waste money on in B2B marketing? For Those Who’ve Sold to Shopify Merchants: What’s their biggest pain point when evaluating tools? What type of content (webinars, blogs, demos) convinces them to buy? The Deal: We’ll compile all advice into a guide and credit contributors. If you're willing to have a virtual coffee chat, please reach out to us, we are always willing to listen to your wisdom!

Unmasking Fake Testimonials on a YC backed company
reddit
LLM Vibe Score0
Human Vibe Score0.75
Far-Amphibian3043This week

Unmasking Fake Testimonials on a YC backed company

As developers, marketeers and builders, we often rely on trusted platforms to guide us in finding tools that meet our unique needs. Recently, I stumbled upon Overlap, a site marketed as a haven for collaboration tools. Its sleek interface and glowing testimonials initially convinced me I had found a gem. But as I dug deeper, I uncovered a jaw-dropping reality: their testimonials featured stock images, all of which were easily identified through a quick reverse image search. Even more shocking was the realization that Overlap is a Y Combinator-backed company—an organization renowned for nurturing some of the most innovative startups in the world. With significant funding at their disposal, the decision to cut corners with fake testimonials felt like a slap in the face to their user base. They could easily afford a robust testimonial platform, yet chose a path that undermined their credibility. As developers, marketeers and builders, we often rely on trusted platforms to guide us in finding tools that meet our unique needs. Recently, I stumbled upon Overlap, a site marketed as a haven for video AI tools. Its sleek interface and glowing testimonials initially convinced me I had found a gem. But as I dug deeper, I uncovered a jaw-dropping reality: their testimonials featured stock images, all of which were easily identified through a quick reverse image search. Even more shocking was the realization that Overlap is a Y Combinator-backed company—an organization renowned for nurturing some of the most innovative startups in the world. With significant funding at their disposal, the decision to cut corners with fake testimonials felt like a slap in the face to their user base. They could easily afford a robust testimonial platform, yet chose a path that undermined their credibility. A screenshot of Overlap's landing page https://preview.redd.it/zosmdl0v01ce1.png?width=1000&format=png&auto=webp&s=83ced4af92ca284486281f00b020f1f0114b4fcd This discovery was nothing short of a wake-up call. For a developer-focused website—an audience that prizes authenticity and technical precision above all else—faking testimonials with stock photos isn’t just misleading, it’s a catastrophic betrayal of trust. It left me questioning the integrity of their entire operation and serves as a stark reminder for businesses everywhere: your audience notices when you’re not authentic, and they won’t forgive it easily. Position of Fake Testimonials One of the stock images https://preview.redd.it/a7ugasrw01ce1.png?width=341&format=png&auto=webp&s=5261df741f1198a92e537f1e61640e7d6ec60a7f Lessons for Startup Founders and Developers This experience offers several critical lessons for startup founders and developers alike: Authenticity is Non-Negotiable: In a competitive market, trust and transparency can make or break your brand. Fake testimonials might provide a short-term boost, but the long-term damage to credibility far outweighs any temporary gains. Invest in Genuine Solutions: If you have the resources, like a Y Combinator-backed company, prioritize tools and practices that enhance authenticity. Platforms like RapidFeedback allow businesses to dynamically update reviews and manage feedback efficiently. Leverage Real User Feedback: Authentic testimonials not only build trust but also provide actionable insights into your product’s strengths and weaknesses. This feedback loop can be invaluable for refining and growing your business. Understand Your Audience: Developers value precision, integrity, and honesty. Catering to this audience requires a commitment to these principles in every aspect of your business. Let’s ensure that the tools we build and the businesses we run prioritize authenticity. In the long run, a commitment to transparency and user trust will always yield greater rewards than any shortcut could provide. Why Fake Testimonials Are a Problem Fake testimonials damage your brand in more ways than one: Loss of Credibility: Developers are a discerning audience. Trust is everything, and losing it can be catastrophic for your reputation. Hurt User Experience: Knowing a platform misrepresents itself makes users skeptical about its features and promises. Missed Opportunities: Genuine feedback can provide valuable insights for growth and improvement, which fake testimonials completely overlook. A Smarter Way: Authentic Testimonials with RapidFeedback This experience reminded me of why tools like RapidFeedback are invaluable. RapidFeedback helps businesses maintain authenticity by dynamically updating reviews and images in real time. Here’s why it stands out: Real-Time Updates: Reviews are fetched and displayed dynamically, ensuring they’re always up-to-date. Dashboard Management: Businesses can monitor and manage good vs. bad reviews from a centralized dashboard, enabling them to address concerns promptly. Authenticity Guaranteed: Dynamic updates ensure that testimonials reflect real users and their experiences, which builds trust and credibility. Lessons for Developers and Businesses If there’s one takeaway from my Overlap experience, it’s this: authenticity isn’t optional. Whether you’re building tools for developers or selling consumer products, your audience values transparency. Using tools like RapidFeedback ensures your business maintains trust while gaining actionable insights to grow. Let’s commit to prioritizing honesty in our work. Because in the end, authentic relationships with users are what truly drive success.

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security

Uzair Javaid, a Ph.D. with a passion for data privacy, co-founded Betterdata to tackle one of AI's most pressing challenges: protecting privacy while enabling innovation. Recently, Betterdata secured a lucrative contract with the US Department of Homeland Security, 1 of only 4 companies worldwide to do so and the only one in Asia. Here's how he did it: The Story So what's your story? I grew up in Peshawar, Pakistan, excelling in coding despite studying electrical engineering. Inspired by my professors, I set my sights on studying abroad and eventually earned a Ph.D. scholarship at NUS Singapore, specializing in data security and privacy. During my research, I ethically hacked Ethereum and published 15 papers—three times the requirement. While wrapping up my Ph.D., I explored startup ideas and joined Entrepreneur First, where I met Kevin Yee. With his expertise in generative models and mine in privacy, we founded Betterdata. Now, nearly three years in, we’ve secured a major contract with the U.S. Department of Homeland Security—one of only four companies globally and the only one from Asia. The Startup In a nutshell, what does your startup do? Betterdata is a startup that uses AI and synthetic data generation to address two major challenges: data privacy and the scarcity of high-quality data for training AI models. By leveraging generative models and privacy-enhancing technologies, Betterdata enables businesses, such as banks, to use customer data without breaching privacy regulations. The platform trains AI on real data, learns its patterns, and generates synthetic data that mimics the real thing without containing any personal or sensitive information. This allows companies to innovate and develop AI solutions safely and ethically, all while tackling the growing need for diverse, high-quality data in AI development. How did you conduct ideation and validation for your startup? The initial idea for Betterdata came from personal experience. During my Ph.D., I ethically hacked Ethereum’s blockchain, exposing flaws in encryption-based data sharing. This led me to explore AI-driven deep synthesis technology—similar to deepfakes but for structured data privacy. With GDPR impacting 28M+ businesses, I saw a massive opportunity to help enterprises securely share data while staying compliant. To validate the idea, I spoke to 50 potential customers—a number that strikes the right balance. Some say 100, but that’s impractical for early-stage founders. At 50, patterns emerge: if 3 out of 10 mention the same problem, and this repeats across 50, you have 10–15 strong signals, making it a solid foundation for an MVP. Instead of outbound sales, which I dislike, we used three key methods: Account-Based Marketing (ABM)—targeting technically savvy users with solutions for niche problems, like scaling synthetic data for banks. Targeted Content Marketing—regular customer conversations shaped our thought leadership and outreach. Raising Awareness Through Partnerships—collaborating with NUS, Singapore’s PDPC, and Plug and Play to build credibility and educate the market. These strategies attracted serious customers willing to pay, guiding Betterdata’s product development and market fit. How did you approach the initial building and ongoing product development? In the early stages, we built synthetic data generation algorithms and a basic UI for proof-of-concept, using open-source datasets to engage with banks. We quickly learned that banks wouldn't share actual customer data due to privacy concerns, so we had to conduct on-site installations and gather feedback to refine our MVP. Through continuous consultation with customers, we discovered real enterprise data posed challenges, such as missing values, which led us to adapt our prototype accordingly. This iterative approach of listening to customer feedback and observing their usage allowed us to improve our product, enhance UX, and address unmet needs while building trust and loyalty. Working closely with our customers also gives us a data advantage. Our solution’s effectiveness depends on customer data, which we can't fully access, but bridging this knowledge gap gives us a competitive edge. The more customers we test on, the more our algorithms adapt to diverse use cases, making it harder for competitors to replicate our insights. My approach to iteration is simple: focus solely on customer feedback and ignore external noise like trends or advice. The key question for the team is: which customer is asking for this feature or solution? As long as there's a clear answer, we move forward. External influences, such as AI hype, often bring more confusion than clarity. True long-term success comes from solving real customer problems, not chasing trends. Customers may not always know exactly what they want, but they understand their problems. Our job is to identify these problems and solve them in innovative ways. While customers may suggest specific features, we stay focused on solving the core issue rather than just fulfilling their exact requests. The idea aligns with the quote often attributed to Henry Ford: "If I asked people what they wanted, they would have said faster horses." The key is understanding their problems, not just taking requests at face value. How do you assess product-market fit? To assess product-market fit, we track two key metrics: Customers' Willingness to Pay: We measure both the quantity and quality of meetings with potential customers. A high number of meetings with key decision-makers signals genuine interest. At Betterdata, we focused on getting meetings with people in banks and large enterprises to gauge our product's resonance with the target market. How Much Customers Are Willing to Pay: We monitor the price customers are willing to pay, especially in the early stages. For us, large enterprises, like banks, were willing to pay a premium for our synthetic data platform due to the growing need for privacy tech. This feedback guided our product refinement and scaling strategy. By focusing on these metrics, we refined our product and positioned it for scaling. What is your business model? We employ a structured, phase-driven approach for out business model, as a B2B startup. I initially struggled with focusing on the core value proposition in sales, often becoming overly educational. Eventually, we developed a product roadmap with models that allowed us to match customer needs to specific offerings and justify our pricing. Our pricing structure includes project-based pilots and annual contracts for successful deployments. At Betterdata, our customer engagement unfolds across three phases: Phase 1: Trial and Benchmarking \- We start with outreach and use open-source datasets to showcase results, offering customers a trial period to evaluate the solution. Phase 2: Pilot or PoC \- After positive trial results, we conduct a PoC or pilot using the customer’s private data, with the understanding that successful pilots lead to an annual contract. Phase 3: Multi-Year Contracts \- Following a successful pilot, we transition to long-term commercial contracts, focusing on multi-year agreements to ensure stability and ongoing partnerships. How do you do marketing for your brand? We take a non-conventional approach to marketing, focusing on answering one key question: Which customers are willing to pay, and how much? This drives our messaging to show how our solution meets their needs. Our strategy centers around two main components: Building a network of lead magnets \- These are influential figures like senior advisors, thought leaders, and strategic partners. Engaging with institutions like IMDA, SUTD, and investors like Plug and Play helps us gain access to the right people and foster warm introductions, which shorten our sales cycle and ensure we’re reaching the right audience. Thought leadership \- We build our brand through customer traction, technology evidence, and regulatory guidelines. This helps us establish credibility in the market and position ourselves as trusted leaders in our field. This holistic approach has enabled us to navigate diverse market conditions in Asia and grow our B2B relationships. By focusing on these areas, we drive business growth and establish strong trust with stakeholders. What's your advice for fundraising? Here are my key takeaways for other founders when it comes to fundraising: Fundraise When You Don’t Need To We closed our seed round in April 2023, a time when we weren't actively raising. Founders should always be in fundraising mode, even when they're not immediately in need of capital. Don’t wait until you have only a few months of runway left. Keep the pipeline open and build relationships. When the timing is right, execution becomes much easier. For us, our investment came through a combination of referrals and inbound interest. Even our lead investor initially rejected us, but after re-engaging, things eventually fell into place. It’s crucial to stay humble, treat everyone with respect, and maintain those relationships for when the time is right. Be Mindful of How You Present Information When fundraising, how you present information matters a lot. We created a comprehensive, easily digestible investment memo, hosted on Notion, which included everything an investor might need—problem, solution, market, team, risks, opportunities, and data. The goal was for investors to be able to get the full picture within 30 minutes without chasing down extra details. We also focused on making our financial model clear and meaningful, even though a 5-year forecast might be overkill at the seed stage. The key was clarity and conciseness, and making it as easy as possible for investors to understand the opportunity. I learned that brevity and simplicity are often the best ways to make a memorable impact. For the pitch itself, keep it simple and focus on 4 things: problem, solution, team, and market. If you can summarize each of these clearly and concisely, you’ll have a compelling pitch. Later on, you can expand into market segments, traction, and other metrics, but for seed-stage, focus on those four areas, and make sure you’re strong in at least three of them. If you do, you'll have a compelling case. How do you run things day-to-day? i.e what's your operational workflow and team structure? Here's an overview of our team structure and process: Internally: Our team is divided into two main areas: backend (internal team) and frontend (market-facing team). There's no formal hierarchy within the backend team. We all operate as equals, defining our goals based on what needs to be developed, assigning tasks, and meeting weekly to share updates and review progress. The focus is on full ownership of tasks and accountability for getting things done. I also contribute to product development, identifying challenges and clearing obstacles to help the team move forward. Backend Team: We approach tasks based on the scope defined by customers, with no blame or hierarchy. It's like a sports team—sometimes someone excels, and other times they struggle, but we support each other and move forward together. Everyone has the creative freedom to work in the way that suits them best, but we establish regular meetings and check-ins to ensure alignment and progress. Frontend Team: For the market-facing side, we implement a hierarchy because the market expects this structure. If I present myself as "CEO," it signals authority and credibility. This distinction affects how we communicate with the market and how we build our brand. The frontend team is split into four main areas: Business Product (Software Engineering) Machine Learning Engineering R&D The C-suite sits at the top, followed by team leads, and then the executors. We distill market expectations into actionable tasks, ensuring that everyone is clear on their role and responsibilities. Process: We start by receiving market expectations and defining tasks based on them. Tasks are assigned to relevant teams, and execution happens with no communication barriers between team members. This ensures seamless collaboration and focused execution. The main goal is always effectiveness—getting things done efficiently while maintaining flexibility in how individuals approach their work. In both teams, there's an emphasis on accountability, collaboration, and clear communication, but the structure varies according to the nature of the work and external expectations.

10 Side Projects in 10 Years: Lessons from Failures and a $700 Exit
reddit
LLM Vibe Score0
Human Vibe Score1
TheValueProviderThis week

10 Side Projects in 10 Years: Lessons from Failures and a $700 Exit

Hey folks, I'm sharing my journey so far in case it can help others. Entrepreneurship can sometimes be demotivating. In my case, I've always been involved in side projects and what I've realized is that every time you crash a project, the next one makes it a bit further. So this is a long-term game and consistency ends up paying off The $1 Android Game (2015, age 18) What Happened: 500 downloads, 1€ in ad revenue Ugly UI, performance issues Key Lessons: Don’t be afraid of launching. Delaying for “perfection” is often a sign that you fear being ignored. I was trying to perfect every aspect of the game. In reality, I was delaying the launch because I feared no one would download the app. Commit to the project or kill it. At some point, this project was no longer fun (it was just about fixing device responsiveness). Most importantly, I wasn't learning anything new so I moved to smth else. The Forex Bot Regret (2016, age 19) What Happened: Lost months identifying inexistent chart patterns Created a Trading bot that was never profitable Key Lessons: Day trading’s real winners are usually brokers. There are plenty of guys selling a bot or systems that are not making money trading, why would they sell a “money-printing machine” otherwise... Develop an unfair advantage. With these projects, I developed a strong coding foundation that gave me an edge when dealing with non-technical business people. Invest countless hours to create a skills gap between you and others, one that becomes increasingly difficult for them to close (coding, public speaking, networking, etc.) The $700 Instagram Exit (2018, age 21) What Happened: Grew a motivational account to 60k followers Sold it for $700 90% of followers were in low-income countries (hard to monetize) Key Lessons: Follower quality > quantity. I focused on growth and ended up with an audience I couldn’t truly define. If brands don’t see value, you won’t generate revenue. Also, if you do not know who you are creating content for, you'll end up demotivated and stop posting. Great 3rd party product + domain authority = Affiliate marketing works. In this case, I could easily promote an IG growing service because my 50k+ followers conveyed trust. Most importantly, the service I was promoting worked amazingly. The Illegal Amazon Review Marketplace (2020, age 23) What Happened: Sellers were reimbursing buyers for positive reviews Built a WordPress marketplace to facilitate “free products for reviews” Realized it violated Amazon’s terms Key Lessons: Check for “red flags” when doing idea assessment. There will always be red and orange flags. It’s about learning to differentiate between them (e.g. illegality, 100% dependence on a platform, etc.) If there’s competition, it’s good, if they are making money it’s even better. I was thrilled when I saw no competition for my “unique idea”. Later, I discovered the obvious reason. Copying a “Proven” Business Model (2020, age 23) What Happened: Tried recreating an Instagram “comment for comment” growth tool Instagram changed the algorithm and killed the growth strategy that the product used. Key Lessons: Do not build a business that depends 100% on another business, it is too risky. Mr. Musk can increase Twitter on API pricing to $42,000 monthly without notice and Tik Tok can be banned in the US. Due to the IG algorithm change, we had built a product that was not useful, and worse, now we had no idea how to grow an IG account. Consider future project synergies before selling. I regret having sold the 60k follower IG account since it could have saved me a lot of time when convincing users to try the service. NFT Marathon Medals (2021, age 24) What Happened: Created NFT race medals Sold 20 for 5€ each, but spent 95% of meetings explaining “what is an NFT?” Key Lessons: Market timing is crucial. As with every new technology, it is only useful as long as society is ready to adopt it. No matter how promising the tech is in the eyes of SV, society will end up dictating its success (blockchain, AI, etc). In this case, the runner community was not ready to adopt blockchain (it is not even prepared today). Race organizers did not know what they were selling, and runners did not know what they were buying. The 30-day rule in Fanatical Prospecting. Do not stop prospecting. I did prospecting and closed deals 3 months after the outbound efforts. Then I was busy executing the projects and had no clients once the projects were finished. AI Portal & Co-Founder Misalignment (2023, age 26) What Happened: Built a portal for SMEs to find AI use cases Co-founders disagreed on vision and execution Platform still gets \~1 new user/day Key Lessons: Define roles and equity clearly. Our biggest strength ended up killing us. Both founders had strong strategic skills and we were constantly arguing about decisions. NextJS + Vercel + Supabase: Great stack to create a SaaS MVP. (but do not use AI with frameworks unless you know how they work conceptually) SEO is king. One of our users creates a use case on “Changing Song Lyrics with AI.” Not being our target use case, it brings 90% of our traffic. Building an AI Tool & Getting Ghosted (2024, age 27) What Happened: SEO agency wanted to automate rewriting product descriptions Built it in 3 weeks, but the client vanished Key Lessons: Validate manually first. Don’t code a full-blown solution for a problem you haven’t tested in real-world workflows. I kept rewriting code only to throw it away. Jumping straight into building a solution ended up costing more time than it saved. Use templates, no-code, and open-source for prototyping. In my case, using a Next.js template saved me about four weeks of development only to hit the same dead end, but much faster. Fall in love with your ICP or walk away. I realized I didn’t enjoy working with SEO agencies. Looking back, I should have been honest with myself and admitted that I wasn’t motivated enough by this type of customer. Ignoring Code Perfection Doubled Traffic (2025, age 28) What Happened: Partnered with an ex-colleague to build an AI agents directory Focused on content & marketing, not endless bug fixes Traffic soared organically Key Lessons: Measure the impact of your actions and double down on what works. We set up an analytics system with PostHog and found wild imbalances (e.g. 1 post about frameworks outperformed 20 promotional posts). You have to start somewhere. For us, the AI agents directory is much more than just a standalone site, it's a strategic project that will allow us to discover new products, gain domain authority, and boost other projects. It builds the path for bigger opportunities. Less coding, more traction. Every day I have to fight against myself not to code “indispensable features”. Surprisingly, the directory keeps gaining consistent traffic despite being far from perfect Quitting My Job & Looking Ahead (2025, age 28) What Happened: Left full-time work to go all-in Plan to build vertical AI agents that handle entire business workflows (support, marketing, sales) Key Lessons: Bet on yourself. The opportunity cost of staying in my full-time job outweighed the benefits. It might be your case too I hope this post helps anyone struggling with their project and inspires those considering quitting their full-time job to take the leap with confidence.

Why the value of writing code and other digital services is going to zero
reddit
LLM Vibe Score0
Human Vibe Score1
BalloonWheelieThis week

Why the value of writing code and other digital services is going to zero

I must preface this with a trigger warning because I make some statements in this post that might be upsetting to some. This post discusses my experience building in the new era of entrepreneurship, which is one where the founder is the center of the universe, and the consultants, overpriced SaaS, and corporate swamp creatures are replaced by single-user custom software, bots, and self-hosted automations. If you work in the legacy economy, I really don't intend to stress you out or say things you are doing are quickly becoming irrelevant, but I must share the reality of how I am operating, because I would like to hear from others who are doing the same, or desire to do the same. I am currently operating with the belief that AI-powered tools are going to make 1-person million dollar businesses much more common. Building anything digital is becoming extremely easy, cheap, and quick to implement. The value of code and digital tools is approaching zero, or at most 5% of what it currently is. Right now, the most powerful AI tools are aimed at developers, so folks who have some technical and business ability basically have nothing holding them back aside from the speed of their brain right now. I happen to be a part of the cohort, and am building like there is no tomorrow, but I don't believe this cohort is actually all that big. The next hurdle to unlock the new era of entrepreneurship is empowering every entrepreneur to build at the same pace that is currently locked behind having technical ability. This cohort is huge (millions, if the number of people in this sub is any indication). This post is aimed at them (you?). If you are part of this cohort, what is holding you back from launching a new product for near-zero cost? What is too complicated, too expensive, too unknown for you to be able to build your new/current business at maximum speed? I look forward to seeing the replies, I hope some insights shared can help the community, and be a catalyst for more tools to enable non-technical founders to launch. I will now share some of how I am testing, launching, and selling as a one-man-show. This will be a little bit technical, but if the output of any layer of my stack is something you want, please comment because maybe someone will build a cheap way of accessing it without needing to manage the code yourself. \#1 BOTS I cannot overstate how much leverage bots have created for me. I run all of my bots locally and interface with with via Telegram. Bots do things like: \- watch social media pages, forums, subreddits, etc related to my customers and notify me of what is going on, and suggest SEO blog posts that could be published to capture traffic related to the topic. with a single message, my bot will generate a blog post, send it to me for review, apply edits i suggest, and then publish it live, all from within telegram \- pay attention to all my key metrics/analytics, and attempt to find insights/corrolations (ex. there is a lot of traffic on this page, blog post, video, etc. here's why, and how we can take advantage of it to drive business goals) \- repurposing content. i have dozens of social media profiles that are 100% run by bots, they are all related to my customer niches and will do things like post news, snippets from my blogs, interact with human creators in the niche, etc. this builds my audience automatically which I can then advertise to/try to convert into paying customers, since they are interested in the things my bot is posting and become followers, it's like automated qualified lead gen 24/7 across every social platform and every niche I care about. you may be thinking by now that this post is made by a bot, but you will have to trust me that this is 100% hand-written by my sleep-deprived brain. let's continue: \#2 replacing every SaaS with a shitty version of it designed for what i need out of it it's absurd that we pay ten's of dollars per seat per month for basic digital functions like chat (slack), CRM (active camppaign, sales force, hubspot, etc), email stuff (mailchip, etc), link sharing (linktree, etc), website builders (wix, squarespace, etc), etc. all of these SaaS tools are overpriced and overbuilt. I believe many of them are going to be caught in the innovators dilemma and will go to 0. I don't use any of these anymore, I build and self-host my own shitty version of each of them that does only what i need out of the tool. for example, my CRM doesn't have a fancy drag and drop email builder and 10000 3rd party plugins, because i dont need any of that shit I just need to segment and communicate with my customers. if i need more features, i can generate them on the fly. \#3 working alone I have worked with cofounders in the past, raised money from investors, hired consultants, burned money and time, suffered sleepless nights from stress caused by other people not delivering, trying to convince others they are wrong, or they are pushing the company off a cliff, waste waste waste. no more of that. In the new age of entrepreneurship, the BUILDER (you and I) are the ones creating the value, and AI empowers us to do it alone. this might seem daunting, but there is no business problem that can't be solved with a detailed discussion sesh with chatgpt, no facts that can't be found with perplexity, and no task that can't be automated with claude. there is no need for anymore swamp creatures. you are the start and the end point, you don't need to rely on anyone else for anything. this may sound ignorant, but this is the conclusion I have come to believe, and it continues to be proven every day my businesses progress with me being the only human involved. This is getting quite long so I'll cut it here. I look forward to hearing about how you are operating in this new era and hopefully getting inspired/learning some new ideas to add to my current stack.

I Watched My Startup Slowly Dying Over Two Years: Mistakes and Lessons Learned
reddit
LLM Vibe Score0
Human Vibe Score0.429
Personal-Expression3This week

I Watched My Startup Slowly Dying Over Two Years: Mistakes and Lessons Learned

If you are tired of reading successful stories, you may want to listen to my almost failure story. Last year in April, I went full-time on my startup. Nearly two years later, I’ve seen my product gradually dying. I want to share some of the key mistakes I made and the lessons I’ve taken from them so you don't have to go through them. Some mistakes were very obvious in hindsight; others, I’m still not sure if they were mistakes or just bad luck. I’d love to hear your thoughts and advice as well. Background I built an English-learning app, with both web and mobile versions. The idea came from recognizing how expensive it is to hire an English tutor in most countries, especially for practicing speaking skills. With the rise of AI, I saw an opportunity in the education space. My target market was Japan, though I later added support for multiple languages and picked up some users from Indonesia and some Latin American countries too. Most of my users came from influencer marketing on Twitter. The MVP for the web version launched in Japan and got great feedback. People were reposting it on Twitter, and growth was at its peak in the first few weeks. After verifying the requirement with the MVP, I decided to focus on the mobile app to boost user retention, but for various reasons, the mobile version didn’t launch until December 2023— 8 months after the web version. Most of this year has been spent iterating on the mobile app, but it didn’t make much of an impact in the end. Key Events and Lessons Learned Here are some takeaways: Find co-founders as committed as you are I started with two co-founders—both were tech people and working Part-Time. After the web version launched, one dropped out due to family issues. Unfortunately, we didn’t set clear rules for equity allocation, so even after leaving, they still retained part of the equity. The other co-founder also effectively dropped out this year, contributing only minor fixes here and there. So If you’re starting a company with co-founders, make sure they’re as committed as you are. Otherwise, you might be better off going solo. I ended up teaching myself programming with AI tools, starting with Flutter and eventually handling both front-end and back-end work using Windsurf. With dev tools getting more advanced, being a solo developer is becoming a more viable option. Also, have crystal-clear rules for equity—especially around what happens if someone leaves. Outsourcing Pitfalls Outsourcing development was one of my biggest mistakes. I initially hired a former colleague from India to build the app. He dragged the project on for two months with endless excuses, and the final output was unusable. Then I hired a company, but they didn’t have enough skilled Flutter developers. The company’s owner scrambled to find people, which led to rushed work and poor-quality code which took a lot of time revising myself. Outsourcing is a minefield. If you must do it, break the project into small tasks, set clear milestones, and review progress frequently. Catching issues early can save you time and money. Otherwise, you’re often better off learning the tools yourself—modern dev tools are surprisingly beginner-friendly. Trust, but Verify I have a bad habit of trusting people too easily. I don’t like spending time double-checking things, so I tend to assume people will do what they say they’ll do. This mindset is dangerous in a startup. For example, if I had set up milestones and regularly verified the progress of my first outsourced project, I would’ve realized something was wrong within two weeks instead of two months. That would’ve saved me a lot of time and frustration. Like what I mentioned above, set up systems to verify their work—milestones, deliverables, etc.—to minimize risk. Avoid red ocean if you are small My team was tiny (or non-existent, depending on how you see it), with no technical edge. Yet, I chose to enter Japan’s English-learning market, which is incredibly competitive. It’s a red ocean, dominated by big players who’ve been in the game for years. Initially, my product’s AI-powered speaking practice and automatic grammar correction stood out, but within months, competitors rolled out similar features. Looking back, I should’ve gone all-in on marketing during the initial hype and focused on rapidly launching the mobile app. But hindsight is 20/20. 'Understanding your user' helps but what if it's not what you want? I thought I was pretty good at collecting user feedback. I added feedback buttons everywhere in the app and made changes based on what users said. But most of these changes were incremental improvements—not the kind of big updates that spark excitement. Also, my primary users were from Japan and Indonesia, but I’m neither Japanese nor Indonesian. That made it hard to connect with users on social media in an authentic way. And in my opinion, AI translations can only go so far—they lack the human touch and cultural nuance that builds trust. But honestly I'm not sure if the thought is correct to assume that they will not get touched if they recognize you are a foreigner...... Many of my Japanese users were working professionals preparing for the TOEIC exam. I didn’t design any features specifically for that; instead, I aimed to build a general-purpose English-learning tool since I dream to expand it to other markets someday. While there’s nothing wrong with this idealistic approach, it didn’t give users enough reasons to pay for the app. Should You Go Full-Time? From what I read, a lot of successful indie developers started part-time, building traction before quitting their jobs. But for me, I jumped straight into full-time mode, which worked for my lifestyle but might’ve hurt my productivity. I value work-life balance and refused to sacrifice everything for the startup. The reason I chose to leave the corp is I want to escape the 996 toxic working environment in China's internet companies. So even during my most stressful periods, I made time to watch TV with my partner and take weekends off. Anyways, if you’re also building something or thinking about starting a business, I hope my story helps. If I have other thoughts later, I will add them too. Appreciate any advice.

101 best SEO tips to help you drive traffic in 2k21
reddit
LLM Vibe Score0
Human Vibe Score0.543
DrJigsawThis week

101 best SEO tips to help you drive traffic in 2k21

Hey guys! I don't have to tell you how SEO can be good for your business - you can drive leads to your SaaS on autopilot, drive traffic to your store/gym/bar/whatever, etc. The thing with SEO, though, is that most SEO tips on the internet are just not that good. Most of the said tips: Are way too simple & basic (“add meta descriptions to your images”*) Are not impactful. Sure, adding that meta tag to an image is important, but that’s not what’s going to drive traffic to your website Don’t talk much about SEO strategy (which is ultimately the most important thing for SEO). Sure, on-page SEO is great, but you sure as hell won't drive much traffic if you can't hire the right writers to scale your content. And to drive serious SEO traffic, you'll need a LOT more than that. Over the past few years, my and my co-founder have helped grow websites to over 200k+ monthly traffic (check out our older Reddit post if you want to learn more about us, our process, and what we do), and we compiled all our most important SEO tips and tricks, as well as case studies, research, and experiments from the web, into this article. Hope you like it ;) If you think we missed something super important, let us know and we'll add it to the list. And btw, we also published this article on our own blog with images, smart filters, and all that good stuff. If you want to check it out, click here. That said, grab some coffee (or beer) & let's dive in - this is going to be a long one. SEO Strategy Tips Tip #1. A Lot of SEO Tips On The Internet Are NOT Necessarily Factual A lot of the SEO content you’ll read on the internet will be based on personal experiences and hearsay. Unfortunately, Google is a bit vague about SEO advice, so you have to rely more on experiments conducted by SEO pros in the community. So, sometimes, a lot of this information is questionable, wrong, or simply based on inaccurate data.  What we’re getting at here is, whenever you hear some new SEO advice, take it with a grain of salt. Google it to double-check other sources, and really understand what this SEO advice is based on (instead of just taking it at face value). Tip #2. SEO Takes Time - Get Used to It Any way you spin it, SEO takes time.  It can take around 6 months to 2 years (depending on the competition in your niche) before you start seeing some serious results.  So, don’t get disappointed if you don’t see any results within 3 months of publishing content. Tip #3. SEO Isn’t The Best Channel for Everyone That said, if you need results for your business tomorrow, you might want to reconsider SEO altogether.  If you just started your business, for example, and are trying to get to break-even ASAP, SEO is a bad idea - you’ll quit before you even start seeing any results.  If that’s the case, focus on other marketing channels that can have faster results like content marketing, PPC, outreach, etc. Tip #4. Use PPC to Validate Keywords Not sure if SEO is right for your business? Do this: set up Google Search ads for the most high-intent keywords in your niche. See how well the traffic converts and then decide if it’s worthwhile to focus on SEO (and rank on these keywords organically). Tip #5. Use GSC to See If SEO Is Working While it takes a while to see SEO results, it IS possible to see if you’re going in the right direction. On a monthly basis, you can use Search Console to check if your articles are indexed by Google and if their average position is improving over time. Tip #6. Publish a TON of Content The more content you publish on your blog, the better. We recommend a minimum of 10,000 words per month and optimally 20,000 - 30,000 (especially if your website is fresh). If an agency offers you the typical “4 500-word articles per month” deal, stay away. No one’s ever gotten results in SEO with short, once-per-week articles. Tip #7. Upgrade Your Writers Got a writer that’s performing well? Hire them as an editor and get them to oversee content operations / edit other writers’ content. Then, upgrade your best editor to Head of Content and get them to manage the entire editor / writer ops. Tip #8. Use Backlink Data to Prioritize Content When doing keyword research, gather the backlink data of the top 3 ranking articles and add it to your sheet. Then, use this data to help you prioritize which keywords to focus on first. We usually prioritize keywords that have lower competition, high traffic, and a medium to high buyer intent. Tip #9. Conduct In-Depth Keyword Research Make your initial keyword research as comprehensive as possible. This will give you a much more realistic view of your niche and allow you to prioritize content the right way. We usually aim for 100 to 300 keywords (depending on the niche) for the initial keyword research when we start working with a client. Tip #10. Start With Competitive Analysis Start every keyword research with competitive analysis. Extract the keywords your top 3 competitors are ranking on.  Then, use them as inspiration and build upon it. Use tools like UberSuggest to help generate new keyword ideas. Tip #11. Get SEMrush of Ahrefs You NEED SEMrush or Ahrefs, there’s no doubt about it. While they might seem expensive at a glance (99 USD per month billed annually), they’re going to save you a lot of manpower doing menial SEO tasks. Tip #12. Don’t Overdo It With SEO Tools Don’t overdo it with SEO tools. There are hundreds of those out there, and if you’re the type that’s into SaaS, you might be tempted to play around with dozens at a time. And yes, to be fair, most of these tools ARE helpful one way or another. To effectively do organic SEO, though, you don’t really need that many tools. In most cases, you just need the following: SEMrush/Ahrefs Screaming Frog RankMath/Yoast SEO Whichever outreach tool you prefer (our favorite is snov.io). Tip #13. Try Some of the Optional Tools In addition to the tools we mentioned before, you can also try the following 2 which are pretty useful & popular in the SEO community: Surfer SEO - helps with on-page SEO and creating content briefs for writers. ClusterAI - tool that helps simplify keyword research & save time. Tip #14. Constantly Source Writers Want to take your content production to the next level? You’ll need to hire more writers.  There is, however, one thing that makes this really, really difficult: 95 - 99% of writers applying for your gigs won’t be relevant. Up to 80% will be awful at writing, and the remainder just won’t be relevant for your niche. So, in order to scale your writing team, we recommend sourcing constantly, and not just once every few months. Tip #15. Create a Process for Writer Filtering As we just mentioned, when sourcing writers, you’ll be getting a ton of applicants, but most won’t be qualified. Fun fact \- every single time we post a job ad on ProBlogger, we get around 300 - 500 applications (most of which are totally not relevant). Trust us, you don’t want to spend your time going through such a huge list and checking out the writer samples. So, instead, we recommend you do this: Hire a virtual assistant to own the process of evaluating and short-listing writers. Create a process for evaluating writers. We recommend evaluating writers by: Level of English. If their samples aren’t fluent, they’re not relevant. Quality of Samples. Are the samples engaging / long-form content, or are they boring 500-word copy-pastes? Technical Knowledge. Has the writer written about a hard-to-explain topic before? Anyone can write about simple topics like traveling - you want to look for someone who knows how to research a new topic and explain it in a simple and easy to read way. If someone’s written about how to create a perfect cover letter, they can probably write about traveling, but the opposite isn’t true. The VA constantly evaluates new applicants and forwards the relevant ones to the editor. The editor goes through the short-listed writers and gives them trial tasks and hires the ones that perform well. Tip #16. Use The Right Websites to Source Writers “Is UpWork any good?” This question pops up on social media time and time again. If you ask us, no, UpWork is not good at all. Of course, there are qualified writers there (just like anywhere else), but from our experience, those writers are few and far in-between. Instead, here are some of our favorite ways to source writers: Cult of Copy Job Board ProBlogger Headhunting on LinkedIn If you really want to use UpWork, use it for headhunting (instead of posting a job ad) Tip #17. Hire Writers the Right Way If you want to seriously scale your content production, hire your writers full-time. This (especially) makes sense if you’re a content marketing agency that creates a TON of content for clients all the time. If you’re doing SEO just for your own blog, though, it usually makes more sense to use freelancers. Tip #18. Topic Authority Matters Google keeps your website's authoritativeness in mind. Meaning, if you have 100 articles on digital marketing, you’re probably more of an authority on the topic than someone that has just 10. Hence, Google is a lot more likely to reward you with better rankings. This is also partially why content volume really matters: the more frequently you publish content, the sooner Google will view you as an authority. Tip #19. Focus on One Niche at a Time Let’s say your blog covers the following topics: sales, accounting, and business management.  You’re more likely to rank if you have 30 articles on a single topic (e.g. accounting) than if you have 10 articles on each. So, we recommend you double-down on one niche instead of spreading your content team thin with different topics. Tip #20. Don’t Fret on the Details While technical SEO is important, you shouldn’t get too hung up on it.  Sure, there are thousands of technical tips you can find on the internet, and most of them DO matter. The truth, though, is that Google won’t punish you just because your website doesn’t load in 3 milliseconds or there’s a meta description missing on a single page. Especially if you have SEO fundamentals done right: Get your website to run as fast as possible. Create a ton of good SEO content. Get backlinks for your website on a regular basis. You’ll still rank, even if your website isn’t 100% optimized. Tip #21. Do Yourself a Favor and Hire a VA There are a TON of boring SEO tasks that your team should really not be wasting time with. So, hire a full-time VA to help with all that. Some tasks you want to outsource include gathering contacts to reach out to for link-building, uploading articles on WordPress, etc. Tip #22. Google Isn’t Everything While Google IS the dominant search engine in most parts of the world, there ARE countries with other popular search engines.  If you want to improve your SEO in China, for example, you should be more concerned with ranking on Baidu. Targeting Russia? Focus on Yandex. Tip #23. No, Voice Search is Still Not Relevant Voice search is not and will not be relevant (no matter what sensationalist articles might say). It’s just too impractical for most search queries to use voice (as opposed to traditional search). Tip #24. SEO Is Not Dead SEO is not dead and will still be relevant decades down the line. Every year, there’s a sensationalist article talking about this.  Ignore those. Tip #25. Doing Local SEO? Focus on Service Pages If you’re doing local SEO, focus on creating service-based landing pages instead of content.  E.g. if you’re an accounting firm based in Boston, you can make a landing page about /accounting-firm-boston/, /tax-accounting-boston/, /cpa-boston/, and so on. Thing is, you don’t really need to rank on global search terms - you just won’t get leads from there. Even if you ranked on the term “financial accounting,” it wouldn’t really matter for your bottom line that much. Tip #26. Learn More on Local SEO Speaking of local SEO, we definitely don’t do the topic justice in this guide. There’s a lot more you need to know to do local SEO effectively and some of it goes against the general SEO advice we talk about in this article (e.g. you don't necessarily need blog content for local SEO). We're going to publish an article on that soon enough, so if you want to check it out, DM me and I'll hit you up when it's up. Tip #27. Avoid Vanity Metrics Don’t get side-tracked by vanity metrics.  At the end of the day, you should care about how your traffic impacts your bottom line. Fat graphs and lots of traffic are nice and all, but none of it matters if the traffic doesn’t have the right search intent to convert to your product/service. Tip #28. Struggling With SEO? Hire an Expert Failing to make SEO work for your business? When in doubt, hire an organic SEO consultant or an SEO agency.  The #1 benefit of hiring an SEO agency or consultant is that they’ve been there and done that - more than once. They might be able to catch issues an inexperienced SEO can’t. Tip #29. Engage With the Community Need a couple of SEO questions answered?  SEO pros are super helpful & easy to reach! Join these Facebook groups and ask your question - you’ll get about a dozen helpful answers! SEO Signals Lab SEO & Content Marketing The Proper SEO Group. Tip #30. Stay Up to Date With SEO Trends SEO is always changing - Google is constantly pumping out new updates that have a significant impact on how the game is played.  Make sure to stay up to date with the latest SEO trends and Google updates by following the Google Search Central blog. Tip #31. Increase Organic CTR With PPC Want to get the most out of your rankings? Run PPC ads for your best keywords. Googlers who first see your ad are more likely to click your organic listing. Content & On-Page SEO Tips Tip #32. Create 50% Longer Content On average, we recommend you create an article that’s around 50% longer than the best article ranking on the keyword.  One small exception, though, is if you’re in a super competitive niche and all top-ranking articles are already as comprehensive as they can be. For example, in the VPN niche, all articles ranking for the keyword “best VPN” are around 10,000 - 11,000 words long. And that’s the optimal word count - even if you go beyond, you won’t be able to deliver that much value for the reader to make it worth the effort of creating the content. Tip #33. Longer Is Not Always Better Sometimes, a short-form article can get the job done much better.  For example, let’s say you’re targeting the keyword “how to tie a tie.”  The reader expects a short and simple guide, something under 500 words, and not “The Ultimate Guide to Tie Tying for 2021 \[11 Best Tips and Tricks\]” Tip #34. SEO is Not Just About Written Content Written content is not always best. Sometimes, videos can perform significantly better. E.g. If the Googler is looking to learn how to get a deadlift form right, they’re most likely going to be looking for a video. Tip #35. Don’t Forget to Follow Basic Optimization Tips For all your web pages (articles included), follow basic SEO optimization tips. E.g. include the keyword in the URL, use the right headings etc.  Just use RankMath or YoastSEO for this and you’re in the clear! Tip #36. Hire Specialized Writers When hiring content writers, try to look for ones that specialize in creating SEO content.  There are a LOT of writers on the internet, plenty of which are really good.  However, if they haven’t written SEO content before, chances are, they won’t do that good of a job. Tip #37. Use Content Outlines Speaking of writers - when working with writers, create a content outline that summarizes what the article should be about and what kind of topics it needs to cover instead of giving them a keyword and asking them to “knock themselves out.”   This makes it a lot more likely for the writer to create something that ranks. When creating content outlines, we recommend you include the following information: Target keyword Related keywords that should be mentioned in the article Article structure - which headings should the writer use? In what order? Article title Tip #38. Find Writers With Niche Knowledge Try to find a SEO content writer with some experience or past knowledge about your niche. Otherwise, they’re going to take around a month or two to become an expert. Alternatively, if you’re having difficulty finding a writer with niche knowledge, try to find someone with experience in technical or hard to explain topics. Writers who’ve written about cybersecurity in the past, for example, are a lot more likely to successfully cover other complicated topics (as opposed to, for example, a food or travel blogger). Tip #39. Keep Your Audience’s Knowledge in Mind When creating SEO content, always keep your audience’s knowledge in mind. If you’re writing about advanced finance, for example, you don’t need to teach your reader what an income statement is. If you’re writing about income statements, on the other hand, you’d want to start from the very barebone basics. Tip #40. Write for Your Audience If your readers are suit-and-tie lawyers, they’re going to expect professionally written content. 20-something hipsters? You can get away with throwing a Rick and Morty reference here and there. Tip #41. Use Grammarly Trust us, it’ll seriously make your life easier! Keep in mind, though, that the app is not a replacement for a professional editor. Tip #42. Use Hemingway Online content should be very easy to read & follow for everyone, whether they’re a senior profession with a Ph.D. or a college kid looking to learn a new topic. As such, your content should be written in a simple manner - and that’s where Hemingway comes in. It helps you keep your blog content simple. Tip #43. Create Compelling Headlines Want to drive clicks to your articles? You’ll need compelling headlines. Compare the two headlines below; which one would you click? 101 Productivity Tips \[To Get Things Done in 2021\] VS Productivity Tips Guide Exactly! To create clickable headlines, we recommend you include the following elements: Keyword Numbers Results Year (If Relevant) Tip #44. Nail Your Blog Content Formatting Format your blog posts well and avoid overly long walls of text. There’s a reason Backlinko content is so popular - it’s extremely easy to read and follow. Tip #45. Use Relevant Images In Your SEO Content Key here - relevant. Don’t just spray random stock photos of “office people smiling” around your posts; no one likes those.  Instead, add graphs, charts, screenshots, quote blocks, CSS boxes, and other engaging elements. Tip #46. Implement the Skyscraper Technique (The Right Way) Want to implement Backlinko’s skyscraper technique?  Keep this in mind before you do: not all content is meant to be promoted.  Pick a topic that fits the following criteria if you want the internet to care: It’s on an important topic. “Mega-Guide to SaaS Marketing” is good, “top 5 benefits of SaaS marketing” is not. You’re creating something significantly better than the original material. The internet is filled with mediocre content - strive to do better. Tip #47. Get The URL Slug Right for Seasonal Content If you want to rank on a seasonal keyword with one piece of content (e.g. you want to rank on “saas trends 2020, 2021, etc.”), don’t mention the year in the URL slug - keep it /saas-trends/ and just change the headline every year instead.  If you want to rank with separate articles, on the other hand (e.g. you publish a new trends report every year), include the year in the URL. Tip #48. Avoid content cannibalization.  Meaning, don’t write 2+ articles on one topic. This will confuse Google on which article it should rank. Tip #49. Don’t Overdo Outbound Links Don’t include too many outbound links in your content. Yes, including sources is good, but there is such a thing as overdoing it.  If your 1,000 word article has 20 outbound links, Google might consider it as spam (even if all those links are relevant). Tip #50. Consider “People Also Ask” To get the most out of SERP, you want to grab as many spots on the search result as possible, and this includes “people also ask (PAA):” Make a list of the topic’s PAA questions and ensure that your article answers them.  If you can’t fit the questions & answers within the article, though, you can also add an FAQ section at the end where you directly pose these questions and provide the answers. Tip #51. Optimize For Google Snippet Optimize your content for the Google Snippet. Check what’s currently ranking as the snippet. Then, try to do something similar (or even better) in terms of content and formatting. Tip #52. Get Inspired by Viral Content Want to create content that gets insane shares & links?  Reverse-engineer what has worked in the past. Look up content in your niche that went viral on Reddit, Hacker News, Facebook groups, Buzzsumo, etc. and create something similar, but significantly better. Tip #53. Avoid AI Content Tools No, robots can’t write SEO content.  If you’ve seen any of those “AI generated content tools,” you should know to stay away. The only thing those tools are (currently) good for is creating news content. Tip #54. Avoid Bad Content You will never, ever, ever rank with one 500-word article per week.  There are some SEO agencies (even the more reputable ones) that offer this as part of their service. Trust us, this is a waste of time. Tip #55. Update Your Content Regularly Check your top-performing articles annually and see if there’s anything you can do to improve them.  When most companies finally get the #1 ranking for a keyword, they leave the article alone and never touch it again… ...Until they get outranked, of course, by someone who one-upped their original article. Want to prevent this from happening? Analyze your top-performing content once a year and improve it when possible. Tip #56. Experiment With CTR Do your articles have low CTR? Experiment with different headlines and see if you can improve it.  Keep in mind, though, that what a “good CTR” is really depends on the keyword.  In some cases, the first ranking will drive 50% of the traffic. In others, it’s going to be less than 15%. Link-Building Tips Tip #57. Yes, Links Matter. Here’s What You Need to Know “Do I need backlinks to rank?” is probably one of the most common SEO questions.  The answer to the question (alongside all other SEO-related questions) is that it depends on the niche.  If your competitors don’t have a lot of backlinks, chances are, you can rank solely by creating superior content. If you’re in an extremely competitive niche (e.g. VPN, insurance, etc.), though, everyone has amazing, quality content - that’s just the baseline.  What sets top-ranking content apart from the rest is backlinks. Tip #58. Sometimes, You’ll Have to Pay For Links Unfortunately, in some niches, paying for links is unavoidable - e.g. gambling, CBD, and others. In such cases, you either need a hefty link-building budget, or a very creative link-building campaign (create a viral infographic, news-worthy story based on interesting data, etc.). Tip #59. Build Relationships, Not Links The very best link-building is actually relationship building.  Make a list of websites in your niche and build a relationship with them - don’t just spam them with the standard “hey, I have this amazing article, can you link to it?”.  If you spam, you risk ruining your reputation (and this is going to make further outreach much harder). Tip #60. Stick With The Classics At the end of the day, the most effective link-building tactics are the most straightforward ones:  Direct Outreach Broken Link-Building Guest Posting Skyscraper Technique Creating Viral Content Guestposting With Infographics Tip #61. Give, Don’t Just Take! If you’re doing link-building outreach, don’t just ask for links - give something in return.  This will significantly improve the reply rate from your outreach email. If you own a SaaS tool, for example, you can offer the bloggers you’re reaching out to free access to your software. Or, alternatively, if you’re doing a lot of guest posting, you can offer the website owner a link from the guest post in exchange for the link to your website. Tip #62. Avoid Link Resellers That guy DMing you on LinkedIn, trying to sell you links from a Google Sheet?  Don’t fall for it - most of those links are PBNs and are likely to backfire on you. Tip #63. Avoid Fiverr Like The Plague Speaking of spammy links, don’t touch anything that’s sold on Fiverr - pretty much all of the links there are useless. Tip #64. Focus on Quality Links Not all links are created equal. A link is of higher quality if it’s linked from a page that: Is NOT a PBN. Doesn’t have a lot of outbound links. If the page links to 20 other websites, each of them gets less link juice. Has a lot of (quality) backlinks. Is part of a website with a high domain authority. Is about a topic relevant to the page it’s linking to. If your article about pets has a link from an accounting blog, Google will consider it a bit suspicious. Tip #65. Data-Backed Content Just Works Data-backed content can get insane results for link-building.  For example, OKCupid used to publish interesting data & research based on how people interacted with their platform and it never failed to go viral. Each of their reports ended up being covered by dozens of news media (which got them a ton of easy links). Tip #66. Be Creative - SEO Is Marketing, After All Be novel & creative with your link-building initiatives.  Here’s the thing: the very best link-builders are not going to write about the tactics they’re using.  If they did, you’d see half the internet using the exact same tactic as them in less than a week! Which, as you can guess, would make the tactic cliche and significantly less effective. In order to get superior results with your link-building, you’ll need to be creative - think about how you can make your outreach different from what everyone does. Experiment it, measure it, and improve it till it works! Tip #67. Try HARO HARO, or Help a Reporter Out, is a platform that matches journalists with sources. You get an email every day with journalists looking for experts in specific niches, and if you pitch them right, they might feature you in their article or link to your website. Tip #68. No-Follow Links Aren’t That Bad Contrary to what you might’ve heard, no-follow links are not useless. Google uses no-follow as more of a suggestion than anything else.  There have been case studies that prove Google can disregard the no-follow tag and still reward you with increased rankings. Tip #69. Start Fresh With an Expired Domain Starting a new website? It might make sense to buy an expired one with existing backlinks (that’s in a similar niche as yours). The right domain can give you a serious boost to how fast you can rank. Tip #70. Don’t Overspend on Useless Links “Rel=sponsored” links don’t pass pagerank and hence, won’t help increase your website rankings.  So, avoid buying links from media websites like Forbes, Entrepreneur, etc. Tip #71. Promote Your Content Other than link-building, focus on organic content promotion. For example, you can repost your content on Facebook groups, LinkedIn, Reddit, etc. and focus on driving traffic.  This will actually lead to you getting links, too. We got around 95 backlinks to our SEO case study article just because of our successful content promotion. Tons of people saw the article on the net, liked it, and linked to it from their website. Tip #72. Do Expert Roundups Want to build relationships with influencers in your niche, but don’t know where to start?  Create an expert roundup article. If you’re in the sales niche, for example, you can write about Top 21 Sales Influencers in 2021 and reach out to the said influencers letting them know that they got featured. Trust us, they’ll love you for this! Tip #73. .Edu Links are Overhyped .edu links are overrated. According to John Mueller, .edu domains tend to have a ton of outbound links, and as such, Google ignores a big chunk of them. Tip #74. Build Relationships With Your Customers Little-known link-building hack: if you’re a SaaS company doing SEO, you can build relationships with your customers (the ones that are in the same topical niche as you are) and help each other build links! Tip #75. Reciprocal Links Aren’t That Bad Reciprocal links are not nearly as bad as Google makes them out to be. Sure, they can be bad at scale (if trading links is all you’re doing). Exchanging a link or two with another website / blog, though, is completely harmless in 99% of cases. Tip #76. Don’t Overspam Don’t do outreach for every single post you publish - just the big ones.  Most people already don’t care about your outreach email. Chances are, they’re going to care even less if you’re asking them to link to this new amazing article you wrote (which is about the top 5 benefits of adopting a puppy). Technical SEO Tips Tip #77. Use PageSpeed Insights If your website is extremely slow, it’s definitely going to impact your rankings. Use PageSpeed Insights to see how your website is currently performing. Tip #78. Load Speed Matters While load speed doesn’t impact rankings directly, it DOES impact your user experience. Chances are, if your page takes 5 seconds to load, but your competition’s loads instantly, the average Googler will drop off and pick them over you. Tip #79. Stick to a Low Crawl Depth Crawl depth of any page on your website should be lower than 4 (meaning, any given page should be possible to reach in no more than 3 clicks from the homepage).  Tip #80. Use Next-Gen Image Formats Next-gen image formats such as JPEG 2000, JPEG XR, and WebP can be compressed a lot better than PNG or JPG. So, when possible, use next-get formats for images on your website. Tip #81. De-Index Irrelevant Pages Hide the pages you don’t want Google to index (e.g: non-public, or unimportant pages) via your Robots.txt. If you’re a SaaS, for example, this would include most of your in-app pages or your internal knowledge base pages. Tip #82. Make Your Website Mobile-Friendly Make sure that your website is mobile-friendly. Google uses “mobile-first indexing.” Meaning, unless you have a working mobile version of your website, your rankings will seriously suffer. Tip #83. Lazy-Load Images Lazy-load your images. If your pages contain a lot of images, you MUST activate lazy-loading. This allows images that are below the screen, to be loaded only once the visitor scrolls down enough to see the image. Tip #84. Enable Gzip Compression Enable Gzip compression to allow your HTML, CSS and JS files to load faster. Tip #85. Clean Up Your Code If your website loads slowly because you have 100+ external javascript files and stylesheets being requested from the server, you can try minifying, aggregating, and inlining some of those files. Tip 86. Use Rel-Canonical Have duplicate content on your website? Use rel-canonical to show Google which version is the original (and should be prioritized for search results). Tip #87. Install an SSL Certificate Not only does an SSL certificate help keep your website safe, but it’s also a direct ranking factor. Google prioritizes websites that have SSL certificates over the ones that don’t. Tip #88. Use Correct Anchor Texts for Internal Links When linking to an internal page, mention the keyword you’re trying to rank for on that page in the anchor text. This helps Google understand that the page is, indeed, about the keyword you’re associating it with. Tip #89. Use GSC to Make Sure Your Content is Interlinked Internal links can have a serious impact on your rankings. So, make sure that all your blog posts (especially the new ones) are properly linked to/from your past content.  You can check how many links any given page has via Google Search Console. Tip #90. Bounce rate is NOT a Google ranking factor. Meaning, you can still rank high-up even with a high bounce rate. Tip #91. Don’t Fret About a High Bounce Rate Speaking of the bounce rate, you’ll see that some of your web pages have a higher-than-average bounce rate (70%+).  While this can sometimes be a cause for alarm, it’s not necessarily so. Sometimes, the search intent behind a given keyword means that you WILL have a high bounce rate even if your article is the most amazing thing ever.  E.g. if it’s a recipe page, the reader gets the recipe and bounces off (since they don’t need anything else). Tip #92. Google Will Ignore Your Meta Description More often than not, Google won’t use the meta description you provide - that’s normal. It will, instead, automatically pick a part of the text that it thinks is most relevant and use it as a meta description. Despite this, you should always add a meta description to all pages. Tip #93. Disavow Spammy & PBN Links Keep track of your backlinks and disavow anything that’s obviously spammy or PBNy. In most cases, Google will ignore these links anyway. However, you never know when a competitor is deliberately targeting you with too many spammy or PBN links (which might put you at risk for being penalized). Tip #94. Use The Correct Redirect  When permanently migrating your pages, use 301 redirect to pass on the link juice from the old page to the new one. If the redirect is temporary, use a 302 redirect instead. Tip #95. When A/B Testing, Do This A/B testing two pages? Use rel-canonical to show Google which page is the original. Tip #96. Avoid Amp DON’T use Amp.  Unless you’re a media company, Amp will negatively impact your website. Tip #97. Get Your URL Slugs Right Keep your blog URLs short and to-the-point. Good Example: apollodigital.io/blog/seo-case-study Bad Example: apollodigital.io/blog/seo-case-study-2021-0-to-200,000/ Tip #98. Avoid Dates in URLs An outdated date in your URL can hurt your CTR. Readers are more likely to click / read articles published recently than the ones written years back. Tip #99. Social Signals Matter Social signals impact your Google rankings, just not in the way you think. No, your number of shares and likes does NOT impact your ranking at all.  However, if your article goes viral and people use Google to find your article, click it, and read it, then yes, it will impact your rankings.  E.g. you read our SaaS marketing guide on Facebook, then look up “SaaS marketing” on Google, click it, and read it from there. Tip #100. Audit Your Website Frequently Every other month, crawl your website with ScreamingFrog and see if you have any broken links, 404s, etc. Tip #101. Use WordPress Not sure which CMS platform to use?  99% of the time, you’re better off with WordPress.  It has a TON of plugins that will make your life easier.  Want a drag & drop builder? Use Elementor. Wix, SiteGround and similar drag & drops are bad for SEO. Tip #102. Check Rankings the Right Way When checking on how well a post is ranking on Google Search Console, make sure to check Page AND Query to get the accurate number.  If you check just the page, it’s going to give you the average ranking on all keywords the page is ranking for (which is almost always going to be useless data). Conclusion Aaand that's about it - thanks for the read! Now, let's circle back to Tip #1 for a sec. Remember when we said a big chunk of what you read on SEO is based on personal experiences, experiments, and the like? Well, the tips we've mentioned are part of OUR experience. Chances are, you've done something that might be different (or completely goes against) our advice in this article. If that's the case, we'd love it if you let us know down in the comments. If you mention something extra-spicy, we'll even include it in this article.

I realized that AI will create equal footing for non-technical / non-coders compared to coders
reddit
LLM Vibe Score0
Human Vibe Score1
MatanNahmaniThis week

I realized that AI will create equal footing for non-technical / non-coders compared to coders

Hey fellow entrepreneurs, I started my current entrepreneurial journey following the advice to “build something that solves a problem you have.” As a coder, I wanted to code faster/better/stronger/etc. So I tried out dozens of AI coding tools to see the state of the market.  I took the best components I saw and started making my own flavor of tool, but sort of shelved it because as a coder I felt that the results were a bit alien (such as getting the AI to follow my code style, write idiomatic code, or refactor the same way I would.) I concluded that building AI coding tools for coders is tricky because as coders we’re so particular about the specifics of our code. Meanwhile, my absolutely non-technical friend was hitting me up to help him build a website for a new real-estate company that he’s launching, and he wanted my help. I really respect his hustle, but I was swamped trying to figure out my own product/market, so I told him he could use my AI coder and I would try to help out when he got stuck. He didn’t get stuck though, not once, and he launched his site over the weekend. I was truly shocked he did it all on his own, so I asked him to share his logs. It was wild – he managed to code a more or less state of the art website (good design, SEO, well-structured source code, Google Analytics, mailing lists. etc.) with absolutely no help. It cost him less than $100 in AI credits, instead of the price quotes of $20,000 - $50,000 from freelancers and agencies. Now I’m seriously pursuing AI coding tools again, but this time with a new passion: AI for non-coder / non-technical people is a 100x game changer. I think 2025 is going to be the year of the entrepreneur, where there will be a hundred times the businesses started because what held people back before was the lack of a technical co-founder or the cash to compensate engineers. Now it costs next to nothing to get started. I’m curious if anyone else has had a similar realization? Anyway, I’ve put the link below to my GitHub if you want to try it (open source, you pay for AI credits). But the main reason for my post is that I feel like I’m living in this new world of realization that being a human on earth is going to get a LOT more interesting in the coming years. There’s literally no excuse to take a job you hate, and nothing stopping people from launching a business. For anyone interested in checking it out or providing feedback you can search for kodu ai on github or kodu ai on google Best of luck to everyone on your entrepreneurial journey! P.s not sure if this is the right flair

Secret behind Airbnb's Billion-Dollar Empire? Spamming Craigslist
reddit
LLM Vibe Score0
Human Vibe Score1
deadcoder0904This week

Secret behind Airbnb's Billion-Dollar Empire? Spamming Craigslist

Silicon Valley wants you to believe that their unicorn startups succeeded doing things legally. But that couldn't be far from truth. For starters, Airbnb used multiple Gmail accounts to spam Craigslist. "They posted unrealistically (fake) cheap rentals of beautiful apartments in places where normal rent should be 10x more. Once people replied, they auto-responded that the unit has been rented, but they should be looking for another unit on AirBnB." The Game of Blackhat is a cat-and-mouse game. You need a lot of guardrails to protect yourself from people using your Social Site by spamming their products. Craigslist is a team of 30 people. There's stuff AI can automate now with such a small team but back then, it wasn't possible. Airbnb used Craigslist as its playground to spam Craigslist visitors to grow their supply-side. In a 2-sided marketplace, growing both supply and demand is very important. And both must grow at the same time for the marketplace to work. A Blackhat Marketer created a new test site to get vacation rental owners to sign-up so that he can test his Airbnb theory. He grabbed their real email-addresses (not Craigslist anonymous addresses) via Craigslist by specifically targeting those who were advertising their vacation rentals on Craigslist. He skipped over the other categories that were directly related to AirBnB's business model because they didn't fit with the test site he built. Once he got 1000+ sign-ups, he then took it upon himself to post it to the advertising section on Craigslist. The email said this: I am emailing you because you have one of the nicest listings on Craigslist in Idaho and I want to recommend you feature it (for free) on one of the largest Idaho housing sites on the web, Airbnb. The site already has 3,000,000 pages views a month. Check it out here to list now: airbnb(dot)com Sarah Surpisingly, all emails were by ladies. He did the same in Week 2 and Week 3 to test if it wasn't a one-time thing. Surely, it wasn't a fluke. After posting 4 ads on Craigslist in 3 weeks, he received 5 identical emails from 2 ladies who were raving fans of AirBnB and spent their days emailing Craigslist advertisers. This is one of the greatest blackhat strategies used in the real world to build a billion-dollar marketplace by growing the supply-side with pure blackhat. These strategies are not mentioned in Press Interviews, Media, or any Founder stories but this is probably the most important piece of the puzzle. Without it, Airbnb probably wouldn't have survived. "Some very famous investors have alluded to the fact that they look for a dangerous streak in the entrepreneurs they invest in…and while those investors will never come out and tell you what they mean, this kind of thing is probably what they mean." It definitely violates CAN-SPAM act. Some comments from Hacker News: "CAN-SPAM, sending from a fake address (illegal headers). CA has a specific law that pre-empts CAN-SPAM that definitely makes this illegal if sent from CA." But I guess it worked in Airbnb's favour lol as they were never caught or fined until after. "It's commercial email 100%. Probably a fake sender name (illegal), against gmail ToS, against CL ToS and no unsubscribe link and no one even subscribed in the first place. 100% against CAN-SPAM." Thanks for reading. If you'd like to learn more blackhat tactics like this, check this site which is a growth hacking newsletter with real-world blackhat examples. PS: Actual emails & screenshots from the Airbnb x Craigslist spam can be found here.

AI Will Make You Extremely Rich or Kill Your Business in 2024
reddit
LLM Vibe Score0
Human Vibe Score1
AntsyNursery58This week

AI Will Make You Extremely Rich or Kill Your Business in 2024

Preface: I'm a solo-founder in the AI space and previously worked as an ML scientist; the new advancements in AI that I'm seeing are going to impact everyone here. It doesn't matter if you're just starting out, or a bootstrapped brick and mortar founder, or even a VC backed hard tech founder. Last year was when the seeds were laid, and this is the year we'll see them bloom. There will be an onslaught of advancements that take place that are borderline inconceivable due to the nature of exponential progress. This will change every single vertical. I'm making this post because I think AI execution strategy will make or break businesses. Dramatically. Over $50B was put into AI startups in 2023 alone. This figure excludes the hundreds of billions poured into AI from enterprises. So, let's follow the money: &#x200B; 1) AI enterprise software. There's a lot to unpack here and this is what I’m currently working on. AI enterprise software will encompass everything from hyper personalized email outbound to AI cold calls to AI that A/B tests ads on synthetic data to vertical specific software. The impact of the former is relatively self explanatory, so I'll focus on the latter. To illustrate vertical specific AI software, I'll use a simple example in the legal space. Lawyers typically have to comb through thousands of pages of documents. Now, using an LLM + a VDB, an AI can instantly answer all of those questions while surfacing the source and highlighting the specific answer in the contract/document. There are dozens of AI startups for this use case alone. This saves lawyers an immense amount of time and allows them to move faster. Firms that adopt this have a fundamental advantage over law firms that don't adopt this. This was 2023 technology. I'm seeing vertical AI software getting built by my friends in areas from construction, to real estate, to even niche areas like chimney manufacturing. This will exist everywhere. Now, this can be extrapolated much further to be applicable to systems that can do reports and even browse the Internet. This brings me to my next point. &#x200B; 2) AI information aggregation and spread. My gut tells me that this will have a crescendo moment in the future with hardware advancements (Rabbit, Tab, etc.). You won't have to google things because it will be surfaced to you. It's predictive in nature. The people who can get information the fastest will grow their business the fastest. This part is semi-speculative, but due to the nature of LLMs being so expensive to train, I have a strong feeling that large institutions will have access to the \fastest\ and \best\ models that can do this quicker than you and I can. This is why it's important to stay on top. &#x200B; 3) AI content generation This is relevant to running advertisements and any digital marketing aspect of your business. If you can rapidly make content faster than your competitors to put in social media, you will outpace your competitors rapidly. I think most folks are familiar with MidJourney, Stable diffusion, etc. but don't know how to use it. You can generate consistent models for a clothing brand or generate images of a product that you would normally need to hire a professional photographer to take. There's also elevenlabs which is relatively easy to use and can be used to make an MP3 clip as a narration for an ad; this is something I've already done. I'm also still shocked by how many people are unfamiliar with tools like Pika which can do video generation. You could imagine companies having fleets of digital influencers that they control or conjuring up the perfect ad for a specific demographic using a combination of all of the aforementioned tools. &#x200B; In summary, if you feel like I'm being hyperbolic or propagating science fiction fantasies, you're likely already behind. I truly recommend that everyone stays up to date on these advancements as much as possible. If your competitor comes across an AI tool that can increase their ROAS by 5x they can crush you. If your competitor uses a tool that increases the rate at which they receive and aggregate information by 200% (modest estimate) they will crush you. If your competitors have a tool that can reduce their employee size, then they will use it. They'll fire their employees to cut costs and reinvest the money back into their business. It will compound to the point where you're outpaced, and this isn't a level of innovation we've seen since the birth of the industrial revolution. Your customers can get stolen overnight, or you can steal your competition’s customers overnight. TL;DR: This is an opportunity for entrepreneurs to scale faster than they could have possibly imagined, but this also comes with the potential for your company to be obliterated. We've never seen advancements that can have this drastic of an impact this quickly. Adoption will happen fast, and first movers will have a disproportionate and compounding advantage. Watch guides, meet with startups, follow the news, and get rich.

I sold my AI tool for $35,000
reddit
LLM Vibe Score0
Human Vibe Score1
marclouvThis week

I sold my AI tool for $35,000

Hey Entrepreneurs, Marc here. Last month I wrote here about how sold a habit tracker for $10,000 in October. Earlier this month, I got $35,000 in my bank account after selling a landing page maker with AI. Here's the story: &#x200B; April 2023: Just like everyone, I get massive FOMO with AI. I played with GPT and decided to build a landing page generator with AI: Input text and the AI prefills a template with copy and AI-generated images. I'm working on it with a good friend of mine named Martin. May: The product is called LandingAI. It's an MVP but we launched and made \~$8,000. Unfortunately, Martin and I had different visions for the project so we forked. &#x200B; June: LandingAI is the name of a big corp (bummer) so I rebranded it to MakeLanding. I ditch 90% of the code because users want a very different product: So here I am, building an entire website builder powered with AI... &#x200B; July: I launched again, but made a BIG mistake: I swapped the one-time payment for a monthly subscription and got $20 MRR for 15k visitors... If you can avoid subscriptions, do it New pricing means new positioning—users compared the app to Framer & Webflow August: I removed the subscription and sales came back: \~$7,000 in 3 months. But I realized this was going nowhere... September: I don't use the product The market is gigantic and crowded As a solopreneur, nothing is more important for me than building cool stuff for people I care about. And I didn't really care about this big market so... October: I called my friend Dan and he said: SELL. He was right. I bought my shares of LandingAI from Martin and listed MakeLanding on Acquire: Asking $38,000 for $14,000 TTM (3x profit) Within hours, I received dozens of NDAs and a buyer started the process 🤯 After a few weeks of NDA, LOI, Escrow, etc. the buyer sent the money but... Only a fraction of the transaction. Then he ghosted me. So I canceled the transition. Back to Acquire... Luckily, in 24 hours I got another buyer! &#x200B; November: Within weeks, the money was in my bank account. The buyer and I never called, just a few messages. It's mind-blowing. &#x200B; My takeaways: Don't build AI products just because Don't go on a massive market you don't care Sell if you don't know how to grow the product It's my 3rd acquisition this year. I love the freedom of build, sell, repeat.

How I went from $27 to $3K as a solopreneur still in a 9-5
reddit
LLM Vibe Score0
Human Vibe Score1
jottrledThis week

How I went from $27 to $3K as a solopreneur still in a 9-5

My journey started back in November 2023. I was scrolling through Twitter and YouTube and saw a word that I had never come across before. Solopreneur. The word caught my eye. Mainly because I was pretty sure I knew what it meant even though it's not a word you'll find in the dictionary. I liked what it was describing. A solo entrepreneur. A one man business. It completely resonated with me. As a software engineer by trade I'm used to working alone, especially since the pandemic hit and we were forced to work remotely. See, I always wanted to ditch the 9-5 thing but thought that was too big and too scary for a single person to do. Surely you would need a lot of money to get started, right? Surely you would need investors? The whole concept seemed impossible to me. That was until I found all the success stories. I became obsessed with the concept of solopreneurship. As I went further down the rabbit hole I found people like Justin Welsh, Kieran Drew and Marc Louvion to name a few. All of whom have one person businesses making huge money every year. So I thought, if they can do it, why can't I? People like this have cleared the pathway for those looking to escape the 9-5 grind. I decided 2024 would be the year I try this out. My main goal for the year? Build a one man business, earn my first $ online and learn a sh\*t ton along the way. My main goal in general? Build my business to $100K per year, quit my 9-5 and live with freedom. From December 2023 to February 2024 I began brainstorming ideas. I was like a lost puppy looking for his ball. How on earth did people find good ideas? I began writing everything and anything that came to mind down in my notes app on my phone. By February I would have approximately 70 ideas. Each as weird and whacky as the other. I was skeptical though. If I went through all the trouble of building a product for one of these ideas how would I know if anyone would even be interested in using it? I got scared and took a break for a week. All these ideas seemed too big and the chance that they would take off into the atmosphere was slim (in my mind anyways). I was learning more and more about solopreneurship as the weeks went on so I decided to build a product centered around everything I was learning about. The idea was simple. Enter a business idea and use AI to give the user details about how to market it, who their target customers were, what to write on their landing page, etc. All for a measly $27 per use. I quickly built it and launched on March 3rd 2024. I posted about it on Indie Hackers, Reddit and Hacker News. I was so excited about the prospect of earning my first internet $! Surely everyone wanted to use my product! Nope...all I got was crickets. I was quickly brought back down to earth. That was until 5 days later. I looked at my phone and had a new Stripe notification! Cha-ching! My first internet $. What a feeling! That was goal number 1 complete. It would be another 6 days before I would get my second sale...and then another 15 days to get my third. It was an emotional rollercoaster. I went from feeling like quitting the 9-5 was actually possible to thinking that maybe the ups and downs aren't worth it. On one hand I had made my first internet dollar so I should my ecstatic, and don't get me wrong, I was but I wanted more. More validation that I could do this long term. By May I was starting to give up on the product. I had learned so much in the past few months about marketing, SEO, building an audience, etc. and I wanted to build something that I thought could have more success so I focused on one critical thing that I had learned about. What was it? Building a product that had SEO potential. A product that I knew hundreds of people were looking for. See this was my thinking - If I could find a keyword that people were searching for on Google hundreds/thousands of times every month and it was easy to rank high on search engines then I would go all in (in SEO land this equates to a Keyword that has a Keyword Difficulty of = 500). I began researching and found that the keyword "micro saas ideas" was being searched for around 600 times each month. Micro Saas was something that really interested me. It was perfect for solopreneurs. Small software products that 1 person could build. What's not to like if you're in the game of software and solopreneurship? Researching keywords like this became like a game for me. I was hooked. I was doing it every day, finding gems that were being searched for hundreds and thousands of times every month that still had potential. That's when I came up with my next product idea. I decided to create a database of Micro Saas Ideas all with this sort of SEO potential. See if you can build a product that you know people are looking for then that's all the validation you need. So I put this theory to the test. I created a database of Micro Saas Ideas with SEO Potential and launched it in June 2024. This time it was different. I made $700 in the first week of launching. A large contrast to my previous failed attempt at becoming the worlds greatest solopreneur. Since launch I have grown the product to $3K and I couldn't be happier. I know what you're saying, $3K isn't a lot. But it's validation. It's validation that I can earn $ online. Validation that I can grow a business and it gives me hope that one day I'll be able to quit that 9-5 grind. My plan is to keep growing the business. I expect there to be a few challenges up ahead but I'll tackle them as I go and learn from the failures and successes. I have a newsletter where I share Micro Saas Ideas with SEO potential every week which I'll leave below in the first comment. Feel free to come along for the ride. If not I hope this post brings you some value If you're thinking about starting as a solopreneur, stop thinking and start doing, you won't regret it.

I have reviewed over 900+ AI Tools for my directory. Here are some of the best ones I have seen for entrepreneurs and startups.
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

I have reviewed over 900+ AI Tools for my directory. Here are some of the best ones I have seen for entrepreneurs and startups.

As one of the co-founders at AI Scout, a platform for AI discovery, I've had the privilege (and challenge) of reviewing over 900 AI tools submitted to our directory. I've filtered these down to some of the top AI tools that I believe could bring value to startups and entrepreneurs. It's worth noting that while these tools are great right out the box, the power of AI is truly realized when these tools are used in tandem and strategically aligned with your business needs. The challenge most people face is not about the lack of AI tools available, but the difficulty in finding the right one that fits their specific needs and workflows. Without further ado, here's my top pick of AI tools you should consider looking into if you are an entrepreneur or run a startup. Chatbase - Custom ChatGPT (Trained on Your Own Data) Taking a step up from traditional support bots, Chatbase combines the power of GPT and your own knowledge base. The result is a ChatGPT-like chatbot that is trained on your own websites and documents. You can embed the chatbot into your own website via an iframe or script in the header of your website code. They also have an API you can take advantage of. We use this personally at AI Scout for ScoutBud (AI assistant to find AI tools), which we trained based on our directory site. It would also work great if you have extensive documentation, papers, etc. that you want to quickly reference by simply asking a chatbot for the info you need instead of having to go through dozens of PDFs. Reply - AI-Powered Sales Engagement Platform Great AI tool to manage your entire sales engagement cycle. They have a large database with about a dozen filters to discover optimal B2B leads. From here, you can use their GPT integration to generate cold emails as well as handle responses and meeting scheduling. What I like personally about Reply are the endless integrations available, including Gmail, Outlook, Zoho, and major social platforms such as Twitter and LinkedIn. Instapage - AI Landing Page Generation, Testing, and Personalization This AI tool allows users to generate content variations for landing pages including headlines, paragraphs, and CTAs based on the target audience. You can also conduct A/B testing for more effective and efficient campaigns. Paired with hundreds of professional and cutomizable layouts, Instapage is definitely something I would recommend for entrepreneurs who want to get a high-converting landing page set up quickly and effectively. SaneBox - AI Emails Management If you feel overwhelmed by the sheer volume of emails you receive like myself and many entrepreneurs, this could be something for you. SaneBox’s AI identifies important emails and declutters your inbox, helping you to stay focused on what truly matters. SocialBee - AI Social Media Manager Think of SocialBee as your all-in-one social media command center, powered by AI. You can manage multiple social media accounts from one platform and generate captions with AI as well. SocialBee not only allows you to schedule posts but also helps you analyze growth and engagement with detailed reports. Works well with all social media platforms, including Facebook, Twitter, Instagram, and Linkedin. I believe they also have integrations for TikTok and YouTube, although I haven't tried these personally. MeetGeek - AI Meeting Assistant Lifesaver if you attend a lot of meetings or calls. Great for transcribing, summarizing, and sharing key insights from meetings. The AI also creates meeting highlights, which I've personally fouund quite useful if you ever need to get a very quick and dirty overview of what happened in a call. It also provides analysis (including sentiment evaluation) for meetings. Taskade - AI Productivity Tool for Task Management An all-in-one AI productivity tool. Multiple AI features available, including a chatbot, writing assistant, and workflow creator. It's a great all-around tool for real-time collaboration and efficient task management. Scribe AI (ScribeHow) - AI Documentation Generator Great for any SaaS applications where you need to create resources/documentations/guides for your app. You simply record your process and Scribe generates a written guide for you. Remember, while AI is an excellent assistant, it's also just a tool. The ultimate success of your venture depends on how effectively you leverage these tools. Happy experimenting!

Why Ignoring AI Agents in 2025 Will Kill Your Marketing Strategy
reddit
LLM Vibe Score0
Human Vibe Score1
frankiemuiruriThis week

Why Ignoring AI Agents in 2025 Will Kill Your Marketing Strategy

If you're still focusing solely on grabbing the attention of human beings with your marketing efforts, you're already behind. In 2025, the game will change. Good marketing will demand an in-depth understanding of the AI space, especially the AI Agent space. Why? Your ads and content won’t just be seen by humans anymore. They’ll be analyzed, indexed, and often acted upon by AI agents—automated systems that will be working on behalf of companies and consumers alike. Your New Audience: Humans + AI Agents It’s not just about appealing to people. Companies are employing AI robots to research, negotiate, and make purchasing decisions. These AI agents are fast, thorough, and unrelenting. Unlike humans, they can analyze millions of options in seconds. And if your marketing isn’t optimized for them, you’ll get filtered out before you even reach the human decision-maker. How to Prepare Your Marketing for AI Agents The companies that dominate marketing in 2025 will be the ones that master the art of capturing AI attention. To do this, marketers will need to: Understand the AI agents shaping their industry. Research how AI agents function in your niche. What are they prioritizing? How do they rank options? Create AI-friendly content. Design ads and messaging that are easily understandable and accessible to AI agents. This means clear metadata, structured data, and AI-readable formats. Invest in AI analytics. AI agents leave behind footprints. Tracking and analyzing their behavior is critical. Stay ahead of AI trends. The AI agent space is evolving rapidly. What works today might be obsolete tomorrow. How My Agency Adapted and Thrived in the AI Space At my digital agency, we saw this shift coming and decided to act early. In 2023, we started integrating AI optimization into our marketing strategies. One of our clients—a B2B SaaS company—struggled to get traction because their competitors were drowning them out in Google search rankings and ad platforms. By analyzing the algorithms and behaviors of AI agents in their space, we: Rewrote their website copy with structured data and optimized metadata that was more AI-agent friendly. Created ad campaigns with clear, concise messaging and technical attributes that AI agents could quickly process and index. Implemented predictive analytics to understand what AI agents would prioritize based on past behaviors. The results? Their website traffic doubled in three months, and their lead conversion rate skyrocketed by 40%. Over half of the traffic increase was traced back to AI agents recommending their platform to human users. The Takeaway In 2025, marketing won’t just be about human attention. It’ll be about AI attention—and that requires a completely different mindset. AI agents are not your enemy; they’re your new gatekeepers. Learn to speak their language, and you’ll dominate the marketing game.

Detailed Guide - How I've Been Self Employed for 2 Years Selling Posters
reddit
LLM Vibe Score0
Human Vibe Score1
tommo278This week

Detailed Guide - How I've Been Self Employed for 2 Years Selling Posters

Hey everyone, bit of context before you read through this. I have been selling POD posters full time for over 2 years now. My next venture is that I have started my own Print on Demand company for posters, PrintShrimp. As one way of creating customers for our service, we are teaching people for free how to also sell posters. Here is a guide I have written on how to sell posters on Etsy. Feel free to have a read through and then check out PrintShrimp, hopefully can help some of you guys out (and get us some more customers!) All of this is also available in video format on our website too, if you prefer to learn that way. Thanks guys! And as some people asked in other subs, no this isn't written with AI 😅 This took a couple of weeks to put together! Through this guide, we will teach you everything you need to know about starting to sell posters and generate some income. We will also show you why PrintShrimp is the best POD supplier for all of your poster needs. Trust me, you won’t need much convincing.  So, why are posters the best product to sell? Also, just thought I’d quickly answer the question - why posters? If you’ve been researching Print on Demand you’ve probably come across the infinite options of t-shirts, mugs, hats, phone cases, and more. All of these are viable options, however we think posters are the perfect place to start. You can always expand into other areas further down the line! So a brief summary of why posters are the perfect product for Print on Demand: \-They are very easy to design! Posters are a very easy shape to deal with - can’t go wrong with a rectangle. This makes designing products very easy. \-Similarly to this, what you see is what you get with a poster. You can literally see your finished product as you design it in either canva or photoshop. With T-Shirts for example, you have to make your design, and then place it on a t-shirt. Then you have to coordinate with your printers the size you would like the design on the tshirt and many other variables like that. There is no messing about with posters - what you see is what you get. \-The same high quality, everywhere. With other products, if you want to reap the benefits of a printing in various countries, you need to ensure each of your global suppliers stocks the same t-shirts, is able to print in the same way, carries the same sizes etc. Again with posters you avoid all of this hassle- your products will come out the same, no matter which of our global locations are used. \-They have a very favorable profit margin. As you will see later, the cost price of posters is very low. And people are prepared to pay quite a lot for a decent bit of wall art! I have tried out other products, and the profit margin combined with the order quantity of posters makes them my most profitable product, every single time. Using PrintShrimp, you can be sure to enjoy profits of anywhere between £6 - £40 pure profit per sale.  \-They are one of the easiest to print white label. This makes them perfect for Print on Demand. Your posters are simply put in a tube, and off they go. There are no extras you need to faff around with, compared to the extra elements other products come with, such as clothing labels on t-shirts.  Picking your poster niche So, you are ready to start selling posters. Great! Now, the blessing and curse with selling posters is that there are infinite possibilities regarding what you can sell. So, it can easily be quite overwhelming at first.  The first thing I would recommend doing is having a look at what others are selling. Etsy is a wonderful place for this (and will likely be a key part of your poster selling journey). So, log on to Etsy and simply type in ‘poster’ in the search bar. Get ready to write a massive list of the broad categories and type of posters that people are selling.  If you do not have more than 50 categories written down by the end, you are doing something wrong. There are seriously an infinite amount of posters! For example, here are some popular ones to get you started: Star sign posters, Kitchen posters, World map posters, Custom Dog Portrait posters, Music posters, Movie posters, Fine art posters, Skiing posters, Girl Power posters and Football posters.  Now, you have a huge list of potential products to sell. What next? There are a few important things you need to bear in mind when picking your niche: \-Does this interest me?  Don’t make the mistake of going down a niche that didn’t actually interest you just because it would probably be a money maker. Before you know it, what can be a very fun process of making designs can become incredibly \\\monotonous, and feel like a chore\\\. You need to bear in mind that you will be spending a lot of time creating designs - if it is something you are interested in you are much less likely to get burnt out! As well, \\\creativity will flow\\\ far better if it is something you are interested in, which at the end of the day will lead to better designs that are more likely to be purchased by customers.  \-Is this within my design range? Don’t let this put you off too much. We will go through how to get started on design later on in this guide. However, it is important to note that the plain truth of it is that some niches and designs are a hell of a lot more complicated than others. For example, quote posters can essentially be designed by anyone when you learn about how to put nice fonts together in a good color scheme. On the other hand, some posters you see may have been designed with complex illustrations in a program like Illustrator. To start with, it may be better to pick a niche that seems a bit more simple to get into, as you can always expand your range with other stores further down the line. A good way of evaluating the design complexity is by identifying if this poster is \\\a lot of elements put together\\\ or is \\\a lot of elements created by the designer themselves\\\\\.\\ Design can in a lot of cases be like a jigsaw - putting colours, shapes and text together to create an image. This will be a lot easier to start with and can be learnt by anyone, compared to complex drawings and illustrations.  \-Is this niche subject to copyright issues? Time to delve deep into good old copyright. Now, when you go through Etsy, you will without a doubt see hundreds of sellers selling music album posters, car posters, movie posters and more. Obviously, these posters contain the property of musicians, companies and more and are therefore copyrighted. The annoying thing is - these are \\\a complete cash cow.\\\ If you go down the music poster route, I will honestly be surprised if you \\don’t\\ make thousands. However it is only a matter of time before the copyright strikes start rolling in and you eventually get banned from Etsy.  So I would highly recommend \\\not making this mistake\\\. Etsy is an incredible platform for selling posters, and it is a hell of a lot easier to make sales on there compared to advertising your own website. And, you \\\only get one chance on Etsy.\\\ Once you have been banned once, you are not allowed to sign up again (and they do ID checks - so you won’t be able to rejoin again under your own name).  So, don’t be shortsighted when it comes to entering Print on Demand. If you keep your designs legitimate, they will last you a lifetime and you will then later be able to crosspost them to other platforms, again without the worry of ever getting shut down.  So, how do I actually design posters? Now you have an idea of what kind of posters you want to be making, it’s time to get creative and make some designs! Photoshop (and the creative cloud in general) is probably the best for this. However, when starting out it can be a scary investment (it costs about £30 a month unless you can get a student rate!).  So, while Photoshop is preferable in the long term, when starting out you can learn the ropes of design and get going with Canva. This can be great at the start as they have a load of templates that you can use to get used to designing and experimenting (while it might be tempting to slightly modify these and sell them - this will be quite saturated on places like Etsy so we would recommend doing something new).  What size format should I use? The best design format to start with is arguably the A sizes - as all the A sizes (A5, A4, A3, A2, A1, A0) are scalable. This means that you can make all of your designs in one size, for example A3, and these designs will be ready to fit to all other A sizes. For example, if you design an A3 poster and someone orders A1, you can just upload this A3 file to PrintShrimp and it will be ready to print. There is a wide range of other sizes you should consider offering on your shop, especially as these sizes are very popular with the American market. They have a wide range of popular options, which unfortunately aren’t all scalable with each other. This does mean that you will therefore have to make some slight modifications to your design in order to be able to offer them in American sizing, in a few different aspect ratios. What you can do however is design all of your products in UK sizing, and simply redesign to fit American sizing once you have had an order. Essentially: design in UK sizing, but list in both UK and US sizing. Then when you get a non-A size order, you can quickly redesign it on demand. This means that you don’t have to make a few different versions of each poster when first designing, and can simply do a quick redesign for US sizing when you need to. Below is PrintShrimps standard size offering. We can also offer any custom sizing too, so please get in touch if you are looking for anything else. With these sizes, your poster orders will be dispatched domestically in whatever country your customer orders from. Our recommendations for starting design One thing that will not be featured in this guide is a written out explanation or guide on how to design. Honestly, I can’t think of a more boring, or frankly worse, way to learn design. When it comes to getting started, experimenting is your best friend! Just have a play around and see what you can do. It is a really fun thing to get started with, and the satisfaction of when a poster design comes together is like no other. A good way to start is honestly by straight up copying a poster you see for sale online. And we don’t mean copying to sell! But just trying to replicate other designs is a great way to get a feel for it and what you can do. We really think you will be surprised at how easy it is to pull together a lot of designs that at first can appear quite complicated! Your best friend throughout this whole process will be google. At the start you will not really know how to do anything - but learning how to look into things you want to know about design is all part of the process. At first, it can be quite hard to even know how to search for what you are trying to do, but this will come with time (we promise). Learning how to google is a skill that you will learn throughout this process.  Above all, what we think is most important is this golden rule: take inspiration but do not steal. You want to be selling similar products in your niche, but not copies. You need to see what is selling in your niche and get ideas from that, but if you make designs too similar to ones already available, you won’t have much luck. At the end of the day, if two very similar posters are for sale and one shop has 1000 reviews and your newer one has 2, which one is the customer going to buy? You need to make yours offer something different and stand out enough to attract customers. Etsy SEO and maximizing your sales You may have noticed in this guide we have mentioned Etsy quite a few times! That is because we think it is hands down the best place to start selling posters. Why? Etsy is a go to place for many looking to decorate their homes and also to buy gifts. It might be tempting to start selling with your own website straight away, however we recommend Etsy as it brings the customers to you. For example, say you start selling Bathroom Posters. It is going to be a hell of a lot easier to convert sales when you already have customers being shown your page after searching ‘bathroom decor’, compared to advertising your own website. This is especially true as it can be hard to identify your ideal target audience to then advertise to via Meta (Facebook/Instagram) for example. Websites are a great avenue to explore eventually like I now have, but we recommend starting with Etsy and going from there. What costs do I need to be aware of? So, setting up an Etsy sellers account is currently costs £15. The only other upfront cost you will have is the cost of listing a product - this is 20 cents per listing. From then on, every time you make a sale you will be charged a transaction fee of 6.5%, a small payment processing fee, plus another 20 cents for a renewed listing fee. It normally works out to about 10% of each order, a small price to pay for all the benefits Etsy brings. No matter what platform you sell on, you will be faced with some form of transaction fee. Etsy is actually quite reasonable especially as they do not charge you to use their platform on a monthly basis.  What do I need to get selling? Getting your shop looking pretty \-Think of a shop name and design (now you are a professional designer) a logo \-Design a banner for the top of your shop \-Add in some about me info/shop announcement \-I recommend running a sale wherein orders of 3+ items get a 20% of discount. Another big benefit of PrintShrimp is that you receive large discounts when ordering multiple posters. This is great for attracting buyers and larger orders.  Making your products look attractive That is the bulk of the ‘decor’ you will need to do. Next up is placing your posters in mock ups! As you may notice on Etsy, most shops show their posters framed and hanging on walls. These are 99% of the time not real photos, but digital mock ups. This is where Photoshop comes in really handy, as you can automate this process through a plug in called Bulk Mock Up. If you don’t have photoshop, you can do this on Canva, you will just have to do it manually which can be rather time consuming.  Now, where can you get the actual Mock Ups? One platform we highly recommend for design in general is platforms like Envato Elements. These are design marketplaces where you have access to millions of design resources that you are fully licensed to use!  Titles, tags, and descriptions  Now for the slightly more nitty gritty part. You could have the world's most amazing looking poster, however, if you do not get the Etsy SEO right, no one is going to see it! We will take you through creating a new Etsy listing field by field so you can know how to best list your products.  The key to Etsy listing optimisation is to maximise. Literally cram in as many key words as you possibly can! Before you start this process, create a word map of anything you can think of relating to your listing. And come at this from the point of view of, if I was looking for a poster like mine, what would I search? Titles \-Here you are blessed with 140 characters to title your listing. Essentially, start off with a concise way of properly describing your poster. And then afterwards, add in as many key words as you can! Here is an example of the title of a well selling Skiing poster: Les Arcs Skiing Poster, Les Arcs Print, Les Alpes, France Ski Poster, Skiing Poster, Snowboarding Poster, Ski Resort Poster Holiday, French This is 139 characters out of 140 - you should try and maximise this as much as possible! As you can see, this crams in a lot of key words and search terms both related to Skiing as a whole, the poster category, and then the specifics of the poster itself (Les Arcs resort in France). Bear in mind that if you are listing a lot of listings that are of the same theme, you won’t have to spend time creating an entirely new title. For example if your next poster was of a ski resort in Italy, you can copy this one over and just swap out the specifics. For example change “France ski poster” to “Italy ski poster”, change “Les Arcs” to “The Dolomites”, etc.  Description \-Same logic applies for descriptions - try and cram in as many key words as you can! Here is an example for a Formula One poster: George Russell, Mercedes Formula One Poster  - item specific keywords Bright, modern and vibrant poster to liven up your home.  - Describes the style of the poster All posters are printed on high quality, museum grade 200gsm poster paper. Suitable for framing and frames. - Shows the quality of the print. Mentions frames whilst showing it comes unframed Experience the thrill of the racetrack with this stunning Formula One poster. Printed on high-quality paper, this racing car wall art print features a dynamic image of a Formula One car in action, perfect for adding a touch of speed and excitement to any motorsports room or man cave. Whether you're a die-hard fan or simply appreciate the adrenaline of high-speed racing, this poster is sure to impress. Available in a range of sizes, it makes a great addition to your home or office, or as a gift for a fellow Formula One enthusiast. Each poster is carefully packaged to ensure safe delivery, so you can enjoy your new piece of art as soon as possible. - A nice bit of text really highlighting a lot of key words such as gift, motorsports, racetrack etc.  You could go further with this too, by adding in extra things related to the poster such as ‘Perfect gift for a Mercedes F1 fan’ etc.  Tags Now, these are actually probably the most important part of your listing! You get 13 tags (20 character limit for each) and there are essentially search terms that will match your listing with what customers search for when shopping.  You really need to maximize these - whilst Title and Description play a part, these are the main things that will bring buyers to your listing. Once again, it is important to think about what customers are likely to be searching when looking for a poster similar to yours. Life hack alert! You can actually see what tags other sellers are using. All you need to do is go to a listing similar to yours that is selling well, scroll down and you can actually see them listed out at the bottom of the page! Here is an example of what this may look like: So, go through a few listings of competitors and make notes on common denominators that you can integrate into your listing. As you can see here, this seller uses tags such as ‘Birthday Gift’ and ‘Poster Print’. When you first start out, you may be better off swapping these out for more listing specific tags. This seller has been on Etsy for a few years however and has 15,000+ sales, so are more likely to see success from these tags.  If it’s not clear why, think about it this way. If you searched ‘poster print’ on Etsy today, there will be 10s of thousands of results. However, if you searched ‘Russell Mercedes Poster’, you will (as of writing) get 336 results. Etsy is far more likely to push your product to the top of the latter tag, against 300 other listings, rather than the top of ‘Poster Print’ where it is incredibly competitive. It is only when you are a more successful shop pulling in a high quantity of orders that these larger and more generic tags will work for you, as Etsy has more trust in your shop and will be more likely to push you to the front.  SKUs \-One important thing you need to do is add SKUs to all of your products! This is worth doing at the start as it will make your life so much easier when it comes to making sales and using PrintShrimp further down the line. What is an SKU? It is a ‘stock keeping unit’, and is essentially just a product identifier. Your SKUs need to match your file name that you upload to PrintShrimp. For example, if you made a poster about the eiffel tower, you can literally name the SKU eiffel-tower. There is no need to complicate things! As long as your file name (as in the image name of your poster on your computer) matches your SKU, you will be good to go.  \-It may be more beneficial to set up a system with unique identifiers, to make organising your files a lot easier further down the line. Say you get to 1000 posters eventually, you’ll want to be able to quickly search a code, and also ensure every SKU is always unique, so you won’t run into accidentally using the same SKU twice further down the line. For example, you can set it up so at the start of each file name, you have \[unique id\]\[info\], so your files will look like -  A1eiffeltower A2france And further down the line: A99aperolspritz B1potatoart This not only removes the potential issue of duplicating SKUs accidentally (for example if you made a few posters of the same subject), but also keeps your files well organised. If you need to find a file, you can search your files according to the code, so just by searching ‘a1’ for example, rather than having to trawl through a load of different files until you find the correct one. \-If your poster has variations, for example color variations, you can set a different SKU for each variation. Just click the little box when setting up variations that says ‘SKUs vary for each (variation)’. So if you have a poster available either in a white or black background, you can name each file, and therefore each SKU, a1eiffel-tower-black and a1eiffel-tower-white for example. \-The same goes for different sizes. As different American sizes have different aspect ratios, as mentioned above you may have to reformat some posters if you get a sale for one of these sizes. You can then add in the SKU to your listing once you have reformatted your poster. So for example if you sell a 16x20” version of the eiffel tower poster, you can name this file eiffel-tower-white-1620. Whilst this involves a little bit of set up, the time it saves you overall is massive!  Variations and Prices \-So, when selling posters there is a huge variety of sizes that you can offer, as mentioned previously. Non-negotiable is that you should be offering A5-A1. These will likely be your main sellers! Especially in the UK. It is also a good idea to offer inch sizing to appeal to a global audience (as bear in mind with PrintShrimp you will be able to print in multiple countries around the world!).  Below is a recommended pricing structure of what to charge on Etsy. Feel free to mess around with these! You may notice on Etsy that many shops charge a whole lot more for sizes such as A1, 24x36” etc. In my experience I prefer charging a lower rate to attract more sales, but there is validity in going for a lower amount of sales with higher profits. As mentioned above, you can also offer different variations on items - for example different colour schemes on posters. This is always a decent idea (if it suits the design) as it provides the customer with more options, which might help to convert the sale. You can always add this in later however if you want to keep it simple while you start! Setting up shipping profiles Etsy makes it very easy to set up different shipping rates for different countries. However, luckily with PrintShrimp you can offer free shipping to the majority of the major countries that are active on Etsy!  Using PrintShrimp means that your production costs are low enough in each domestic market to justify this. If you look on Etsy you can see there are many shops that post internationally to countries such as the US or Australia. Therefore, they often charge £8-10 in postage, and have a delivery time of 1-2 weeks. This really limits their customer base to their domestic market.  Using PrintShrimp avoids this and means you can offer free shipping (as we absorb the shipping cost in our prices) to the major markets of the UK, Australia, and USA (Europe coming soon!).  We also offer a 1 day processing time, unlike many POD poster suppliers. This means you can set your Etsy processing time to just one day, which combined with our quick shipping, means you will be one of the quickest on Etsy at sending out orders. This is obviously very attractive for customers, who are often very impatient with wanting their orders!  Getting the sales and extra tips \-Don’t list an insane amount of listings when you first get started. Etsy will be like ‘hang on a second’ if a brand new shop suddenly has 200 items in the first week. Warm up your account, and take things slow as you get going. We recommend 5 a day for the first week or so, and then you can start uploading more. You don’t want Etsy to flag your account for suspicious bot-like activity when you first get going.  \-It is very easy to copy listings when creating a new one. Simply select an old listing and press copy, and then you can just change the listing specific details to create a new one, rather than having to start from scratch. It can feel like a bit of a ball-ache setting up your first ever listing, but from then on you can just copy it over and just change the specifics.  \-Try and organize your listings into sections! This really helps the customer journey. Sometimes a customer will click onto your shop after seeing one of your listings, so it really helps if they can easily navigate your shop for what they are looking for. So, you now have a fully fledged Etsy shop. Well done! Time to start making £3,000 a month straight away right? Not quite. Please bear in mind, patience is key when starting out. If you started doing this because you are £10,000 in debt to the Albanian mafia and need to pay it off next week, you have come into this in the wrong frame of mind. If you have however started this to slowly build up a side hustle which hopefully one day become your full time gig, then winner winner chicken dinner.  Starting out on Etsy isn’t always easy. It takes time for your shop to build up trust! As I’ve said before, a buyer is far more likely to purchase from a shop with 1000s of reviews, than a brand new one with 0. But before you know it, you can become one of these shops! One thing you can do at the very start is to encourage your friends and family to buy your posters! This is a slightly naughty way of getting a few sales at the start, of course followed by a few glowing 5\* reviews. It really helps to give your shop this little boost at the start, so if this is something you can do then I recommend it.  Okay, so once you have a fully fledged shop with a decent amount of listings, you might be expecting the sales to start rolling in. And, if you are lucky, they indeed might. However, in my experience, you need to give your listings a little boost. So let us introduce you to: The wonderful world of Etsy ads Ads!! Oh no, that means money!! We imagine some of you more risk averse people are saying to yourself right now. And yes, it indeed does. But more often than not unfortunately you do have to spend money to make money.  Fortunately, in my experience anyway, Etsy ads do tend to work. This does however only apply if your products are actually good however, so if you’re back here after paying for ads for 2 months and are losing money at the same rate as your motivation, maybe go back to the start of this guide and pick another niche.  When you first start out, there are two main strategies.  Number 1: The Safer Option So, with PrintShrimp, you will essentially be making a minimum of £6 profit per order. With this in mind, I normally start a new shop with a safer strategy of advertising my products with a budget of $3-5 dollars a day. This then means that at the start, you only need to make 1 sale to break even, and anything above that is pure profit! This might not seem like the most dazzling proposition right now, but again please bear in mind that growth will be slow at the start. This means that you can gradually grow your shop, and therefore the trust that customers have in your shop, over time with a very small risk of ever actually losing money. Number 2: The Billy Big Balls Option If you were yawning while reading the first option, then this strategy may be for you. This will be better suited to those of you that are a bit more risk prone, and it also helps if you have a bit more cash to invest at the start. Through this strategy, you can essentially pay your way to the top of Etsy's rankings. For this, you’ll probably be looking at spending $20 a day on ads. So, this can really add up quickly and is definitely the riskier option. In my experience, the level of sales with this may not always match up to your spend every day. You may find that some days you rake in about 10 sales, and other days only one. But what this does mean is that as your listings get seen and purchased more, they will begin to rank higher in Etsy’s organic search rankings, at a much quicker rate than option one. This is the beauty of Etsy’s ads. You can pay to boost your products, but then results from this paid promotion feed into the organic ranking of your products. So you may find that you can splash the cash for a while at the start in order to race to the top, and then drop your ad spending later on when your products are already ranking well.  Sending your poster orders So, you’ve now done the hard bit. You have a running Etsy store, and essentially all you need to now on a daily basis is send out your orders and reply to customer messages! This is where it really becomes passive income.  \-Check out the PrintShrimp order portal. Simply sign up, and you can place individual orders through there. \-Bulk upload: We have an option to bulk upload your Esty orders via csv.  Seriously, when you are up and running with your first store, it is really as easy as that.  Once you have your first Etsy store up and running, you can think about expanding. There are many ways to expand your income. You can set up other Etsy stores, as long as the type of posters you are selling varies. You can look into setting up your own Shopify stores, and advertise them through Facebook, Instagram etc. Through this guide, we will teach you everything you need to know about starting to sell posters and generate some income. We will also show you why PrintShrimp is the best POD supplier for all of your poster needs. Trust me, you won’t need much convincing.

Ai C-Level team
reddit
LLM Vibe Score0
Human Vibe Score1
thestoicdesignerThis week

Ai C-Level team

I've been exploring ways to run a company where I'm essentially the only internal team member, relying entirely on a suite of specialized AIs for executive roles, supported occasionally by external consultants for niche expertise. My goal is to stay lean, agile, and highly creative, especially in a fashion / tech brand context. Essentially, I'm building an AI-driven C-Level team, or what I like to call a "C-Level AI Wallet." Here's what I'm thinking for the key executive roles I'd need to cover with AI: CEO AI – Responsible for overall strategy, decision-making, trend analysis, and guiding the company's vision. I'd probably lean on something advanced like Gemini, GPT-4, or similar models, fine-tuned with market-specific data. COO AI (Operations): I'd need tools that streamline and automate logistics, supply chain management, and day-to-day operations (think something along the lines of Zapier AI integrations or Make). CMO AI (Marketing & Content): For branding, content creation, digital marketing, and consumer insights, I'd use Jasper or Copy . ai, combined with predictive analytics tools like Google Vertex AI to understand trends better. Additionally, for generating engaging visual and multimedia content, tools like Midjourney, DALL·E, Adobe Firefly, and Runway ML would be perfect. CFO AI (Financial Management): For financial management, cash flow control, and investment decisions, I'd probably leverage AI tools like Bloomberg GPT, combined with AI-powered forecasting platforms. CHRO AI (Human Resources & Culture): Although the internal team is minimal (just myself!), I'd still rely on AI for tasks like project management, freelancer hiring, and performance tracking—tools like HireVue AI, Motion, or even Notion's AI could be beneficial here. CSO AI (Sustainability & Compliance): Since sustainability and ethical sourcing are critical, I'd integrate ESG-focused AI tools to ensure transparency and responsible sourcing. My idea is that, with the right AI tools seamlessly integrated, I can manage the strategic vision and creative direction personally, leveraging external consultants only when necessary. This setup would ideally allow me to operate as a one-person internal team supported by a robust "wallet" of AI executives. Has anyone tried a similar approach? What AI tools would you recommend for a truly lean, innovative brand structure? I'm very curious about your experiences or suggestions—let me know your thoughts!

This is why most of AI wrappers will die
reddit
LLM Vibe Score0
Human Vibe Score1
ecommerce_itThis week

This is why most of AI wrappers will die

We began building our AI product in public as a tool to help people quickly build online stores using AI during June of 2023. It was quite a hot AI time. The tool was using ChatGPT to create a fully-functinal eCommerce store with a demo products from Amazon. And we managed to get such impression among people so they started to share it with words: "Look, I made my own store in 20 seconds." We got about 2,000 users that way, mainly people telling their friends to try it out. We built a toy Back in 2023, this idea was exciting. It was great for getting people to talk about us and for getting random people to check us out. We burned \~2k$ on various API we used then with an expectations: people will start to pay. Nobody paid. It was a train called AI and we all were the passengers, but not all of us were able to understand how to monitize this and in reality most of AI wrappers have the lack of this. Most of AI wrappers would be eaten by a bigger players, other will be not able to proceed due to fact of investment. We had a few benefits: 1) We are developers with skills in design and a bit in marketing 2) We spent years in development of eCommerce products So to keep things going it was important to focus on: 1) Longer game, there is no quick wins, unfortunatelly or fortunatelly 2) Narrower niche and smaller auditory 3) Patience 4) Building network and product authority The road to actual product So to attract real users, we had to start solving a real problem for them, to offer them something valuable. We do this already 5 months since October. We made like 5 pivots... Today our product proposition "Marketsy allows busy people to own a business: a simple in management store of digital products as a source of income" So all AI thing right now is hidden under "busy", AI helps to automate the process, but not the primary thing in the product anymore. Even eCommerce SaaS market is huge and comeptition is hight. We are going to test this approach upcoming weeks, we believe it will be a right step. Anyway we are sure we will find the right proposition and our audience, one way or another. All the best to other product builders here!

[CASE STUDY] From 217/m to $2,836/m in 9 months - Sold for $59,000; I grow and monetise web traffic of 5, 6, 7 figures USD valued passive income content sites [AMA]
reddit
LLM Vibe Score0
Human Vibe Score1
jamesackerman1234This week

[CASE STUDY] From 217/m to $2,836/m in 9 months - Sold for $59,000; I grow and monetise web traffic of 5, 6, 7 figures USD valued passive income content sites [AMA]

Hello Everyone (VERY LONG CASE STUDY AHEAD) - 355% return in 9 months Note: I own a 7-figures USD valued portfolio of 41+ content sites that generates 5-6 figures USD a month in passive income. This is my first time posting in this sub and my goal is to NOT share generic advice but precise numbers, data and highly refined processes so you can also get started with this business yourself or if you already have an existing business, drive huge traffic to it and scale it substantially (get more customers). I will use a case study to explain the whole process. As most of us are entrepreneurs here, explaining an actual project would be more meaningful. In this case study I used AI assisted content to grow an existing site from $217/m to $2,836/m in 9 months (NO BACKLINKS) and sold it for $59,000. ROI of 3 months: 355% Previous case studies (before I give an overview of the model) Amazon Affiliate Content Site: $371/m to $19,263/m in 14 MONTHS - $900K CASE STUDY \[AMA\] Affiliate Website from $267/m to $21,853/m in 19 months (CASE STUDY - Amazon?) \[AMA\] Amazon Affiliate Website from $0 to $7,786/month in 11 months Amazon Affiliate Site from $118/m to $3,103/m in 8 MONTHS (SOLD it for $62,000+) Note: You can check pinned posts on my profile. Do go through the comments as well as a lot of questions are answered in those. However, if you still have any questions, feel free to reach out. This is an \[AMA\]. Quick Overview of the Model Approach: High traffic, niche specific, informative content websites that monetise its traffic through highly automated methods like display ads and affiliate. The same model can be applied to existing businesses to drive traffic and get customers. Main idea: Make passive income in a highly automated way Easy to understand analogy You have real estate (here you have digital asset like a website) You get rental income (here you get ads and affiliate income with no physical hassle, in case you have a business like service, product etc. then you can get customers for that too but if not, it's alright) Real estate has value (this digital asset also has value that can be appreciated with less effort) Real estate can be sold (this can be sold too but faster) IMPORTANT NOTE: Search traffic is the BEST way to reach HUGE target audience and it's important when it comes to scaling. This essentially means that you can either monetise that via affiliate, display etc. or if you have a business then you can reach a bigger audience to scale. Overview of this website's valuation (then and now: Oct. 2022 and June 2023) Oct 2022: $217/m Valuation: $5,750.5 (26.5x) - set it the same as the multiple it was sold for June 2023: $2,836/m Traffic and revenue trend: growing fast Last 3 months avg: $2,223 Valuation now: $59,000 (26.5x) Description: The domain was registered in 2016, it grew and then the project was left unattended. I decided to grow it again using properly planned AI assisted content. Backlink profile: 500+ Referring domains (Ahrefs). Backlinks mean the sites linking back to you. This is important when it comes to ranking. Summary of Results of This Website - Before and After Note: If the terms seem technical, do not worry. I will explain them in detail later. Still if you have any questions. Feel free to comment or reach out. |Metric|Oct 2022|June 2023|Difference|Comments| |:-|:-|:-|:-|:-| |Articles|314|804|\+490|AI Assisted content published in 3 months| |Traffic|9,394|31,972|\+22,578|Organic| |Revenue|$217|$2,836|\+$2,619|Multiple sources| |RPM (revenue/1000 web traffic)|23.09|$88.7|\+$65.61|Result of Conversion rate optimisation (CRO). You make changes to the site for better conversions| |EEAT (expertise, experience, authority and trust of website)|2 main authors|8 authors|6|Tables, video ads and 11 other fixations| |CRO|Nothing|Tables, video ads |Tables, video ads and 11 other fixations || &#x200B; Month by Month Growth |Month|Revenue|Steps| |:-|:-|:-| |Sept 2022|NA|Content Plan| |Oct 2022|$217|Content Production| |Nov 2022|$243|Content production + EEAT authors| |Dec 2022|$320|Content production + EEAT authors| |Jan 2023|$400|Monitoring| |Feb 2023|$223|Content production + EEAT authors| |Mar 2023|$2,128|CRO & Fixations| |April 2023|$1,609|CRO & Fixations| |May 2023|$2,223|Content production + EEAT authors| |June 2023|$2,836|CRO and Fixations| |Total|$10,199|| &#x200B; What will I share Content plan and Website structure Content Writing Content Uploading, formatting and onsite SEO Faster indexing Conversion rate optimisation Guest Posting EEAT (Experience, Expertise, Authority, Trust) Costing ROI The plans moving forward with these sites &#x200B; Website Structure and Content Plan This is probably the most important important part of the whole process. The team spends around a month just to get this right. It's like defining the direction of the project. Description: Complete blueprint of the site's structure in terms of organisation of categories, subcategories and sorting of articles in each one of them. It also includes the essential pages. The sorted articles target main keyword, relevant entities and similar keywords. This has to be highly data driven and we look at over 100 variables just to get it right. It's like beating Google's algorithm to ensure you have a blueprint for a site that will rank. It needs to be done right. If there is a mistake, then even if you do everything right - it's not going to work out and after 8-16 months you will realise that everything went to waste. Process For this project, we had a niche selected already so we didn't need to do a lot of research pertaining to that. We also knew the topic since the website was already getting good traffic on that. We just validated from Ahrefs, SEMRUSH and manual analysis if it would be worth it to move forward with that topic. &#x200B; Find entities related to the topic: We used Ahrefs and InLinks to get an idea about the related entities (topics) to create a proper topical relevance. In order to be certain and have a better idea, we used ChatGPT to find relevant entities as well \> Ahrefs (tool): Enter main keyword in keywords explorer. Check the left pain for popular topics \> Inlinks (tool): Enter the main keyword, check the entity maps \> ChatGPT (tool): Ask it to list down the most important and relevant entities in order of their priority Based on this info, you can map out the most relevant topics that are semantically associated to your main topic Sorting the entities in topics (categories) and subtopics (subcategories): Based on the information above, cluster them properly. The most relevant ones must be grouped together. Each group must be sorted into its relevant category. \> Example: Site about cycling. \> Categories/entities: bicycles, gear and equipment, techniques, safety, routes etc. \> The subcategories/subentities for let's say "techniques" would be: Bike handling, pedaling, drafting etc. Extract keywords for each subcategory/subentity: You can do this using Ahrefs or Semrush. Each keyword would be an article. Ensure that you target the similar keywords in one article. For example: how to ride a bicycle and how can I ride a bicycle will be targeted by one article. Make the more important keyword in terms of volume and difficulty as the main keyword and the other one(s) as secondary Define main focus vs secondary focus: Out of all these categories/entities - there will be one that you would want to dominate in every way. So, focus on just that in the start. This will be your main focus. Try to answer ALL the questions pertaining to that. You can extract the questions using Ahrefs. \> Ahrefs > keywords explorer \> enter keyword \> Questions \> Download the list and cluster the similar ones. This will populate your main focus category/entity and will drive most of the traffic. Now, you need to write in other categories/subentities as well. This is not just important, but crucial to complete the topical map loop. In simple words, if you do this Google sees you as a comprehensive source on the topic - otherwise, it ignores you and you don't get ranked Define the URLs End result: List of all the entities and sub-entities about the main site topic in the form of categories and subcategories respectively. A complete list of ALL the questions about the main focus and at around 10 questions for each one of the subcategories/subentities that are the secondary focus Content Writing So, now that there's a plan. Content needs to be produced. Pick out a keyword (which is going to be a question) and... Answer the question Write about 5 relevant entities Answer 10 relevant questions Write a conclusion Keep the format the same for all the articles. Content Uploading, formatting and onsite SEO Ensure the following is taken care of: H1 Permalink H2s H3s Lists Tables Meta description Socials description Featured image 2 images in text \\Schema Relevant YouTube video (if there is) Note: There are other pointers link internal linking in a semantically relevant way but this should be good to start with. Faster Indexing Indexing means Google has read your page. Ranking only after this step has been done. Otherwise, you can't rank if Google hasn't read the page. Naturally, this is a slow process. But, we expedite it in multiple ways. You can use RankMath to quickly index the content. Since, there are a lot of bulk pages you need a reliable method. Now, this method isn't perfect. But, it's better than most. Use Google Indexing API and developers tools to get indexed. Rank Math plugin is used. I don't want to bore you and write the process here. But, a simple Google search can help you set everything up. Additionally, whenever you post something - there will be an option to INDEX NOW. Just press that and it would be indexed quite fast. Conversion rate optimisation Once you get traffic, try adding tables right after the introduction of an article. These tables would feature a relevant product on Amazon. This step alone increased our earnings significantly. Even though the content is informational and NOT review. This still worked like a charm. Try checking out the top pages every single day in Google analytics and add the table to each one of them. Moreover, we used EZOIC video ads as well. That increased the RPM significantly as well. Both of these steps are highly recommended. Overall, we implemented over 11 fixations but these two contribute the most towards increasing the RPM so I would suggest you stick to these two in the start. Guest Posting We made additional income by selling links on the site as well. However, we were VERY careful about who we offered a backlink to. We didn't entertain any objectionable links. Moreover, we didn't actively reach out to anyone. We had a professional email clearly stated on the website and a particularly designated page for "editorial guidelines" A lot of people reached out to us because of that. As a matter of fact, the guy who bought the website is in the link selling business and plans to use the site primarily for selling links. According to him, he can easily make $4000+ from that alone. Just by replying to the prospects who reached out to us. We didn't allow a lot of people to be published on the site due to strict quality control. However, the new owner is willing to be lenient and cash it out. EEAT (Experience, Expertise, Authority, Trust) This is an important ranking factor. You need to prove on the site that your site has authors that are experienced, have expertise, authority and trust. A lot of people were reaching out to publish on our site and among them were a few established authors as well. We let them publish on our site for free, added them on our official team, connected their socials and shared them on all our socials. In return, we wanted them to write 3 articles each for us and share everything on all the social profiles. You can refer to the tables I shared above to check out the months it was implemented. We added a total of 6 writers (credible authors). Their articles were featured on the homepage and so were their profiles. Costing Well, we already had the site and the backlinks on it. Referring domains (backlinks) were already 500+. We just needed to focus on smart content and content. Here is the summary of the costs involved. Articles: 490 Avg word count per article: 1500 Total words: 735,000 (approximately) Cost per word: 2 cents (includes research, entities, production, quality assurance, uploading, formatting, adding images, featured image, alt texts, onsite SEO, publishing/scheduling etc.) Total: $14,700 ROI (Return on investment) Earning: Oct 22 - June 23 Earnings: $10,199 Sold for: $59,000 Total: $69,199 Expenses: Content: $14,700 Misc (hosting and others): $500 Total: $15,200 ROI over a 9 months period: 355.25% The plans moving forward This website was a part of a research and development experiment we did. With AI, we wanted to test new waters and transition more towards automation. Ideally, we want to use ChatGPT or some other API to produce these articles and bulk publish on the site. The costs with this approach are going to be much lower and the ROI is much more impressive. It's not the the 7-figures projects I created earlier (as you may have checked the older case studies on my profile), but it's highly scalable. We plan to refine this model even further, test more and automate everything completely to bring down our costs significantly. Once we have a model, we are going to scale it to 100s of sites. The process of my existing 7-figures websites portfolio was quite similar. I tested out a few sites, refined the model and scaled it to over 41 sites. Now, the fundamentals are the same however, we are using AI in a smarter way to do the same but at a lower cost, with a smaller team and much better returns. The best thing in my opinion is to run numerous experiments now. Our experimentation was slowed down a lot in the past since we couldn't write using AI but now it's much faster. The costs are 3-6 times lower so when it used to take $50-100k to start, grow and sell a site. Now you can pump 3-6 more sites for the same budget. This is a good news for existing business owners as well who want to grow their brand. Anyway, I am excited to see the results of more sites. In the meantime, if you have any questions - feel free to let me know. Best of luck for everything. Feel free to ask questions. I'd be happy to help. This is an AMA.

100 best ai sustainable business ideas in 2025
reddit
LLM Vibe Score0
Human Vibe Score1
Low_Philosopher1792This week

100 best ai sustainable business ideas in 2025

AI in Renewable Energy AI-powered smart solar panel optimization Predictive maintenance for wind turbines AI-driven energy storage management AI-based microgrid optimization Smart grid energy forecasting AI-powered water desalination efficiency AI-driven carbon footprint reduction software AI-powered hydropower efficiency monitoring AI for geothermal energy exploration AI-driven green hydrogen production optimization AI in Waste Management & Recycling AI-based waste sorting robots Smart recycling bins with AI recognition AI-powered food waste management AI-driven upcycling marketplace AI-enabled e-waste management solutions AI-powered sustainable packaging optimization AI-driven landfill management systems AI-powered plastic waste tracking and reduction AI-based waste-to-energy conversion AI-driven composting automation AI in Water Conservation AI-powered leak detection and water conservation AI-driven smart irrigation systems AI-based flood prediction and mitigation AI-powered ocean plastic cleanup robots AI-driven rainwater harvesting optimization AI-based groundwater level monitoring AI-powered desalination energy efficiency AI-driven smart water meters AI-powered wastewater treatment optimization AI-based water pollution monitoring AI in Sustainable Agriculture AI-driven precision farming AI-powered vertical farming automation AI-based pest and disease prediction AI-powered livestock health monitoring AI-driven soil health analysis AI-powered regenerative agriculture analytics AI-driven smart greenhouses AI-powered crop rotation optimization AI-based carbon farming solutions AI-powered sustainable aquaculture AI in Transportation & Mobility AI-powered electric vehicle (EV) battery optimization AI-driven smart traffic management AI-powered EV charging station optimization AI-based sustainable urban mobility planning AI-powered drone delivery for carbon reduction AI-driven logistics and supply chain sustainability AI-powered smart public transport systems AI-driven sustainable aviation fuel optimization AI-powered bicycle-sharing optimization AI-driven AI carpooling and ride-sharing efficiency AI in Green Manufacturing AI-powered energy-efficient manufacturing AI-driven supply chain sustainability analytics AI-based material waste reduction AI-powered sustainable fashion production AI-driven predictive demand to reduce overproduction AI-powered eco-friendly textile manufacturing AI-driven 3D printing for sustainable manufacturing AI-powered emission reduction in factories AI-driven green construction material optimization AI-based lifecycle assessment for eco-products AI in Carbon Offsetting & Climate Action AI-powered carbon credit marketplaces AI-driven tree planting optimization AI-based carbon capture efficiency enhancement AI-powered reforestation tracking and monitoring AI-driven climate risk prediction AI-powered environmental compliance software AI-driven sustainable investment analysis AI-based corporate sustainability tracking AI-powered carbon accounting and reporting AI-driven decarbonization roadmaps for businesses AI in Sustainable Smart Cities AI-powered urban energy efficiency monitoring AI-driven AI-powered smart lighting for cities AI-based pollution monitoring and reduction AI-driven green building automation AI-powered smart HVAC energy optimization AI-driven urban tree canopy management AI-powered digital twins for sustainable city planning AI-based urban noise pollution monitoring AI-powered public waste management optimization AI-driven citizen engagement for sustainability AI in Eco-Friendly Consumer Solutions AI-powered sustainable shopping assistant AI-driven personal carbon footprint tracking app AI-powered second-hand marketplace optimization AI-driven sustainable food delivery services AI-powered ethical supply chain transparency AI-driven zero-waste grocery stores AI-powered green subscription services AI-driven sustainable tourism planning AI-powered smart home energy efficiency optimization AI-driven personal finance for sustainability investments AI in Sustainable Healthcare & Well-being AI-powered climate impact on health analytics AI-driven sustainable hospital management AI-based predictive disease outbreak prevention AI-powered mental health solutions for eco-anxiety AI-driven green pharmaceutical production AI-powered sustainable medical waste management AI-based air quality health impact monitoring AI-driven climate-friendly diet and nutrition planning AI-powered fitness and well-being optimization for sustainability AI-driven telemedicine to reduce healthcare emissions These AI-driven sustainable business ideas offer high growth potential while making a positive impact on the planet. Let me know if you want details on a specific idea or need help with implementation strategies!

Why the value of writing code and other digital services is going to zero
reddit
LLM Vibe Score0
Human Vibe Score1
BalloonWheelieThis week

Why the value of writing code and other digital services is going to zero

I must preface this with a trigger warning because I make some statements in this post that might be upsetting to some. This post discusses my experience building in the new era of entrepreneurship, which is one where the founder is the center of the universe, and the consultants, overpriced SaaS, and corporate swamp creatures are replaced by single-user custom software, bots, and self-hosted automations. If you work in the legacy economy, I really don't intend to stress you out or say things you are doing are quickly becoming irrelevant, but I must share the reality of how I am operating, because I would like to hear from others who are doing the same, or desire to do the same. I am currently operating with the belief that AI-powered tools are going to make 1-person million dollar businesses much more common. Building anything digital is becoming extremely easy, cheap, and quick to implement. The value of code and digital tools is approaching zero, or at most 5% of what it currently is. Right now, the most powerful AI tools are aimed at developers, so folks who have some technical and business ability basically have nothing holding them back aside from the speed of their brain right now. I happen to be a part of the cohort, and am building like there is no tomorrow, but I don't believe this cohort is actually all that big. The next hurdle to unlock the new era of entrepreneurship is empowering every entrepreneur to build at the same pace that is currently locked behind having technical ability. This cohort is huge (millions, if the number of people in this sub is any indication). This post is aimed at them (you?). If you are part of this cohort, what is holding you back from launching a new product for near-zero cost? What is too complicated, too expensive, too unknown for you to be able to build your new/current business at maximum speed? I look forward to seeing the replies, I hope some insights shared can help the community, and be a catalyst for more tools to enable non-technical founders to launch. I will now share some of how I am testing, launching, and selling as a one-man-show. This will be a little bit technical, but if the output of any layer of my stack is something you want, please comment because maybe someone will build a cheap way of accessing it without needing to manage the code yourself. \#1 BOTS I cannot overstate how much leverage bots have created for me. I run all of my bots locally and interface with with via Telegram. Bots do things like: \- watch social media pages, forums, subreddits, etc related to my customers and notify me of what is going on, and suggest SEO blog posts that could be published to capture traffic related to the topic. with a single message, my bot will generate a blog post, send it to me for review, apply edits i suggest, and then publish it live, all from within telegram \- pay attention to all my key metrics/analytics, and attempt to find insights/corrolations (ex. there is a lot of traffic on this page, blog post, video, etc. here's why, and how we can take advantage of it to drive business goals) \- repurposing content. i have dozens of social media profiles that are 100% run by bots, they are all related to my customer niches and will do things like post news, snippets from my blogs, interact with human creators in the niche, etc. this builds my audience automatically which I can then advertise to/try to convert into paying customers, since they are interested in the things my bot is posting and become followers, it's like automated qualified lead gen 24/7 across every social platform and every niche I care about. you may be thinking by now that this post is made by a bot, but you will have to trust me that this is 100% hand-written by my sleep-deprived brain. let's continue: \#2 replacing every SaaS with a shitty version of it designed for what i need out of it it's absurd that we pay ten's of dollars per seat per month for basic digital functions like chat (slack), CRM (active camppaign, sales force, hubspot, etc), email stuff (mailchip, etc), link sharing (linktree, etc), website builders (wix, squarespace, etc), etc. all of these SaaS tools are overpriced and overbuilt. I believe many of them are going to be caught in the innovators dilemma and will go to 0. I don't use any of these anymore, I build and self-host my own shitty version of each of them that does only what i need out of the tool. for example, my CRM doesn't have a fancy drag and drop email builder and 10000 3rd party plugins, because i dont need any of that shit I just need to segment and communicate with my customers. if i need more features, i can generate them on the fly. \#3 working alone I have worked with cofounders in the past, raised money from investors, hired consultants, burned money and time, suffered sleepless nights from stress caused by other people not delivering, trying to convince others they are wrong, or they are pushing the company off a cliff, waste waste waste. no more of that. In the new age of entrepreneurship, the BUILDER (you and I) are the ones creating the value, and AI empowers us to do it alone. this might seem daunting, but there is no business problem that can't be solved with a detailed discussion sesh with chatgpt, no facts that can't be found with perplexity, and no task that can't be automated with claude. there is no need for anymore swamp creatures. you are the start and the end point, you don't need to rely on anyone else for anything. this may sound ignorant, but this is the conclusion I have come to believe, and it continues to be proven every day my businesses progress with me being the only human involved. This is getting quite long so I'll cut it here. I look forward to hearing about how you are operating in this new era and hopefully getting inspired/learning some new ideas to add to my current stack.

AI Content Campaign Got 4M impressions, Thousands of Website Views, Hundreds of Customers for About $100 — This is the future of marketing
reddit
LLM Vibe Score0
Human Vibe Score0.857
adamkstinsonThis week

AI Content Campaign Got 4M impressions, Thousands of Website Views, Hundreds of Customers for About $100 — This is the future of marketing

Alright. So, a few months ago I tested a marketing strategy for a client that I’ve sense dedicated my life to developing on. The Idea was to take the clients Pillar content (their YouTube videos) and use AI to rewrite the content for all the viable earned media channels (mainly Reddit). The campaign itself was moderately successful. To be specific, after one month it became their 2nd cheapest customer acquisition cost (behind their organic YouTube content). But there is a lot to be done to improve the concept. I will say, having been in growth marketing for a decade, I felt like I had hit something big with the concept. I’m going to detail how I built that AI system, and what worked well and what didn’t here. Hopefully you guys will let me know what you think and whether or not there is something here to keep working on. DEFINING THE GOAL Like any good startup, their marketing budget was minimal. They wanted to see results, fast and cheap. Usually, marketers like me hate to be in this situation because getting results usually either takes time or it takes money. But you can get results fast and cheap if you focus on an earned media strategy - basically getting featured in other people’s publication. The thing is these strategies are pretty hard to scale or grow over time. That was a problem for future me though. I looked through their analytics and saw they were getting referral traffic from Reddit - it was their 5th or 6th largest source of traffic - and they weren’t doing any marketing on the platform. It was all digital word of mouth there. It kind of clicked for me there, that Reddit might be the place to start laying the ground work. So with these considerations in mind the goal became pretty clear: Create content for relevant niche communities on Reddit with the intent of essentially increasing brand awareness. Use an AI system to repurpose their YouTube videos to keep the cost of producing unique content for each subreddit really low. THE HIGH-LEVEL STRATEGY I knew that there are huge amounts of potential customers on Reddit (About 12M people in all the relevant communities combined) AND that most marketers have a really tough time with the platform. I also knew that any earned media strategy, Reddit or not, means Click Through Rates on our content would be extremely low. A lot of people see this as a Reddit specific problem because you can’t self-promote on the platform, but really you have to keep self-promotion to a minimum with any and all earned media. This basically meant we had to get a lot of impressions to make up for it. The thing about Reddit is if your post absolutely crushes it, it can get millions of views. But crushing it is very specific to what the expectations are of that particular subreddit. So we needed to make content that was specifically written for that Subreddit. With that I was able to essentially design how this campaign would work: We would put together a list of channels (specifically subreddits to start) that we wanted to create content for. For each channel, we would write a content guideline that details out how to write great content for this subreddit. These assets would be stored in an AirTable base, along with the transcripts of the YouTube videos that were the base of our content. We would write and optimize different AI Prompts that generated different kinds of posts (discussion starters about a stock, 4-5 paragraph stock analysis, Stock update and what it means, etc…) We would build an automation that took the YouTube transcripts, ran each prompt on it, and then edited each result to match the channel writing guidelines. And then we would find a very contextual way to leave a breadcrumb back to the client. Always as part of the story of the content. At least, this is how I originally thought things would go. CHOOSING THE RIGHT SUBREDDITS Picking the right communities was vital. Here’s the basic rubric we used to pick and prioritize them: • Relevance: We needed communities interested in stock analysis, personal finance, or investing. • Subreddit Size vs. Engagement: Large subreddits offer more potential impressions but can be less focused. Smaller subreddits often have higher engagement rates. • Content Feasibility: We had to ensure we could consistently create high-value posts for each chosen subreddit. We started with about 40 possibilities, then narrowed it down to four or five that consistently delivered upvotes and user signups. CREATING CHANNEL-SPECIFIC GUIDES By the end, creating channel specific writing guidelines looked like a genius decision. Here’s how we approached it and used AI to get it done quickly: Grabbed Top Posts: We filtered the subreddit’s top posts (change filter to “Top” and then “All Time”) of all time to see the kinds of content that performed best Compiled The Relevant Posts: We took the most relevant posts to what we were trying to do and put them all on one document (basically created one document per subreddit that just had the top 10 posts in that subreddit). Had AI Create Writing Guideline Based On Posts: For each channel, we fed the document with the 10 posts with the instructions “Create a writing guideline for this subreddit based on these high performing posts. I had to do some editing on each guideline but this worked pretty well and saved a lot of time. Each subreddit got a custom guideline, and we put these inside the “Channels” table of the AirTable base we were developing with these assets. BUILDING THE AI PROMPTS THAT GENERATED CONTENT Alright this is probably the most important section so I’ll be detailed. Essentially, we took all the assets we developed up until this point, and used them to create unique posts for each channel. This mean each AI prompt was about 2,000 words of context and produced about a 500-word draft. There was a table in our AirTable where we stored the prompts, as I alluded to earlier. And these were basically the instructions for each prompt. More specifically, they detailed out our expectations for the post. In other words, there were different kinds of posts that performed well on each channel. For example, you can write a post that’s a list of resources (5 tools we used to…), or a how to guide (How we built…), etc.. Those weren’t the specific ones we used, but just wanted to really explain what I meant there. That actual automation that generated the content worked as follows: New source content (YouTube video transcript) was added to the Source Content table. This triggered the Automation. The automation grabbed all the prompts in the prompt table. For each prompt in the prompt table, we sent a prompt to OpenAI (gpt-4o) that contained first the prompt and also the source content. Then, for each channel that content prompt could be used on, we sent another prompt to OpenAI that revised the result of the first prompt based on the specific channel guidelines. The output of that prompt was added to the Content table in AirTable. To be clear, our AirTable had 4 tables: Content Channels Prompts Source Content The Source Content, Prompts, and Channel Guidelines were all used in the prompt that generated content. And the output was put in the Content table. Each time the automation ran, the Source Content was turned into about 20 unique posts, each one a specific post type generated for a specific channel. In other words, we were create a ton of content. EDITING & REFINING CONTENT The AI drafts were never perfect. Getting them Reddit-ready took editing and revising The main things I had to go in and edit for were: • Tone Adjustments: We removed excessively cliche language. The AI would say silly things like “Hello fellow redditors!” which sound stupid. • Fact-Checking: Financial data can be tricky. We discovered AI often confused figures, so we fact check all stock related metrics. Probably something like 30-40% error rate here. Because the draft generation was automated, that made the editing and getting publish ready the human bottleneck. In other words, after creating the system I spent basically all my time reviewing the content. There were small things I could do to make this more efficient, but not too much. The bigger the model we used, the less editing the content needed. THE “BREADCRUMB” PROMOTION STRATEGY No where in my prompt to the AI did I mention that we were doing any marketing. I just wanted the AI to focus on creating content that would do well on the channel. So in the editing process I had to find a way to promote the client. I called it a breadcrumb strategy once and that stuck. Basically, the idea was to never overtly promote anything. Instead find a way to leave a breadcrumb that leads back to the client, and let the really interested people follow the trail. Note: this is supposed to be how we do all content marketing. Some examples of how we did this were: Shared Visuals with a Subtle Watermark: Because our client’s product offered stock data, we’d often include a chart or graph showing a company’s financial metric with the client’s branding in the corner. Added Supporting Data from Client’s Website: If we mentioned something like a company’s cash flow statement, we could link to that company’s cash flow statement on the client’s website. It worked only because there was a lot of data on the client’s website that wasn’t gated. These tactics were really specific to the client. Which is should be. For other companies I would rethink what tactics I use here. THE RESULTS I’m pretty happy with the results • Impressions: – Early on posts averaged \~30,000 apiece, but after about a month of optimization, we hit \~70,000 impressions average. Over about two months, we reached 4 million total impressions. • Signups: – In their signups process there was one of those “Where did you find us?” questions and the amount of people who put Reddit jumped into the few hundred a month. Precise tracking of this is impossible. • Cost Efficiency (This is based on what I charged, and not the actual cost of running the campaign which is about $100/mo): – CPM (cost per thousand impressions) was about $0.08, which is far better than most paid channels. – Cost per free user: \~$8-10. After about a 10% conversion rate to a paid plan, our cost per paying user was $80–$100—well below the client’s previous $300–$400. HIGHLIGHTS: WHAT WORKED Subreddit-Specific Content: – Tailoring each post’s format and length to the audience norms boosted engagement. Worked out really well. 1 post got over 1M views alone. We regularly had posts that had hundreds of thousands. Breadcrumbs: – We never had anyone call us out for promoting. And really we weren’t. Our first priority was writing content that would crush on that subreddit. Using the Founder’s Existing Material: – The YouTube transcripts grounded the AI’s content in content we already made. This was really why we were able to produce so much content. CHALLENGES: WHAT DIDN’T WORK AI is still off: – Maybe it’s expecting too much, but still I wish the AI had done a better job. I editing a lot of content. Human oversight was critical. Scheduling all the content was a pain: – Recently I automated this pretty well. But at first I was scheduling everything manually and scheduling a hundred or so posts was a hassle. Getting Data and Analytics: – Not only did we have not very good traffic data, but the data from reddit had to be collected manually. Will probably automate this in the future. COST & TIME INVESTMENT Setup: The setup originally took me a couple weeks. I’ve since figured out how to do much faster (about 1 week). AirTable Setup here was easy and the tools costs $24/mo so not bad. ChatGPT costs were pretty cheap. Less than $75 per month. I’ve sense switched to using o1 which is much more expensive but saves me a lot of editing time Human Editing: Because this is the human part of the process and everything else was automated it mean by default all my time was spent editing content. Still this was a lot better than creating content from scratch probably by a factor of 5 or 10. The main expense was paying an editor (or using your own time) to refine posts. Worth it? Yes even with the editing time I was able to generate way more content that I would have otherwise. LESSONS & ACTIONABLE TAKEAWAYS Reddit as a Growth Channel: – If you genuinely respect each subreddit’s culture, you can achieve massive reach on a tight budget. AI + Human Collaboration: – AI excels at first drafts, but human expertise is non-negotiable for polishing and ensuring factual integrity. Soft Promotion Wins: – The “breadcrumb” approach paid off. It might feel like too light a touch, but is crucial for Reddit communities. Create once, repurpose as many times as possible: – If you have blog posts, videos, podcasts, or transcripts, feed them into AI to keep your message accurate and brand-consistent. CONCLUSION & NEXT STEPS If you try a similar approach: • Begin with smaller tests in a few niches to learn what resonates. • Create a clear “channel guide” for each community. • Carefully fact-check AI-generated posts. • Keep brand mentions low-key until you’ve established credibility.

I spent 18 hours every week tracking marketing trends and latest news. Here are my predictions for 2024
reddit
LLM Vibe Score0
Human Vibe Score0.778
lazymentorsThis week

I spent 18 hours every week tracking marketing trends and latest news. Here are my predictions for 2024

1/ Securing Digital Footprint becomes #1 Priority For Chronically Online Users, Protecting their digital footprint will become one of the main things. We saw influencers getting cancelled over Old Content and Brands used Old Travis Kelce Tweets, we saw what could happen without digital footprint protection. Online Engagement Precautions will be taken again with Twitter & IG showing your usernames above ‘Algorithm Suggested Content’. What you like is more visible to other people in UI Design of these apps, another reason behind why Digital Footprint preservation will matter a lot in 2024. This will impact likes to viewership ratio on your organic and paid content. &#x200B; 2/  TikTok wants Long Videos with Storytelling As I was writing this report, TikTok also released their What’s Next 2024 Report. It focuses heavily on how the audiences on the app demand better storytelling and from the examples in the report, you can judge what TikTok wants. They also rolled out a 30-minute video upload limit. Engaging Content over 1-Minute Mark to keep the audiences longer on the app. I highlighted in the first trend, every social media platform wants the same thing, more time spent. 3/ Use of Shop the Look While Streaming Netflix or Amazon Prime. This year’s one of the most successful TV series, The Bear caused Men to go mad for the T-Shirt worn by Jeremy Allen White in the show. Showing us how TV Shows influence or encourage us to dress in a particular way. It’s nothing new, TV Shows like Friends & Gossip Girl influenced all demographics when they came out. But now, Streamings Services such as Roku & Amazon enable consumers to shop the look while watching the TV Shows. Many Brands will jump on these opportunities in upcoming months. 4/ Brands in Comments & Memes are the new norm By Summer 2024, Most Online Users & Creators will no longer feel too excited or answered when they see your brand in the comments. Why? It’s becoming too common for Brands to show in comments under viral content about them. Or Brands being funny with Internet Culture Trends is known to most users. The Saturation of Every Brand being funny and being present leads to increased competition of levitating the content quality. &#x200B; 5/ Marketers decrease their focus on Traffic & Views With AI recommendations taking over, The Structure of content distributing on social media is changing, the same goes for SEO. Conversational AIs are changing how web traffic is distributed to publishers. An Increased focus on managing the conversion rate and landing page relevancy will be the main focus. 6/ OOH is kind of making a comeback. First, US OOH Ads Industry grew 1.1% in Q3 2023. Second, Outfront Media reported slight revenue increase in Q3 as Billboard Ad Revenue grew in Q3. Many Brands in UK are also aligning more toward traditional media Channels. With Burger King in UK focusing on only OOH for Christmas this year and Fashion Brands like SSENSE launching Billboards as Branding Play. 7/ Rise of Curation Continues This Year, we witnessed success of Pinterest Shuffles App, Gen-Z loved it. Similar Success with formats like IG photo dump & TikTok ‘My Fav Finds’ Carousels being the center of Gen-Z Content. Just look at this recent trend and tell me Curation isn’t personal to Online Teens. Spotify won with their idea of curating Songs with Astrology-type signs. The Fashion Products with Curated Emojis and Stickers on them, that scrappy curated approach is predicted to grow in 2024, data from Pinterest. 8/ Use of AI to Trace Consumers in the wild This year we saw a huge trend of people using Image/ face recognition tools to find or dig dirt about famous people. The biggest example was Dillion Dannis exposing Multiple images of Logan Paul’s girlfriend using AI tools. (Which was Obviously bad) But next year, I believe with better rules, big brands like Adidas or Nike will be able to find worldwide micro-influencers & Online Consumers seen wearing adidas. And partnering with them on a large scale through automated outreach. 9/ More Cartoons than Influencer-Brand Products. All the Cartoon shows are seeing huge rise on IG and TikTok, Shaun the sheep is viral, Snoopy was big this year, Sesame Street’s TikTok is working. Aussie Show Bluey is making a huge spark in the US. More Brand collaborations are on the road. Why? Cartoons have built a very consistent identity and they have social channels. I know many see Cartoons as Kids Content but on social, looking at TikTok Account of Sesame Street & Snoopy. Last month, Powerpuff Girls launched a collaboration with Nike. &#x200B; 10/ The Best Trend to get people off social media &#x200B; Try to get people off the social media apps, build your own loops. You can’t rely on social and you clearly shouldn’t burn out trying to win on social and streaming with Paid Ads or without them. This matters a lot because data shares most of your customers buy from you once or twice a year. And then they interact with your content, how bad will you feel if the only thing they remember as your content is being on TikTok. Nothing about your brand. 11/ The Internet Aesthetic will Die for Cafes & Restaurants When I wrote my post about Instagram Marketing, I mentioned this issue of Every Account looking the same. In reality, It isn’t limited to IG Feeds, This Creator points out the same Problem, mentioning the aesthetic Standards from Internet are changing how new businesses approach their whole business. More Content from Cafes & Restaurants need to be around their people and neighbourhood. 12/ Echo Chambers & Sonic Influence All Podcasts are Echo Chambers because if people wanted a new perspective in form of value. We would have chosen debates, but we chose Podcasts to find new value while being in comfort. People are now looking for more value in comfort than ever, Podcasts will continue to rise. 13/ Clever AI Integration to Better Customer Journeys in B2B & B2C Marketing Agencies can provide clever solutions to B2B Companies, and help them overcome the tag of Boring Ads only. How? Ogilvy India created an AI Ad Campaign for Cadbury, allowing SMBs to have the Bollywood Actor endorse them. They used the AI voice generation allowing businesses to alter the voice and have Shah Rukh Khan endorse their shop. A similar approach was taken by IPG India, An AI Ad with Shah Rukh Khan allowing everyone to add their face in the Branded Content. &#x200B; If I sounded like an Old head in this report or I missed on some elements like Programmatic Advertising and PPC. I will try to include better analysis and new content about future trends. You can find the post shared with examples & research, linked here.

We create AI software and provide AI automation for companies. Here is a list of the best AI tools for sales IMHO
reddit
LLM Vibe Score0
Human Vibe Score1
IntellectualAINCThis week

We create AI software and provide AI automation for companies. Here is a list of the best AI tools for sales IMHO

Here are some AI tools that are useful for sales. I tried to touch as many different parts of the sales process so the tools are all quite different but all useful for sales. I tried to include some of the best and underrated AI tools. Most of them are free so check them out if you want. I did not include ChatGPT as it can basically be used for anything with the right prompts. So these tools will be more research-oriented. A quick disclaimer – I work for the company Idealink where we create custom ChatGPT for businesses and other AI products. Apollo AI Seamless AI CoPilot AI Lavender AI Regie AI Gemini Plusdocs Make Midjourney Fireflies AI Apollo AI - Find potential customers Apollo is a platform for sales and business development. It offers a range of tools to find and engage with ideal customers. The platform has an extensive B2B database and features that streamline the sales process from prospecting to closing deals. Key Features: Extensive B2B Database: Apollo boasts a large, accurate database of over 275 million contacts, providing a wealth of potential leads and opportunities for sales teams. Data Enrichment and Lead Insights: The platform offers data enrichment capabilities, ensuring CRM systems are continuously updated with detailed and actionable lead information. AI-Driven Sales Engagement: Apollo's AI technology assists in crafting effective communication and prioritizing high-value leads, enhancing the overall sales engagement process. Comprehensive Sales Tools: The platform provides an integrated suite of tools for email, call, and social media engagement, combined with analytics and automation features to streamline the sales cycle. Tailored Solutions for Teams: Apollo offers customized solutions for different team types, including sales and business development, founders, and marketing teams, addressing specific needs and goals. Seamless AI - Sale process made easier Seamless.AI is an innovative B2B sales lead generation solution that allows sales teams to efficiently connect with their ideal customers. The platform's features provide accurate and up-to-date contact information and integrate easily with existing sales and marketing tools. Key Features: Real-Time Search Engine: Seamless.AI uses AI to scour the web in real time, ensuring the contact information for sales leads is current and accurate. Comprehensive Integration: Easily integrates with popular CRMs and sales tools like Salesforce, HubSpot, and LinkedIn Sales Navigator, enhancing productivity and eliminating manual data entry. Chrome Extension: Enhances web browsing experience for sales teams, allowing them to build lead lists directly from their browser. Pitch Intelligence and Writer: Tools for crafting effective sales messages and marketing content, personalized for each potential customer. Data Enrichment and Autopilot: Keeps customer data current and automates lead-building, supporting consistent lead generation. Buyer Intent Data and Job Changes: Offers insights into potential customers' buying intentions and keeps track of significant job changes within key accounts. CoPilot AI - Helps sales reps manage leads CoPilot AI is an advanced AI-powered sales support platform designed for B2B sales teams and agencies to drive consistent revenue growth. The tool focuses on using LinkedIn for sales prospecting, engagement, and conversion. Key Features: LinkedIn Lead Generation: Targets and automates outreach to high-intent LinkedIn leads, enhancing efficiency and scalability in lead generation. Personalized Messaging Automation: Facilitates sending of personalized, one-click messages at scale, maintaining a human touch in digital interactions. Sales Conversion Insights: Offers tools to understand and adapt to prospects' communication styles, improving the likelihood of conversion. Sales Process Optimization: Provides analytics to evaluate and refine sales strategies, identifying opportunities for improvement in the sales funnel. Industry Versatility: Adapts to diverse industries, offering tailored solutions for B2B sales, marketing, HR, and financial services sectors. Collaborative Team Tools: Enables team synchronization and collaboration, boosting productivity and synergy in sales teams Lavender AI - Email AI assistant Lavender AI is an AI-powered email tool that helps users write better emails. It provides real-time feedback and personalized suggestions to optimize email communication efficiency. Key Features: Email Coaching and Scoring: Lavender evaluates emails using AI and a vast database of email interactions, offering a score and tips for improvement. It identifies factors that might reduce the likelihood of receiving a reply, helping users refine their email content. Personalization Assistant: This feature integrates prospect data directly into the user's email platform, suggesting personalization strategies based on recipient data and personality insights to foster deeper connections. Adaptive Improvement: Lavender's scoring and recommendations evolve in real-time with changing email behaviors and practices, thanks to its generative AI and extensive data analysis, ensuring users always follow the best practices. Data-Driven Managerial Insights: The platform provides managers with valuable insights derived from actual email interactions, aiding them in coaching their teams more effectively based on real performance and communication trends. Broad Integration Capability: Lavender integrates with various email and sales platforms including Gmail, Outlook, and others, making it versatile for different user preferences and workflows. Regie AI - Great for business intelligence Regie.ai simplifies the sales prospecting process for businesses, using GenAI and automation to improve interactions with prospects. The platform offers tools like Auto-Pilot for automatic prospecting and meeting scheduling, Co-Pilot for sales rep support, and integrations with various CRM and sales engagement platforms. It also includes a Chrome Extension and CMS for content management and customization. Key Features: Automated Prospecting with Auto-Pilot: Regie.ai's Auto-Pilot feature autonomously prospects and schedules meetings, using Generative AI for Sales Agents to enhance outbound sales efforts. Audience Discovery and Content Generation: The platform identifies target accounts not in the CRM, generating relevant, on-brand content for each message, thus ensuring efficiency in list building and message personalization. Outbound Prioritization and Dynamic Engagement: It utilizes engagement and intent data to prioritize outreach to in-market prospects and adjust engagement strategies based on buyer responsiveness. Full Funnel Brand Protection and Analytics: Regie.ai ensures consistent use of marketing-approved language in all sales outreach and provides insights into campaign and document performance, thereby safeguarding brand integrity throughout the sales funnel. Gemini - AI powered conversational platform Gemini is a large language model chatbot developed by Google AI. It can generate text, translate languages, write different creative text formats, and answer your questions in an informative way. It is still under development but has learned to perform many kinds of tasks. Key features: Generate different creative text formats of text content (poems, code, scripts, musical pieces, email, letters, etc.) Answer your questions in an informative way, even if they are open ended, challenging, or strange. Translate languages Follow your instructions and complete your requests thoughtfully. Plusdocs (Plus AI) - AI tool for presentations Plus AI is a versatile tool that helps improve presentations and integrates with Slides in a simple and intuitive way. It simplifies slide creation and customization by converting text into slides and utilizing AI for various languages. Key Features: Text-to-Slide Conversion: Plus AI excels in transforming textual content into visually appealing slides, streamlining the presentation creation process. Multilingual AI Support: The tool is equipped to handle various languages, making it adaptable for a global user base. Professional Design Options: Users have access to professionally designed slide layouts, enabling the creation of polished presentations with ease. Customization and AI Design: Plus AI allows for extensive customization, including the use of AI for designing and editing slides, ensuring unique and personalized presentations. Live Snapshots and Templates: The tool offers live snapshots for real-time updates and a wide range of templates for quick and effective slide creation. Make - AI automation Make is a powerful visual platform that allows users to build and automate tasks, workflows, apps, and systems. It offers an intuitive, no-code interface that empowers users across various business functions to design and implement complex processes without the need for developer resources. Key Features: No-Code Visual Workflow Builder: Make's core feature is its user-friendly interface that allows for the creation of intricate workflows without coding expertise, making it accessible to a wide range of users. Extensive App Integration: The platform boasts compatibility with over 1000 apps, facilitating seamless connections and data sharing across diverse tools and systems. Custom Automation Solutions: Make enables personalized automation strategies, fitting various business needs from marketing automation to IT workflow control. Template Library: Users can jumpstart their automation projects with a vast collection of pre-built templates, which are customizable to fit specific workflow requirements. Enterprise-Level Solutions: Make offers advanced options for larger organizations, including enhanced security, single sign-on, custom functions, and dedicated support. Midjourney - Making sales content Midjourney is an AI-based image generation tool that changes the way we visualise and create digital art. It offers a lot of artistic possibilities, allowing users to create stunning images from text prompts. This innovative service caters to artists, designers, and anyone seeking to bring their creative visions to life. Key Features: Advanced AI Image Generation: Midjourney's core strength lies in its powerful AI algorithms, which interpret text prompts to generate detailed, high-quality images. This feature allows users to explore an endless array of visual concepts and styles. User-driven Customization: The tool offers significant control over the image creation process, enabling users to guide the AI with specific instructions, ensuring that the final output aligns closely with their vision. Diverse Artistic Styles: Midjourney can mimic various artistic styles, from classical to contemporary, providing users with a wide range of aesthetic options for their creations. Collaboration and Community Features: The platform fosters a community of users who can share, critique, and collaborate on artistic projects, enriching the creative experience. Fireflies AI - Sales meeting assistant Fireflies.ai is a powerful tool for improving team productivity and efficiency in managing meetings and voice conversations. It offers a range of features to simplify the process of capturing, organizing, and analyzing meeting content. Key Features: Automatic Meeting Transcription: Fireflies.ai can transcribe meetings held on various video-conferencing platforms and dialers. The tool captures both video and audio, providing transcripts quickly and efficiently. AI-Powered Search and Summarization: It allows users to review long meetings in a fraction of the time, highlighting key action items, tasks, and questions. Users can filter and focus on specific topics discussed in meetings. Improved Collaboration: The tool enables adding comments, pins, and reactions to specific conversation parts. Users can create and share soundbites and integrate meeting notes with popular collaboration apps such as Slack, Notion, and Asana. Conversation Intelligence: Fireflies.ai offers insights into meetings by tracking metrics like speaker talk time and sentiment. It helps in coaching team members and improving performance in sales, recruiting, and other internal processes. Workflow Automation: The AI assistant from Fireflies.ai can log call notes and activities in CRMs, create tasks through voice commands, and share meeting recaps instantly across various platforms. Comprehensive Knowledge Base: It compiles all voice conversations into an easily accessible and updatable knowledge base, with features to organize meetings into channels and set custom privacy controls. I’ll keep updating this little guide, so add your comments and I’ll try to add more tools. This is all just a personal opinion, so it’s completely cool if you disagree with it. Btw here is the link to the full blog post about all the AI tools in a bit more depth.

boring passive site... now 42k monthly visitors and $2540 MRR
reddit
LLM Vibe Score0
Human Vibe Score1
TasAdamsThis week

boring passive site... now 42k monthly visitors and $2540 MRR

people underestimate SEO... It is evergreen... passive... digital real estate. it can do magic... if you are consistent. Especially now with AI you can 2X your traffic growth and automate 85% of the work. For the past 6 months... we've been building an online directory. we just reached $2540 MRR... with SEO only... from a complete zero. I did share this on other subreddits. Maybe this gives ideas to someone. \+ This can be easily replicated if you have a website lol Current metrics: $2540 MRR - businesses pay us to list on the directory + display ads + pay to be featured. 43k monthly visitors - in the past couple of weeks our SEO growth is a hockey stick. DR (Domain Rating) 35 - it took us 2.5 months to get to that. 51 okay-ish quality referring domains (90% of them are do-follow) and 1.6k backlinks. There are probably 3 main pillars I try to focus on: keywords --> which then is the basis for ALL the content pieces we do blogs, landing pages, about us pages, competitor comparisons etc --> we use a DIY excel file to automate content production at scale. backlinks --> boost DR --> one of the main things to boost ranking on google. website health --> this is technical stuff like internal and external linking, schemas, canonical tags, alt texts, load speeds, compressed images, meta descriptions, titles etc --> do this once... and do it GOOD. $0.07 per SEO optimised blog at scale with AI Yep... we've literally built our own SEO blog tool... and it is a Spreadsheet with bunch of app scripts :D NOTE that we add a little bit of human touch to those blogs that are picked up by Google rank top in 25 How it works... is that we paste in bunch of links (other websites, blogs, news articles) and with a click of a button we can get up to 2000 SEO optimised content pieces... from an Excel file... $0.07 per blog. The spreadsheet is integrated with Chat gpt (obviously). We use GPT-4 for meta descriptions, titles, transforming the content from text to html code since it is more powerful, and GPT-4o for content itself because it is cheaper and faster for "general text". The spreadsheet repurposes content. The spreadsheet generates: Meta descriptions and titles FAQs sections - DON'T skip FAQ sections! They are a must for SEO. On Ahrefs... there is a section of questions people are searching about your keyword... that's your FAQs It can find contextual youtube videos (links to those videos) - to show google that our content is not "just text" thus higher quality. Screenshots and images of the original source (the website link we inputed) I then download a csv version of the excel and import it into our Webflow. The csv file column names match our webflow CMS field names. tbh... we didn't even know that it can be done with a spreadsheet. We "tried" building it because every other tool we were using is (1) expensive from $0.59 per SEO content piece (2) they didn't provide the scale we wanted (3) we wanted more control over the output. Focus on DR 35+ backlinks... easier We bought backlinks only once... rest of the backlinks was a manual work from us. Bunch of free listing databases (about 65% of our backlinks) You can comment on open forums with your link to get a backlink (be careful tho) Post a blog on Medium com --> DR 94 backlink (takes time to Index) If you pay for Notion you can get a DR 94 backlink from Notion If you use Beehiiv you can get a DR 86 backlink from Beehiiv Google product stacking (Google sites, Google notes etc) --> backlink from almighty Google itself A lot of work goes into backlinks because they are THAT important. I have tried bunch of "black hat" strategies as well... but note that all of these strategies won't work if you don't index the primary source from where your backlink is coming from. BIG search volume and low KD Key things I'm looking for in keywords: I use Ahrefs Keyword research tool... it is literally free BIG search volume - 2k+ is oaky-ish for a single keyword EASY to rank - KD (keyword difficulty) below 15 Look for long tail keywords (these are golden nuggets since they have a VERY clear search intent) - "how to edit..." "how to change..." "how to delete..." "how to paint..." I hope you got the idea. on Ahrefs you can use "\" to get BIG volume long tail keywords... like this "my keyword\". Ahrefs then populates the "\" with the tail. Check SERP (Search Engine Result Page) for your keywords - it shows current top 10 pages for those KWs. Check their content. Can you improve it? Have they missed anything? Keyword gap from your competitors - shows EASY keywords that your competitors have missed and also shows what keywords overlap with you. Also one cool thing... if you don't type any keywords on Ahrefs and press "Enter"... you can browse all the keywords out there... it is magical. Once we have the keywords, we run our spreadsheet. And that's pretty much it. I hope that you can get some ideas from this little silly project. Also... if you have any questions about this... I might share the SEO blog automation excel file/help if people are interested...

I run an AI automation agency (AAA). My honest overview and review of this new business model
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

I run an AI automation agency (AAA). My honest overview and review of this new business model

I started an AI tools directory in February, and then branched off that to start an AI automation agency (AAA) in June. So far I've come across a lot of unsustainable "ideas" to make money with AI, but at the same time a few diamonds in the rough that aren't fully tapped into yet- especially the AAA model. Thought I'd share this post to shine light into this new business model and share some ways you could potentially start your own agency, or at the very least know who you are dealing with and how to pick and choose when you (inevitably) get bombarded with cold emails from them down the line. Foreword Running an AAA does NOT involve using AI tools directly to generate and sell content directly. That ship has sailed, and unless you are happy with $5 from Fiverr every month or so, it is not a real business model. Cry me a river but generating generic art with AI and slapping it onto a T-shirt to sell on Etsy won't make you a dime. At the same time, the AAA model will NOT require you to have a deep theoretical knowledge of AI, or any academic degree, as we are more so dealing with the practical applications of generative AI and how we can implement these into different workflows and tech-stacks, rather than building AI models from the ground up. Regardless of all that, common sense and a willingness to learn will help (a shit ton), as with anything. Keep in mind - this WILL involve work and motivation as well. The mindset that AI somehow means everything can be done for you on autopilot is not the right way to approach things. The common theme of businesses I've seen who have successfully implemented AI into their operations is the willingess to work with AI in a way that augments their existing operations, rather than flat out replace a worker or team. And this is exactly the train of thought you need when working with AI as a business model. However, as the field is relatively unsaturated and hype surrounding AI is still fresh for enterprises, right now is the prime time to start something new if generative AI interests you at all. With that being said, I'll be going over three of the most successful AI-adjacent businesses I've seen over this past year, in addition to some tips and resources to point you in the right direction. so.. WTF is an AI Automation Agency? The AI automation agency (or as some YouTubers have coined it, the AAA model) at its core involves creating custom AI solutions for businesses. I have over 1500 AI tools listed in my directory, however the feedback I've received from some enterprise users is that ready-made SaaS tools are too generic to meet their specific needs. Combine this with the fact virtually no smaller companies have the time or skills required to develop custom solutions right off the bat, and you have yourself real demand. I would say in practice, the AAA model is quite similar to Wordpress and even web dev agencies, with the major difference being all solutions you develop will incorporate key aspects of AI AND automation. Which brings me to my second point- JUST AI IS NOT ENOUGH. Rather than reducing the amount of time required to complete certain tasks, I've seen many AI agencies make the mistake of recommending and (trying to) sell solutions that more likely than not increase the workload of their clients. For example, if you were to make an internal tool that has AI answer questions based on their knowledge base, but this knowledge base has to be updated manually, this is creating unnecessary work. As such I think one of the key components of building successful AI solutions is incorporating the new (Generative AI/LLMs) with the old (programmtic automation- think Zapier, APIs, etc.). Finally, for this business model to be successful, ideally you should target a niche in which you have already worked and understand pain points and needs. Not only does this make it much easier to get calls booked with prospects, the solutions you build will have much greater value to your clients (meaning you get paid more). A mistake I've seen many AAA operators make (and I blame this on the "Get Rich Quick" YouTubers) is focusing too much on a specific productized service, rather than really understanding the needs of businesses. The former is much done via a SaaS model, but when going the agency route the only thing that makes sense is building custom solutions. This is why I always take a consultant-first approach. You can only build once you understand what they actually need and how certain solutions may impact their operations, workflows, and bottom-line. Basics of How to Get Started Pick a niche. As I mentioned previously, preferably one that you've worked in before. Niches I know of that are actively being bombarded with cold emails include real estate, e-commerce, auto-dealerships, lawyers, and medical offices. There is a reason for this, but I will tell you straight up this business model works well if you target any white-collar service business (internal tools approach) or high volume businesses (customer facing tools approach). Setup your toolbox. If you wanted to start a pressure washing business, you would need a pressure-washer. This is no different. For those without programming knowledge, I've seen two common ways AAA get setup to build- one is having a network of on-call web developers, whether its personal contacts or simply going to Upwork or any talent sourcing agency. The second is having an arsenal of no-code tools. I'll get to this more in a second, but this works beecause at its core, when we are dealing with the practical applications of AI, the code is quite simple, simply put. Start cold sales. Unless you have a network already, this is not a step you can skip. You've already picked a niche, so all you have to do is find the right message. Keep cold emails short, sweet, but enticing- and it will help a lot if you did step 1 correctly and intimately understand who your audience is. I'll be touching base later about how you can leverage AI yourself to help you with outreach and closing. The beauty of gen AI and the AAA model You don't need to be a seasoned web developer to make this business model work. The large majority of solutions that SME clients want is best done using an API for an LLM for the actual AI aspect. The value we create with the solutions we build comes with the conceptual framework and design that not only does what they need it to but integrates smoothly with their existing tech-stack and workflow. The actual implementation is quite straightforward once you understand the high level design and know which tools you are going to use. To give you a sense, even if you plan to build out these apps yourself (say in Python) the large majority of the nitty gritty technical work has already been done for you, especially if you leverage Python libraries and packages that offer high level abstraction for LLM-related functions. For instance, calling GPT can be as little as a single line of code. (And there are no-code tools where these functions are simply an icon on a GUI). Aside from understanding the capabilities and limitations of these tools and frameworks, the only thing that matters is being able to put them in a way that makes sense for what you want to build. Which is why outsourcing and no-code tools both work in our case. Okay... but how TF am I suppposed to actually build out these solutions? Now the fun part. I highly recommend getting familiar with Langchain and LlamaIndex. Both are Python libraires that help a lot with the high-level LLM abstraction I mentioned previously. The two most important aspects include being able to integrate internal data sources/knowledge bases with LLMs, and have LLMs perform autonomous actions. The two most common methods respectively are RAG and output parsing. RAG (retrieval augmented Generation) If you've ever seen a tool that seemingly "trains" GPT on your own data, and wonder how it all works- well I have an answer from you. At a high level, the user query is first being fed to what's called a vector database to run vector search. Vector search basically lets you do semantic search where you are searching data based on meaning. The vector databases then retrieves the most relevant sections of text as it relates to the user query, and this text gets APPENDED to your GPT prompt to provide extra context to the AI. Further, with prompt engineering, you can limit GPT to only generate an answer if it can be found within this extra context, greatly limiting the chance of hallucination (this is where AI makes random shit up). Aside from vector databases, we can also implement RAG with other data sources and retrieval methods, for example SQL databses (via parsing the outputs of LLM's- more on this later). Autonomous Agents via Output Parsing A common need of clients has been having AI actually perform tasks, rather than simply spitting out text. For example, with autonomous agents, we can have an e-commerce chatbot do the work of a basic customer service rep (i.e. look into orders, refunds, shipping). At a high level, what's going on is that the response of the LLM is being used programmtically to determine which API to call. Keeping on with the e-commerce example, if I wanted a chatbot to check shipping status, I could have a LLM response within my app (not shown to the user) with a prompt that outputs a random hash or string, and programmatically I can determine which API call to make based on this hash/string. And using the same fundamental concept as with RAG, I can append the the API response to a final prompt that would spit out the answer for the user. How No Code Tools Can Fit In (With some example solutions you can build) With that being said, you don't necessarily need to do all of the above by coding yourself, with Python libraries or otherwise. However, I will say that having that high level overview will help IMMENSELY when it comes to using no-code tools to do the actual work for you. Regardless, here are a few common solutions you might build for clients as well as some no-code tools you can use to build them out. Ex. Solution 1: AI Chatbots for SMEs (Small and Medium Enterprises) This involves creating chatbots that handle user queries, lead gen, and so forth with AI, and will use the principles of RAG at heart. After getting the required data from your client (i.e. product catalogues, previous support tickets, FAQ, internal documentation), you upload this into your knowledge base and write a prompt that makes sense for your use case. One no-code tool that does this well is MyAskAI. The beauty of it especially for building external chatbots is the ability to quickly ingest entire websites into your knowledge base via a sitemap, and bulk uploading files. Essentially, they've covered the entire grunt work required to do this manually. Finally, you can create a inline or chat widget on your client's website with a few lines of HTML, or altneratively integrate it with a Slack/Teams chatbot (if you are going for an internal Q&A chatbot approach). Other tools you could use include Botpress and Voiceflow, however these are less for RAG and more for building out complete chatbot flows that may or may not incorporate LLMs. Both apps are essentially GUIs that eliminate the pain and tears and trying to implement complex flows manually, and both natively incoporate AI intents and a knowledge base feature. Ex. Solution 2: Internal Apps Similar to the first example, except we go beyond making just chatbots but tools such as report generation and really any sort of internal tool or automations that may incorporate LLM's. For instance, you can have a tool that automatically generates replies to inbound emails based on your client's knowledge base. Or an automation that does the same thing but for replies to Instagram comments. Another example could be a tool that generates a description and screeenshot based on a URL (useful for directory sites, made one for my own :P). Getting into more advanced implementations of LLMs, we can have tools that can generate entire drafts of reports (think 80+ pages), based not only on data from a knowledge base but also the writing style, format, and author voice of previous reports. One good tool to create content generation panels for your clients would be MindStudio. You can train LLM's via prompt engineering in a structured way with your own data to essentially fine tune them for whatever text you need it to generate. Furthermore, it has a GUI where you can dictate the entire AI flow. You can also upload data sources via multiple formats, including PDF, CSV, and Docx. For automations that require interactions between multiple apps, I recommend the OG zapier/make.com if you want a no-code solution. For instance, for the automatic email reply generator, I can have a trigger such that when an email is received, a custom AI reply is generated by MyAskAI, and finally a draft is created in my email client. Or, for an automation where I can create a social media posts on multiple platforms based on a RSS feed (news feed), I can implement this directly in Zapier with their native GPT action (see screenshot) As for more complex LLM flows that may require multiple layers of LLMs, data sources, and APIs working together to generate a single response i.e. a long form 100 page report, I would recommend tools such as Stack AI or Flowise (open-source alternative) to build these solutions out. Essentially, you get most of the functions and features of Python packages such as Langchain and LlamaIndex in a GUI. See screenshot for an example of a flow How the hell are you supposed to find clients? With all that being said, none of this matters if you can't find anyone to sell to. You will have to do cold sales, one way or the other, especially if you are brand new to the game. And what better way to sell your AI services than with AI itself? If we want to integrate AI into the cold outreach process, first we must identify what it's good at doing, and that's obviously writing a bunch of text, in a short amount of time. Similar to the solutions that an AAA can build for its clients, we can take advantage of the same principles in our own sales processes. How to do outreach Once you've identified your niche and their pain points/opportunities for automation, you want to craft a compelling message in which you can send via cold email and cold calls to get prospects booked on demos/consultations. I won't get into too much detail in terms of exactly how to write emails or calling scripts, as there are millions of resources to help with this, but I will tell you a few key points you want to keep in mind when doing outreach for your AAA. First, you want to keep in mind that many businesses are still hesitant about AI and may not understand what it really is or how it can benefit their operations. However, we can take advantage of how mass media has been reporting on AI this past year- at the very least people are AWARE that sooner or later they may have to implement AI into their businesses to stay competitive. We want to frame our message in a way that introduces generative AI as a technology that can have a direct, tangible, and positive impact on their business. Although it may be hard to quantify, I like to include estimates of man-hours saved or costs saved at least in my final proposals to prospects. Times are TOUGH right now, and money is expensive, so you need to have a compelling reason for businesses to get on board. Once you've gotten your messaging down, you will want to create a list of prospects to contact. Tools you can use to find prospects include Apollo.io, reply.io, zoominfo (expensive af), and Linkedin Sales Navigator. What specific job titles, etc. to target will depend on your niche but for smaller companies this will tend to be the owner. For white collar niches, i.e. law, the professional that will be directly benefiting from the tool (i.e. partners) may be better to contact. And for larger organizations you may want to target business improvement and digital transformation leads/directors- these are the people directly in charge of projects like what you may be proposing. Okay- so you have your message, and your list, and now all it comes down to is getting the good word out. I won't be going into the details of how to send these out, a quick Google search will give you hundreds of resources for cold outreach methods. However, personalization is key and beyond simple dynamic variables you want to make sure you can either personalize your email campaigns directly with AI (SmartWriter.ai is an example of a tool that can do this), or at the very least have the ability to import email messages programmatically. Alternatively, ask ChatGPT to make you a Python Script that can take in a list of emails, scrape info based on their linkedin URL or website, and all pass this onto a GPT prompt that specifies your messaging to generate an email. From there, send away. How tf do I close? Once you've got some prospects booked in on your meetings, you will need to close deals with them to turn them into clients. Call #1: Consultation Tying back to when I mentioned you want to take a consultant-first appraoch, you will want to listen closely to their goals and needs and understand their pain points. This would be the first call, and typically I would provide a high level overview of different solutions we could build to tacke these. It really helps to have a presentation available, so you can graphically demonstrate key points and key technologies. I like to use Plus AI for this, it's basically a Google Slides add-on that can generate slide decks for you. I copy and paste my default company messaging, add some key points for the presentation, and it comes out with pretty decent slides. Call #2: Demo The second call would involve a demo of one of these solutions, and typically I'll quickly prototype it with boilerplate code I already have, otherwise I'll cook something up in a no-code tool. If you have a niche where one type of solution is commonly demanded, it helps to have a general demo set up to be able to handle a larger volume of calls, so you aren't burning yourself out. I'll also elaborate on how the final product would look like in comparison to the demo. Call #3 and Beyond: Once the initial consultation and demo is complete, you will want to alleviate any remaining concerns from your prospects and work with them to reach a final work proposal. It's crucial you lay out exactly what you will be building (in writing) and ensure the prospect understands this. Furthermore, be clear and transparent with timelines and communication methods for the project. In terms of pricing, you want to take this from a value-based approach. The same solution may be worth a lot more to client A than client B. Furthermore, you can create "add-ons" such as monthly maintenance/upgrade packages, training sessions for employeees, and so forth, separate from the initial setup fee you would charge. How you can incorporate AI into marketing your businesses Beyond cold sales, I highly recommend creating a funnel to capture warm leads. For instance, I do this currently with my AI tools directory, which links directly to my AI agency and has consistent branding throughout. Warm leads are much more likely to close (and honestly, much nicer to deal with). However, even without an AI-related website, at the very least you will want to create a presence on social media and the web in general. As with any agency, you will want basic a professional presence. A professional virtual address helps, in addition to a Google Business Profile (GBP) and TrustPilot. a GBP (especially for local SEO) and Trustpilot page also helps improve the looks of your search results immensely. For GBP, I recommend using ProfilePro, which is a chrome extension you can use to automate SEO work for your GBP. Aside from SEO optimzied business descriptions based on your business, it can handle Q/A answers, responses, updates, and service descriptions based on local keywords. Privacy and Legal Concerns of the AAA Model Aside from typical concerns for agencies relating to service contracts, there are a few issues (especially when using no-code tools) that will need to be addressed to run a successful AAA. Most of these surround privacy concerns when working with proprietary data. In your terms with your client, you will want to clearly define hosting providers and any third party tools you will be using to build their solution, and a DPA with these third parties listed as subprocessors if necessary. In addition, you will want to implement best practices like redacting private information from data being used for building solutions. In terms of addressing concerns directly from clients, it helps if you host your solutions on their own servers (not possible with AI tools), and address the fact only ChatGPT queries in the web app, not OpenAI API calls, will be used to train OpenAI's models (as reported by mainstream media). The key here is to be open and transparent with your clients about ALL the tools you are using, where there data will be going, and make sure to get this all in writing. have fun, and keep an open mind Before I finish this post, I just want to reiterate the fact that this is NOT an easy way to make money. Running an AI agency will require hours and hours of dedication and work, and constantly rearranging your schedule to meet prospect and client needs. However, if you are looking for a new business to run, and have a knack for understanding business operations and are genuinely interested in the pracitcal applications of generative AI, then I say go for it. The time is ticking before AAA becomes the new dropshipping or SMMA, and I've a firm believer that those who set foot first and establish themselves in this field will come out top. And remember, while 100 thousand people may read this post, only 2 may actually take initiative and start.

I am starting a startup on AI research automation. Looking for feedback!
reddit
LLM Vibe Score0
Human Vibe Score1
pablonmThis week

I am starting a startup on AI research automation. Looking for feedback!

Hi everyone, I would like to share a product idea that I'm working on. I studied computer science and have worked for Silicon Valley startups for the last 6 years. I'm currently employed full-time at a startup that sells an AI-powered search engine, so I have gained valuable experience in the AI/information retrieval space. I turned 29 last week, and I think it's time for me to start my own business. I've always wanted to run my own tech company, and I feel like now is the right time to begin with an idea. Are you a researcher in any field? Do you often find yourself learning new, highly complex topics and don't know where to start? Google is a great tool for finding answers to specific questions, but what if you don't know what questions to ask? I am developing a "deep search" engine that, given a topic, produces a multi-page report aggregating information from several properly cited sources. It finds and explains different perspectives and ideas related to the topic of interest. You can use it to automate the research process, but it's much more than that because it can help you uncover hidden perspectives, important questions, and ideas that you might not otherwise find when just googling. I welcome any feedback and ideas! Do you think this product would deliver significant value to your life? Why or why not? Would you be willing to pay to use it? I will post updates about this product in this thread in case you want to follow its development and try the product when it's ready.

Secret behind Airbnb's Billion-Dollar Empire? Spamming Craigslist
reddit
LLM Vibe Score0
Human Vibe Score1
deadcoder0904This week

Secret behind Airbnb's Billion-Dollar Empire? Spamming Craigslist

Silicon Valley wants you to believe that their unicorn startups succeeded doing things legally. But that couldn't be far from truth. For starters, Airbnb used multiple Gmail accounts to spam Craigslist. "They posted unrealistically (fake) cheap rentals of beautiful apartments in places where normal rent should be 10x more. Once people replied, they auto-responded that the unit has been rented, but they should be looking for another unit on AirBnB." The Game of Blackhat is a cat-and-mouse game. You need a lot of guardrails to protect yourself from people using your Social Site by spamming their products. Craigslist is a team of 30 people. There's stuff AI can automate now with such a small team but back then, it wasn't possible. Airbnb used Craigslist as its playground to spam Craigslist visitors to grow their supply-side. In a 2-sided marketplace, growing both supply and demand is very important. And both must grow at the same time for the marketplace to work. A Blackhat Marketer created a new test site to get vacation rental owners to sign-up so that he can test his Airbnb theory. He grabbed their real email-addresses (not Craigslist anonymous addresses) via Craigslist by specifically targeting those who were advertising their vacation rentals on Craigslist. He skipped over the other categories that were directly related to AirBnB's business model because they didn't fit with the test site he built. Once he got 1000+ sign-ups, he then took it upon himself to post it to the advertising section on Craigslist. The email said this: I am emailing you because you have one of the nicest listings on Craigslist in Idaho and I want to recommend you feature it (for free) on one of the largest Idaho housing sites on the web, Airbnb. The site already has 3,000,000 pages views a month. Check it out here to list now: airbnb(dot)com Sarah Surpisingly, all emails were by ladies. He did the same in Week 2 and Week 3 to test if it wasn't a one-time thing. Surely, it wasn't a fluke. After posting 4 ads on Craigslist in 3 weeks, he received 5 identical emails from 2 ladies who were raving fans of AirBnB and spent their days emailing Craigslist advertisers. This is one of the greatest blackhat strategies used in the real world to build a billion-dollar marketplace by growing the supply-side with pure blackhat. These strategies are not mentioned in Press Interviews, Media, or any Founder stories but this is probably the most important piece of the puzzle. Without it, Airbnb probably wouldn't have survived. "Some very famous investors have alluded to the fact that they look for a dangerous streak in the entrepreneurs they invest in…and while those investors will never come out and tell you what they mean, this kind of thing is probably what they mean." It definitely violates CAN-SPAM act. Some comments from Hacker News: "CAN-SPAM, sending from a fake address (illegal headers). CA has a specific law that pre-empts CAN-SPAM that definitely makes this illegal if sent from CA." But I guess it worked in Airbnb's favour lol as they were never caught or fined until after. "It's commercial email 100%. Probably a fake sender name (illegal), against gmail ToS, against CL ToS and no unsubscribe link and no one even subscribed in the first place. 100% against CAN-SPAM." Thanks for reading. If you'd like to learn more blackhat tactics like this, check this site which is a growth hacking newsletter with real-world blackhat examples. PS: Actual emails & screenshots from the Airbnb x Craigslist spam can be found here.

boring passive site... now 42k monthly visitors and $2540 MRR
reddit
LLM Vibe Score0
Human Vibe Score1
TasAdamsThis week

boring passive site... now 42k monthly visitors and $2540 MRR

people underestimate SEO... It is evergreen... passive... digital real estate. it can do magic... if you are consistent. Especially now with AI you can 2X your traffic growth and automate 85% of the work. For the past 6 months... we've been building an online directory. we just reached $2540 MRR... with SEO only... from a complete zero. I did share this on other subreddits. Maybe this gives ideas to someone. \+ This can be easily replicated if you have a website lol Current metrics: $2540 MRR - businesses pay us to list on the directory + display ads + pay to be featured. 43k monthly visitors - in the past couple of weeks our SEO growth is a hockey stick. DR (Domain Rating) 35 - it took us 2.5 months to get to that. 51 okay-ish quality referring domains (90% of them are do-follow) and 1.6k backlinks. There are probably 3 main pillars I try to focus on: keywords --> which then is the basis for ALL the content pieces we do blogs, landing pages, about us pages, competitor comparisons etc --> we use a DIY excel file to automate content production at scale. backlinks --> boost DR --> one of the main things to boost ranking on google. website health --> this is technical stuff like internal and external linking, schemas, canonical tags, alt texts, load speeds, compressed images, meta descriptions, titles etc --> do this once... and do it GOOD. $0.07 per SEO optimised blog at scale with AI Yep... we've literally built our own SEO blog tool... and it is a Spreadsheet with bunch of app scripts :D NOTE that we add a little bit of human touch to those blogs that are picked up by Google rank top in 25 How it works... is that we paste in bunch of links (other websites, blogs, news articles) and with a click of a button we can get up to 2000 SEO optimised content pieces... from an Excel file... $0.07 per blog. The spreadsheet is integrated with Chat gpt (obviously). We use GPT-4 for meta descriptions, titles, transforming the content from text to html code since it is more powerful, and GPT-4o for content itself because it is cheaper and faster for "general text". The spreadsheet repurposes content. The spreadsheet generates: Meta descriptions and titles FAQs sections - DON'T skip FAQ sections! They are a must for SEO. On Ahrefs... there is a section of questions people are searching about your keyword... that's your FAQs It can find contextual youtube videos (links to those videos) - to show google that our content is not "just text" thus higher quality. Screenshots and images of the original source (the website link we inputed) I then download a csv version of the excel and import it into our Webflow. The csv file column names match our webflow CMS field names. tbh... we didn't even know that it can be done with a spreadsheet. We "tried" building it because every other tool we were using is (1) expensive from $0.59 per SEO content piece (2) they didn't provide the scale we wanted (3) we wanted more control over the output. Focus on DR 35+ backlinks... easier We bought backlinks only once... rest of the backlinks was a manual work from us. Bunch of free listing databases (about 65% of our backlinks) You can comment on open forums with your link to get a backlink (be careful tho) Post a blog on Medium com --> DR 94 backlink (takes time to Index) If you pay for Notion you can get a DR 94 backlink from Notion If you use Beehiiv you can get a DR 86 backlink from Beehiiv Google product stacking (Google sites, Google notes etc) --> backlink from almighty Google itself A lot of work goes into backlinks because they are THAT important. I have tried bunch of "black hat" strategies as well... but note that all of these strategies won't work if you don't index the primary source from where your backlink is coming from. BIG search volume and low KD Key things I'm looking for in keywords: I use Ahrefs Keyword research tool... it is literally free BIG search volume - 2k+ is oaky-ish for a single keyword EASY to rank - KD (keyword difficulty) below 15 Look for long tail keywords (these are golden nuggets since they have a VERY clear search intent) - "how to edit..." "how to change..." "how to delete..." "how to paint..." I hope you got the idea. on Ahrefs you can use "\" to get BIG volume long tail keywords... like this "my keyword\". Ahrefs then populates the "\" with the tail. Check SERP (Search Engine Result Page) for your keywords - it shows current top 10 pages for those KWs. Check their content. Can you improve it? Have they missed anything? Keyword gap from your competitors - shows EASY keywords that your competitors have missed and also shows what keywords overlap with you. Also one cool thing... if you don't type any keywords on Ahrefs and press "Enter"... you can browse all the keywords out there... it is magical. Once we have the keywords, we run our spreadsheet. And that's pretty much it. I hope that you can get some ideas from this little silly project. Also... if you have any questions about this... I might share the SEO blog automation excel file/help if people are interested...

Only 2 months of cash in the Bank for my business but was able to save it with the help of AI.
reddit
LLM Vibe Score0
Human Vibe Score1
CALLIRDAN90This week

Only 2 months of cash in the Bank for my business but was able to save it with the help of AI.

Hi there! I’m excited to share something very personal with you. We needed to book at least 2 appointments per day in the next 60 days, or my business would fail. We were already trying two acquisition channels, LinkedIn and email. The problem with these channels was that the positive response rate was very low in both. So I decided to focus on LinkedIn and get the attention of the lead by sending videos directly to them via LinkedIn messages. (You can send videos to your connections on LinkedIn if you use your cell phone.) This wasn’t new, but I added a small twist to get the lead’s attention. All the covers of the videos had a picture of me holding a sign with the person’s name and an interesting phrase. This showed some okay results, but the rest of the video was not personalized. Only the picture on the cover was. I even developed a Chrome extension for this because I thought this would be the answer and that I would book tons of appointments.  But after more trial and outreach, my leads responded, telling me that because the video itself wasn’t personalized for them, they felt like I didn’t put enough effort in, so they would not book a call with me. So after investing time and effort into my “new bright idea” and getting developers to make the Chrome extension, I was back to square one with no results. A few weeks went by, and after researching online, I found an online course from a guy who promised to teach me how to book 30+ appointments per month, guaranteed (at the time, I was making 2 or 3 appointments per week, maximum). He promised that I would only pay if he actually booked appointments for me and even offered to give me money if his course didn’t work for me. I never paid attention to internet gurus, but the offer was actually not bad, so I looked into this guy’s website. I found out he had hundreds of reviews from people who had taken his course and were talking amazing things about it. The more I read, the more excited I got. I booked a call that day and talked to a salesperson. The call was very short, and he promised I would get at least 2 appointments per day, easily. He seemed a bit cocky and told me that I just needed to trust him and the 100+ reviews from people who had taken the course. He didn’t share details, a proposal, or anything. I asked the price, and he told me it was close to $10k. (Not kidding, this was the price.) Then he told me that I would make the money back in no time with the clients I would get following his course, and that if it didn’t work, he would give me the money back. But I needed to follow everything the course said for at least 6 months. I had never paid $10k for anything in my life; it was extremely expensive for me. Also, my salary from my business was not in dollars but in a currency that was worth much less than the dollar. I continued to research more and more, but no other course was close to the number of reviews and promises that this guy had. I got desperate and told myself that I would bet everything on this course. If it worked for so many others, surely it would work for me. I got a loan from the bank and paid for the course. You might read this and think it was the most stupid thing ever, but the reality is that after 2 months in the course (I did the course as fast as I could), I learned a lot. The course was not bad; it was very extensive—probably more than 200 hours or so—and they taught a lot of things. I don’t think it was worth $10k for me, but I can see how for other people it might be worth that. Now, to the question you’re all thinking: did it get me the 2 appointments I needed per day? The answer is no. Here’s the thing: most of the techniques they taught were innovative and disruptive, but the focus was always on personalization, and they didn’t teach any way to automate the personalization. (I think, at the time they made the course, the tools didn’t exist yet.) So they taught how to do everything manually, and it took a lot—a lot of time and effort. And most annoyingly: an incredible amount of time doing operational things. I did get 2 appointments on some days, but it wasn’t consistent, and I didn’t have the time to spend 14 hours a day doing everything manually or the money to hire someone to do this for me. (I needed to also spend time delivering our service to our current clients; otherwise, they would leave.) I told them this, and they were very reasonable. After some negotiation, they gave me part of the money back. (To be fair, there was a lot of value in the course, so asking for the full $10k back would have been excessive because, in the end, it really taught me a lot of things I didn’t know.) So in the end, I spent $10k and 200+ hours on an online course, spent time and effort developing a Chrome extension, and was still not able to hit the meetings I needed. Money in the business was running out, and I needed to do something fast, or I was doomed. After investing time and effort in tools, research, and spending $10k and over 200 hours on a course that didn’t deliver the consistent results I needed, I was at a crossroads. My businesses were running out of money, and I knew I needed to find a solution quickly, or everything I had worked for would collapse. It was during this time of desperation that I started exploring other options. One night, while scrolling through the internet, I stumbled upon a 2024 article about how AI was being used to revolutionize various industries. It wasn’t directly related to appointment booking, but it sparked an idea in my mind. What if I could use AI to automate the personalization process that I had learned in the course? It seemed like a long shot, but I had nothing to lose. I started researching AI tools and technologies—YouTube videos, podcasts, pretty much everything related to AI—desperate to find something that could help me scale my outreach without investing too much time, while still maintaining the personalization that was so important. After a lot of trial and error, I found a few tools that showed promise. All of these tools were extremely new. Some of them had just launched the versions I needed just weeks ago. I can say I researched and tested more than 50 AI startups, experimenting with them, testing different approaches, checking prices (the problem was that most of them were cheap but became very expensive when applying the volume I needed to get results), and gradually refining my process. It wasn’t an overnight success, but for the first time, I felt like I was onto something that could truly work. The idea of combining AI personalization with volume was something new, and it gave me hope that I could finally book the meetings I needed without burning out. One day, I sent a video of myself talking—completely AI-generated—to my family chat group and waited for their response. None of them noticed it wasn’t actually me. At that moment, I said to myself: “Okay, I am ready to test this in the real world and see if it works.” Like everything in life, focus is key. As I mentioned earlier, we were already trying outbound strategies on LinkedIn and email, but I decided to narrow my focus to LinkedIn and specifically to video outreach. My goal was to stand out from the crowd, where most people were using text or sending generic videos. I knew that if my videos were 100% personalized, it would make a strong impression on my leads. I focused on two key metrics during my tests: Time spent on manual personalized outreach vs. AI-generated personalized outreach. Positive reply rate for non-personalized manual outreach vs. AI-generated personalized outreach. I ran a test using a sample of 50 one-minute videos sent to 50 leads, and here are the results: Time Spent to Make the Videos: Manual Process: It took me up to 10 hours to create and send 50 personalized videos. This included looking good on camera, brushing my hair, choosing appropriate clothing, ensuring proper lighting, not messing up the script, using a camera holder, recharging the phone, pausing to drink water, avoiding external sounds, being in an appropriate room, downloading the videos, deleting the videos that were not good, and sending the final ones. On average, it took me at least 12.5 minutes per one-minute video. AI Process: With AI, it took me just 32 seconds to create the exact same one-minute personalized video—without saying a word or recording a second of footage. In total, I could make and send the same 50 personalized videos in just 27 minutes. Result: The AI process was 24 times faster. Completely crazy! Positive Reply Rate: Non-Personalized Script (Manual): Using a good script without personalization (no name, job title, city, company, etc.) resulted in a positive reply rate of 4-6% on LinkedIn, including follow-ups. Personalized Script (AI): Using the same script but adding personalized details like the lead's name, company, city, and job title resulted in a positive reply rate of 15-20%, including follow-ups. Result: AI personalization led to 3x (three times) more replies. The best part was the responses. Almost everyone who replied thanked me for taking the time to research them, congratulated me on my speech, and appreciated the personalization and eloquence of my message.  These metrics were a complete breakthrough for me. I researched online to see if anyone else had done something similar, but I couldn’t find anything close. After achieving these metrics, booking the two appointments I desperately needed became easy. In fact, in the last 10 weeks, I’ve been able to consistently book 3-4 appointments per day. This success allowed me to train someone in my company to handle the process, freeing me up to focus on other aspects of the business and ultimately saving it. With the AI appointment machine we built, I even have free time now—time that I’ve been using to develop a methodology and tech tools that I now teach to others. I named the methodology Clip2Lead as a reference to the first Chrome extension I developed that didn’t work but ended up being the first step toward everything that followed. I’ve condensed everything I learned and throughout my experiences into a simple and short FREE training where I cover the entire AI appointment booking process. This includes how to find leads, create scripts, set up follow-up sequences, generate AI videos, clone your voice, compare non-AI metrics with AI metrics, and even navigate AI safety controls. I also offer Chrome extensions that helped me automate the process even further, so you can spend your time closing deals or focusing on other acquisition channels, while your AI machine for booking appointments runs with minimal effort from you. If you’re interested please get in touch with me and thank you for taking the time to read my personal story.

My Roadmap to Success with AI Automation for Small Businesses
reddit
LLM Vibe Score0
Human Vibe Score1
Giggly_ScarlettThis week

My Roadmap to Success with AI Automation for Small Businesses

Hey everybody! 👋 I’ve been working on automating small business workflows for a while now, and I wanted to share how AI and automation can help scale your business with no coding experience required. I started by automating tedious tasks for clients. Things like social media posting, client onboarding, and data transfers by using simple tools like Make and Zapier. The results were amazing! For example: One client cut down 3 hours of daily social media posting to just 15 minutes a day. Another automated follow-ups for proposals, which saved them dozens of hours each month. A boutique business streamlined its customer service by setting up a chatbot for basic FAQs and lead qualification. But here’s the thing—automation isn’t perfect, and it’s crucial to know its limitations. AI might not always get everything right. That’s why I recommend setting up workflows where you still have some oversight—like reviewing AI-generated content before posting or checking data transfers for accuracy. It’s more of a quality-control role, but it ensures the AI doesn’t stain your brand. If you're wondering where to start, here's the roadmap I followed: Start with Make or Zapier: These are perfect for non-programmers and let you automate tasks like transferring data between tools or triggering specific actions. Learn Prompt Engineering: Master how to ask AI the right questions. A little practice goes a long way! Level Up to AI Agents: Once you’re comfortable, you can build more advanced AI systems, like RAG (Retrieval-Augmented Generation) agents, which help businesses create personalized responses. Learn Python (Optional): Want to take your automation to the next level? Learning Python gives you the power to customize AI and automation workflows even further. Automation can be a huge time-saver and growth booster, but it’s not about replacing people—it’s about giving them the tools to work smarter. If you’ve been putting off automation, trust me, it’s worth diving in. Let me know if y'all have any questions and I'd be happy to answer them!

AI Voice Platform Comparison for Small Business Use Cases
reddit
LLM Vibe Score0
Human Vibe Score1
Glad-Syllabub6777This week

AI Voice Platform Comparison for Small Business Use Cases

We provide AI voice agent consultation and solutions in Upwork. One of clients’ frequent questions is which platform is best/perfect for their use cases, like lead qualification, AI receptionist, customer support, etc. This post provides our thoughts on this question. Our overall feeling is that the AI agent technology is still not there yet. It seems close but there are many corner cases the AI bot doesn't handle well. Four major players in the AI voice platforms: Bland ($65M funding) Retell ($4.6M funding) Synthflow ($7.4M funding) VAPI ($20M funding) We will only talk about Bland, Retell and VAPI. We firstly tried Synthflow and found the UI was buggy (the prompt editor froze for 20 seconds to 30 seconds when we were editing the prompt). Currently we don't use it anymore. Recommended use cases based on Upwork jobs we delivered: Bland. We recommend Bland for lead qualification as the lead qualification has a strict conversation flow (like asking questions, extracting variables, and making webhook calls). Clients/contractors can draw flow diagrams to build AI voice agents. We also find Bland is not a good fit for a small business with a monthly budget less than 5K. The reason is that common tools (like warm transfer, SMS sending) for AI voice agents are only available to enterprise clients. But warm transfers are critical for small businesses. Retell. We recommend Retell for customer support in contact centers. Retell has the best voice among competitors. One use case we build in Retell is the live translator in the ambulance call center. We tried the same prompt with the same LLM setup in VAPI. We found Retell performs way better than VAPI in terms of the translation quality and reliability. Another common scenario in the customer support domain is to have 3-way merge so that the agent can tell the summary to the transfer number while the caller can hear the conversation. VAPI. We recommend VAPI for AI receptionists and phone answering use cases. We can write a prompt and ask LLM to do the magic if callers ask questions not included in the prompt. We can set up custom tools to trigger automation (like update CRM) and warm transfer to connect to the stakeholders. One feeling we have is that VAPI is way more complicated than the other two platforms. If you don’t have developer experience and have a budget to hire a contractor, it is better to try Retell as Retell has many integrations. If you have any other questions or we miss anything, feel free to comment. We like to explore AI voice agent space together.

I Recreated An AI Phone Calling Agent That Automated Scheduling And Patient Inquiries For A  Hospital
reddit
LLM Vibe Score0
Human Vibe Score1
Will_feverThis week

I Recreated An AI Phone Calling Agent That Automated Scheduling And Patient Inquiries For A Hospital

AI has been killing it as of recent when it comes to automating repetitive tasks in businesses, and I've been even more fascinated by how AI voice agents have been impacting various industries. I recently came across a case study about a voice agent that helped a hospital with appointment scheduling, cost reduction and much more. Motivated by the potential of this technology, I decided to build a similar system to see how it could be adapted for other industries. I've added the case study below so that you could see the direct impact this technology is having and how fast it is advancing in todays world. Case Study A multi-specialty hospital was facing a range of operational challenges such as high administrative load, limited 24/7 availability, high operation costs, patient follow ups, answering routine questions and long call wait times. Solution To solve these problems, the hospital implemented an AI voice agent capable of handling various aspects of patient interaction and operations such as: Automated Appointment Scheduling: AI agents seamlessly handled patient appointments, rescheduling, and cancellations. This reduced manual effort by 75%, increased appointment adherence by 30%, and allowed patients to reschedule with ease. 24/7 Multilingual Patient Support: The AI agents utilized advanced Natural Language Processing (NLP) to communicate in six languages. This feature eased communication barriers, leading to a significant boost in guest satisfaction. Handling Patient Inquiries: AI agents answered FAQs about hospital services, procedures, insurance, and general health queries with speed and accuracy, improving the overall patient experience. This reduced the burden on front-desk staff by 60%. Proactive Patient Follow-Ups: The Voice AI agents automated follow-up calls for patients post-treatment, providing reminders for medication, check-ups, and future appointments, improving patient engagement and adherence to treatment plans. Enhanced Call Routing: AI agents routed patient calls based on specific needs without requiring additional staff. This eliminated long waits, improved call response times by 60%, and allowed staff to focus on more critical tasks. Elimination of IVR Systems: The hospital replaced outdated touch-tone IVRs with AI agents that routed calls efficiently without requiring patients to wait in long queues or be transferred among departments. This resulted in a 55% reduction in average call-handling times. Outcome The adoption of AI agents resulted in measurable improvements across various operational and patient care metrics: The hospital achieved a 55% reduction in operational costs by decreasing reliance on human agents for routine tasks and minimizing the need for additional staff. Patient satisfaction scores improved by 35% as a result of faster response times, personalized communication, and proactive patient engagement. Automation of appointment scheduling, follow-ups, and call routing increased overall operational efficiency by 75%. The AI agents supported 12 languages which bridged communication gaps with non-English speaking patients, further enhancing the patient experience. The AI agents reduced call center wait times by 60%, significantly improving patient support and reducing frustration. Appointment reminders and follow-up messages sent by AI agents contributed to a 30% reduction in missed appointments By implementing the AI voice agent, the hospital business enhanced its customer communication and scheduling, while significantly reducing operational costs. I’d love to hear some of your thoughts on this technology and how you see it impacting your and/or other industries.

I run an AI automation agency (AAA). My honest overview and review of this new business model
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

I run an AI automation agency (AAA). My honest overview and review of this new business model

I started an AI tools directory in February, and then branched off that to start an AI automation agency (AAA) in June. So far I've come across a lot of unsustainable "ideas" to make money with AI, but at the same time a few diamonds in the rough that aren't fully tapped into yet- especially the AAA model. Thought I'd share this post to shine light into this new business model and share some ways you could potentially start your own agency, or at the very least know who you are dealing with and how to pick and choose when you (inevitably) get bombarded with cold emails from them down the line. Foreword Running an AAA does NOT involve using AI tools directly to generate and sell content directly. That ship has sailed, and unless you are happy with $5 from Fiverr every month or so, it is not a real business model. Cry me a river but generating generic art with AI and slapping it onto a T-shirt to sell on Etsy won't make you a dime. At the same time, the AAA model will NOT require you to have a deep theoretical knowledge of AI, or any academic degree, as we are more so dealing with the practical applications of generative AI and how we can implement these into different workflows and tech-stacks, rather than building AI models from the ground up. Regardless of all that, common sense and a willingness to learn will help (a shit ton), as with anything. Keep in mind - this WILL involve work and motivation as well. The mindset that AI somehow means everything can be done for you on autopilot is not the right way to approach things. The common theme of businesses I've seen who have successfully implemented AI into their operations is the willingess to work with AI in a way that augments their existing operations, rather than flat out replace a worker or team. And this is exactly the train of thought you need when working with AI as a business model. However, as the field is relatively unsaturated and hype surrounding AI is still fresh for enterprises, right now is the prime time to start something new if generative AI interests you at all. With that being said, I'll be going over three of the most successful AI-adjacent businesses I've seen over this past year, in addition to some tips and resources to point you in the right direction. so.. WTF is an AI Automation Agency? The AI automation agency (or as some YouTubers have coined it, the AAA model) at its core involves creating custom AI solutions for businesses. I have over 1500 AI tools listed in my directory, however the feedback I've received from some enterprise users is that ready-made SaaS tools are too generic to meet their specific needs. Combine this with the fact virtually no smaller companies have the time or skills required to develop custom solutions right off the bat, and you have yourself real demand. I would say in practice, the AAA model is quite similar to Wordpress and even web dev agencies, with the major difference being all solutions you develop will incorporate key aspects of AI AND automation. Which brings me to my second point- JUST AI IS NOT ENOUGH. Rather than reducing the amount of time required to complete certain tasks, I've seen many AI agencies make the mistake of recommending and (trying to) sell solutions that more likely than not increase the workload of their clients. For example, if you were to make an internal tool that has AI answer questions based on their knowledge base, but this knowledge base has to be updated manually, this is creating unnecessary work. As such I think one of the key components of building successful AI solutions is incorporating the new (Generative AI/LLMs) with the old (programmtic automation- think Zapier, APIs, etc.). Finally, for this business model to be successful, ideally you should target a niche in which you have already worked and understand pain points and needs. Not only does this make it much easier to get calls booked with prospects, the solutions you build will have much greater value to your clients (meaning you get paid more). A mistake I've seen many AAA operators make (and I blame this on the "Get Rich Quick" YouTubers) is focusing too much on a specific productized service, rather than really understanding the needs of businesses. The former is much done via a SaaS model, but when going the agency route the only thing that makes sense is building custom solutions. This is why I always take a consultant-first approach. You can only build once you understand what they actually need and how certain solutions may impact their operations, workflows, and bottom-line. Basics of How to Get Started Pick a niche. As I mentioned previously, preferably one that you've worked in before. Niches I know of that are actively being bombarded with cold emails include real estate, e-commerce, auto-dealerships, lawyers, and medical offices. There is a reason for this, but I will tell you straight up this business model works well if you target any white-collar service business (internal tools approach) or high volume businesses (customer facing tools approach). Setup your toolbox. If you wanted to start a pressure washing business, you would need a pressure-washer. This is no different. For those without programming knowledge, I've seen two common ways AAA get setup to build- one is having a network of on-call web developers, whether its personal contacts or simply going to Upwork or any talent sourcing agency. The second is having an arsenal of no-code tools. I'll get to this more in a second, but this works beecause at its core, when we are dealing with the practical applications of AI, the code is quite simple, simply put. Start cold sales. Unless you have a network already, this is not a step you can skip. You've already picked a niche, so all you have to do is find the right message. Keep cold emails short, sweet, but enticing- and it will help a lot if you did step 1 correctly and intimately understand who your audience is. I'll be touching base later about how you can leverage AI yourself to help you with outreach and closing. The beauty of gen AI and the AAA model You don't need to be a seasoned web developer to make this business model work. The large majority of solutions that SME clients want is best done using an API for an LLM for the actual AI aspect. The value we create with the solutions we build comes with the conceptual framework and design that not only does what they need it to but integrates smoothly with their existing tech-stack and workflow. The actual implementation is quite straightforward once you understand the high level design and know which tools you are going to use. To give you a sense, even if you plan to build out these apps yourself (say in Python) the large majority of the nitty gritty technical work has already been done for you, especially if you leverage Python libraries and packages that offer high level abstraction for LLM-related functions. For instance, calling GPT can be as little as a single line of code. (And there are no-code tools where these functions are simply an icon on a GUI). Aside from understanding the capabilities and limitations of these tools and frameworks, the only thing that matters is being able to put them in a way that makes sense for what you want to build. Which is why outsourcing and no-code tools both work in our case. Okay... but how TF am I suppposed to actually build out these solutions? Now the fun part. I highly recommend getting familiar with Langchain and LlamaIndex. Both are Python libraires that help a lot with the high-level LLM abstraction I mentioned previously. The two most important aspects include being able to integrate internal data sources/knowledge bases with LLMs, and have LLMs perform autonomous actions. The two most common methods respectively are RAG and output parsing. RAG (retrieval augmented Generation) If you've ever seen a tool that seemingly "trains" GPT on your own data, and wonder how it all works- well I have an answer from you. At a high level, the user query is first being fed to what's called a vector database to run vector search. Vector search basically lets you do semantic search where you are searching data based on meaning. The vector databases then retrieves the most relevant sections of text as it relates to the user query, and this text gets APPENDED to your GPT prompt to provide extra context to the AI. Further, with prompt engineering, you can limit GPT to only generate an answer if it can be found within this extra context, greatly limiting the chance of hallucination (this is where AI makes random shit up). Aside from vector databases, we can also implement RAG with other data sources and retrieval methods, for example SQL databses (via parsing the outputs of LLM's- more on this later). Autonomous Agents via Output Parsing A common need of clients has been having AI actually perform tasks, rather than simply spitting out text. For example, with autonomous agents, we can have an e-commerce chatbot do the work of a basic customer service rep (i.e. look into orders, refunds, shipping). At a high level, what's going on is that the response of the LLM is being used programmtically to determine which API to call. Keeping on with the e-commerce example, if I wanted a chatbot to check shipping status, I could have a LLM response within my app (not shown to the user) with a prompt that outputs a random hash or string, and programmatically I can determine which API call to make based on this hash/string. And using the same fundamental concept as with RAG, I can append the the API response to a final prompt that would spit out the answer for the user. How No Code Tools Can Fit In (With some example solutions you can build) With that being said, you don't necessarily need to do all of the above by coding yourself, with Python libraries or otherwise. However, I will say that having that high level overview will help IMMENSELY when it comes to using no-code tools to do the actual work for you. Regardless, here are a few common solutions you might build for clients as well as some no-code tools you can use to build them out. Ex. Solution 1: AI Chatbots for SMEs (Small and Medium Enterprises) This involves creating chatbots that handle user queries, lead gen, and so forth with AI, and will use the principles of RAG at heart. After getting the required data from your client (i.e. product catalogues, previous support tickets, FAQ, internal documentation), you upload this into your knowledge base and write a prompt that makes sense for your use case. One no-code tool that does this well is MyAskAI. The beauty of it especially for building external chatbots is the ability to quickly ingest entire websites into your knowledge base via a sitemap, and bulk uploading files. Essentially, they've covered the entire grunt work required to do this manually. Finally, you can create a inline or chat widget on your client's website with a few lines of HTML, or altneratively integrate it with a Slack/Teams chatbot (if you are going for an internal Q&A chatbot approach). Other tools you could use include Botpress and Voiceflow, however these are less for RAG and more for building out complete chatbot flows that may or may not incorporate LLMs. Both apps are essentially GUIs that eliminate the pain and tears and trying to implement complex flows manually, and both natively incoporate AI intents and a knowledge base feature. Ex. Solution 2: Internal Apps Similar to the first example, except we go beyond making just chatbots but tools such as report generation and really any sort of internal tool or automations that may incorporate LLM's. For instance, you can have a tool that automatically generates replies to inbound emails based on your client's knowledge base. Or an automation that does the same thing but for replies to Instagram comments. Another example could be a tool that generates a description and screeenshot based on a URL (useful for directory sites, made one for my own :P). Getting into more advanced implementations of LLMs, we can have tools that can generate entire drafts of reports (think 80+ pages), based not only on data from a knowledge base but also the writing style, format, and author voice of previous reports. One good tool to create content generation panels for your clients would be MindStudio. You can train LLM's via prompt engineering in a structured way with your own data to essentially fine tune them for whatever text you need it to generate. Furthermore, it has a GUI where you can dictate the entire AI flow. You can also upload data sources via multiple formats, including PDF, CSV, and Docx. For automations that require interactions between multiple apps, I recommend the OG zapier/make.com if you want a no-code solution. For instance, for the automatic email reply generator, I can have a trigger such that when an email is received, a custom AI reply is generated by MyAskAI, and finally a draft is created in my email client. Or, for an automation where I can create a social media posts on multiple platforms based on a RSS feed (news feed), I can implement this directly in Zapier with their native GPT action (see screenshot) As for more complex LLM flows that may require multiple layers of LLMs, data sources, and APIs working together to generate a single response i.e. a long form 100 page report, I would recommend tools such as Stack AI or Flowise (open-source alternative) to build these solutions out. Essentially, you get most of the functions and features of Python packages such as Langchain and LlamaIndex in a GUI. See screenshot for an example of a flow How the hell are you supposed to find clients? With all that being said, none of this matters if you can't find anyone to sell to. You will have to do cold sales, one way or the other, especially if you are brand new to the game. And what better way to sell your AI services than with AI itself? If we want to integrate AI into the cold outreach process, first we must identify what it's good at doing, and that's obviously writing a bunch of text, in a short amount of time. Similar to the solutions that an AAA can build for its clients, we can take advantage of the same principles in our own sales processes. How to do outreach Once you've identified your niche and their pain points/opportunities for automation, you want to craft a compelling message in which you can send via cold email and cold calls to get prospects booked on demos/consultations. I won't get into too much detail in terms of exactly how to write emails or calling scripts, as there are millions of resources to help with this, but I will tell you a few key points you want to keep in mind when doing outreach for your AAA. First, you want to keep in mind that many businesses are still hesitant about AI and may not understand what it really is or how it can benefit their operations. However, we can take advantage of how mass media has been reporting on AI this past year- at the very least people are AWARE that sooner or later they may have to implement AI into their businesses to stay competitive. We want to frame our message in a way that introduces generative AI as a technology that can have a direct, tangible, and positive impact on their business. Although it may be hard to quantify, I like to include estimates of man-hours saved or costs saved at least in my final proposals to prospects. Times are TOUGH right now, and money is expensive, so you need to have a compelling reason for businesses to get on board. Once you've gotten your messaging down, you will want to create a list of prospects to contact. Tools you can use to find prospects include Apollo.io, reply.io, zoominfo (expensive af), and Linkedin Sales Navigator. What specific job titles, etc. to target will depend on your niche but for smaller companies this will tend to be the owner. For white collar niches, i.e. law, the professional that will be directly benefiting from the tool (i.e. partners) may be better to contact. And for larger organizations you may want to target business improvement and digital transformation leads/directors- these are the people directly in charge of projects like what you may be proposing. Okay- so you have your message, and your list, and now all it comes down to is getting the good word out. I won't be going into the details of how to send these out, a quick Google search will give you hundreds of resources for cold outreach methods. However, personalization is key and beyond simple dynamic variables you want to make sure you can either personalize your email campaigns directly with AI (SmartWriter.ai is an example of a tool that can do this), or at the very least have the ability to import email messages programmatically. Alternatively, ask ChatGPT to make you a Python Script that can take in a list of emails, scrape info based on their linkedin URL or website, and all pass this onto a GPT prompt that specifies your messaging to generate an email. From there, send away. How tf do I close? Once you've got some prospects booked in on your meetings, you will need to close deals with them to turn them into clients. Call #1: Consultation Tying back to when I mentioned you want to take a consultant-first appraoch, you will want to listen closely to their goals and needs and understand their pain points. This would be the first call, and typically I would provide a high level overview of different solutions we could build to tacke these. It really helps to have a presentation available, so you can graphically demonstrate key points and key technologies. I like to use Plus AI for this, it's basically a Google Slides add-on that can generate slide decks for you. I copy and paste my default company messaging, add some key points for the presentation, and it comes out with pretty decent slides. Call #2: Demo The second call would involve a demo of one of these solutions, and typically I'll quickly prototype it with boilerplate code I already have, otherwise I'll cook something up in a no-code tool. If you have a niche where one type of solution is commonly demanded, it helps to have a general demo set up to be able to handle a larger volume of calls, so you aren't burning yourself out. I'll also elaborate on how the final product would look like in comparison to the demo. Call #3 and Beyond: Once the initial consultation and demo is complete, you will want to alleviate any remaining concerns from your prospects and work with them to reach a final work proposal. It's crucial you lay out exactly what you will be building (in writing) and ensure the prospect understands this. Furthermore, be clear and transparent with timelines and communication methods for the project. In terms of pricing, you want to take this from a value-based approach. The same solution may be worth a lot more to client A than client B. Furthermore, you can create "add-ons" such as monthly maintenance/upgrade packages, training sessions for employeees, and so forth, separate from the initial setup fee you would charge. How you can incorporate AI into marketing your businesses Beyond cold sales, I highly recommend creating a funnel to capture warm leads. For instance, I do this currently with my AI tools directory, which links directly to my AI agency and has consistent branding throughout. Warm leads are much more likely to close (and honestly, much nicer to deal with). However, even without an AI-related website, at the very least you will want to create a presence on social media and the web in general. As with any agency, you will want basic a professional presence. A professional virtual address helps, in addition to a Google Business Profile (GBP) and TrustPilot. a GBP (especially for local SEO) and Trustpilot page also helps improve the looks of your search results immensely. For GBP, I recommend using ProfilePro, which is a chrome extension you can use to automate SEO work for your GBP. Aside from SEO optimzied business descriptions based on your business, it can handle Q/A answers, responses, updates, and service descriptions based on local keywords. Privacy and Legal Concerns of the AAA Model Aside from typical concerns for agencies relating to service contracts, there are a few issues (especially when using no-code tools) that will need to be addressed to run a successful AAA. Most of these surround privacy concerns when working with proprietary data. In your terms with your client, you will want to clearly define hosting providers and any third party tools you will be using to build their solution, and a DPA with these third parties listed as subprocessors if necessary. In addition, you will want to implement best practices like redacting private information from data being used for building solutions. In terms of addressing concerns directly from clients, it helps if you host your solutions on their own servers (not possible with AI tools), and address the fact only ChatGPT queries in the web app, not OpenAI API calls, will be used to train OpenAI's models (as reported by mainstream media). The key here is to be open and transparent with your clients about ALL the tools you are using, where there data will be going, and make sure to get this all in writing. have fun, and keep an open mind Before I finish this post, I just want to reiterate the fact that this is NOT an easy way to make money. Running an AI agency will require hours and hours of dedication and work, and constantly rearranging your schedule to meet prospect and client needs. However, if you are looking for a new business to run, and have a knack for understanding business operations and are genuinely interested in the pracitcal applications of generative AI, then I say go for it. The time is ticking before AAA becomes the new dropshipping or SMMA, and I've a firm believer that those who set foot first and establish themselves in this field will come out top. And remember, while 100 thousand people may read this post, only 2 may actually take initiative and start.

Looking for Social Media Marketing Partner(s) for High-Potential AI App Business
reddit
LLM Vibe Score0
Human Vibe Score1
Altruistic-Flan-8222This week

Looking for Social Media Marketing Partner(s) for High-Potential AI App Business

Hello everyone! I am Mak, and I'm a software engineer and AI developer with a few years of experience. I'm pretty young like the most of you and have an amazing idea. I'm sure that some of you have heard of Rizz, Plug, Wigman and similar apps. Those are simple AI apps that generate pickup lines for people, and I worked as an AI developer for one of the above. I got this business idea after analyzing more about this industry and realizing that these apps make TONS of money—like the one I worked for, which is making about $50k per WEEK using my AI solutions. That's crazy. The point is that I took a pause from working as a software engineer for clients and researched how to do the same thing. It took me a few months to actually understand everything about this business model, and Rizz apps are just one example of this type of business. There is one 17 yo guy I found who made "Cal AI" I guess, basically a simple AI app that analyzes your meal and provides info like calories, etc. I also created AI solutions for a guy who made an AI app that analyzes your face, provides Sigma analytics, and suggests how to improve your face, etc. So the point is that there are tons of AI app ideas that you can create for this industry. And the important fact is that the AI market is growing. Some important AI analytics say that in 2024, there were 1.5B AI app downloads, and mobile AI app consumer spending was $1.8B. That's huge. So, what am I looking for? I need someone, hopefully from the US, or someone who knows how to post social media content for US users, to help me out with my business idea. I'm self-funded and have already spent a lot on important requirements and equipment, which is why I need someone interested in revenue sharing. We can come up with a deal such as capped/tiered revenue share, profit share, deferred model, etc. We could discuss this privately since everyone has different experience levels and thoughts about this. Also, since I'm talking about experience, you don't need huge experience at all. You can be 16-25 years old just like me and only have marketing skills. However, to make it easier for those who don't have marketing skills, I am planning to create code that will automatically generate content for you, and all you need to do is post the content. But this is only for posting content without creating it and is for interested people from the US since I need US customers. However, if you have marketing skills and an idea for getting organic US views, please let's talk. Short info about my app: It is an AI app like the previous examples, which doesn’t yet exist. There is pretty big potential for app growth (60% of Americans could use this app), and it should be pretty easy to market. Good niche, good idea and overall solid market for this app idea. TL;DR I need someone interested in marketing my AI app in exchange for revenue share. No huge experience is needed. I would prefer someone from the US. If you are interested, feel free to contact me here on Reddit via private messages or below. We can talk here, on Discord, LinkedIn, or anywhere you prefer. Thanks once again!

AI Voice Platform Comparison for Small Business Use Cases
reddit
LLM Vibe Score0
Human Vibe Score1
Glad-Syllabub6777This week

AI Voice Platform Comparison for Small Business Use Cases

We provide AI voice agent consultation and solutions in Upwork. One of clients’ frequent questions is which platform is best/perfect for their use cases, like lead qualification, AI receptionist, customer support, etc. This post provides our thoughts on this question. Our overall feeling is that the AI agent technology is still not there yet. It seems close but there are many corner cases the AI bot doesn't handle well. Four major players in the AI voice platforms: Bland ($65M funding) Retell ($4.6M funding) Synthflow ($7.4M funding) VAPI ($20M funding) We will only talk about Bland, Retell and VAPI. We firstly tried Synthflow and found the UI was buggy (the prompt editor froze for 20 seconds to 30 seconds when we were editing the prompt). Currently we don't use it anymore. Recommended use cases based on Upwork jobs we delivered: Bland. We recommend Bland for lead qualification as the lead qualification has a strict conversation flow (like asking questions, extracting variables, and making webhook calls). Clients/contractors can draw flow diagrams to build AI voice agents. We also find Bland is not a good fit for a small business with a monthly budget less than 5K. The reason is that common tools (like warm transfer, SMS sending) for AI voice agents are only available to enterprise clients. But warm transfers are critical for small businesses. Retell. We recommend Retell for customer support in contact centers. Retell has the best voice among competitors. One use case we build in Retell is the live translator in the ambulance call center. We tried the same prompt with the same LLM setup in VAPI. We found Retell performs way better than VAPI in terms of the translation quality and reliability. Another common scenario in the customer support domain is to have 3-way merge so that the agent can tell the summary to the transfer number while the caller can hear the conversation. VAPI. We recommend VAPI for AI receptionists and phone answering use cases. We can write a prompt and ask LLM to do the magic if callers ask questions not included in the prompt. We can set up custom tools to trigger automation (like update CRM) and warm transfer to connect to the stakeholders. One feeling we have is that VAPI is way more complicated than the other two platforms. If you don’t have developer experience and have a budget to hire a contractor, it is better to try Retell as Retell has many integrations. If you have any other questions or we miss anything, feel free to comment. We like to explore AI voice agent space together.

How I went from $27 to $3K as a solopreneur still in a 9-5
reddit
LLM Vibe Score0
Human Vibe Score1
jottrledThis week

How I went from $27 to $3K as a solopreneur still in a 9-5

My journey started back in November 2023. I was scrolling through Twitter and YouTube and saw a word that I had never come across before. Solopreneur. The word caught my eye. Mainly because I was pretty sure I knew what it meant even though it's not a word you'll find in the dictionary. I liked what it was describing. A solo entrepreneur. A one man business. It completely resonated with me. As a software engineer by trade I'm used to working alone, especially since the pandemic hit and we were forced to work remotely. See, I always wanted to ditch the 9-5 thing but thought that was too big and too scary for a single person to do. Surely you would need a lot of money to get started, right? Surely you would need investors? The whole concept seemed impossible to me. That was until I found all the success stories. I became obsessed with the concept of solopreneurship. As I went further down the rabbit hole I found people like Justin Welsh, Kieran Drew and Marc Louvion to name a few. All of whom have one person businesses making huge money every year. So I thought, if they can do it, why can't I? People like this have cleared the pathway for those looking to escape the 9-5 grind. I decided 2024 would be the year I try this out. My main goal for the year? Build a one man business, earn my first $ online and learn a sh\*t ton along the way. My main goal in general? Build my business to $100K per year, quit my 9-5 and live with freedom. From December 2023 to February 2024 I began brainstorming ideas. I was like a lost puppy looking for his ball. How on earth did people find good ideas? I began writing everything and anything that came to mind down in my notes app on my phone. By February I would have approximately 70 ideas. Each as weird and whacky as the other. I was skeptical though. If I went through all the trouble of building a product for one of these ideas how would I know if anyone would even be interested in using it? I got scared and took a break for a week. All these ideas seemed too big and the chance that they would take off into the atmosphere was slim (in my mind anyways). I was learning more and more about solopreneurship as the weeks went on so I decided to build a product centered around everything I was learning about. The idea was simple. Enter a business idea and use AI to give the user details about how to market it, who their target customers were, what to write on their landing page, etc. All for a measly $27 per use. I quickly built it and launched on March 3rd 2024. I posted about it on Indie Hackers, Reddit and Hacker News. I was so excited about the prospect of earning my first internet $! Surely everyone wanted to use my product! Nope...all I got was crickets. I was quickly brought back down to earth. That was until 5 days later. I looked at my phone and had a new Stripe notification! Cha-ching! My first internet $. What a feeling! That was goal number 1 complete. It would be another 6 days before I would get my second sale...and then another 15 days to get my third. It was an emotional rollercoaster. I went from feeling like quitting the 9-5 was actually possible to thinking that maybe the ups and downs aren't worth it. On one hand I had made my first internet dollar so I should my ecstatic, and don't get me wrong, I was but I wanted more. More validation that I could do this long term. By May I was starting to give up on the product. I had learned so much in the past few months about marketing, SEO, building an audience, etc. and I wanted to build something that I thought could have more success so I focused on one critical thing that I had learned about. What was it? Building a product that had SEO potential. A product that I knew hundreds of people were looking for. See this was my thinking - If I could find a keyword that people were searching for on Google hundreds/thousands of times every month and it was easy to rank high on search engines then I would go all in (in SEO land this equates to a Keyword that has a Keyword Difficulty of = 500). I began researching and found that the keyword "micro saas ideas" was being searched for around 600 times each month. Micro Saas was something that really interested me. It was perfect for solopreneurs. Small software products that 1 person could build. What's not to like if you're in the game of software and solopreneurship? Researching keywords like this became like a game for me. I was hooked. I was doing it every day, finding gems that were being searched for hundreds and thousands of times every month that still had potential. That's when I came up with my next product idea. I decided to create a database of Micro Saas Ideas all with this sort of SEO potential. See if you can build a product that you know people are looking for then that's all the validation you need. So I put this theory to the test. I created a database of Micro Saas Ideas with SEO Potential and launched it in June 2024. This time it was different. I made $700 in the first week of launching. A large contrast to my previous failed attempt at becoming the worlds greatest solopreneur. Since launch I have grown the product to $3K and I couldn't be happier. I know what you're saying, $3K isn't a lot. But it's validation. It's validation that I can earn $ online. Validation that I can grow a business and it gives me hope that one day I'll be able to quit that 9-5 grind. My plan is to keep growing the business. I expect there to be a few challenges up ahead but I'll tackle them as I go and learn from the failures and successes. I have a newsletter where I share Micro Saas Ideas with SEO potential every week which I'll leave below in the first comment. Feel free to come along for the ride. If not I hope this post brings you some value If you're thinking about starting as a solopreneur, stop thinking and start doing, you won't regret it.

This founder was about to shut down his startup and open a restaurant. He pivoted the business and grew it to $45m ARR in 12 months. What else have you seen grow that fast?
reddit
LLM Vibe Score0
Human Vibe Score1
CountryPitifulThis week

This founder was about to shut down his startup and open a restaurant. He pivoted the business and grew it to $45m ARR in 12 months. What else have you seen grow that fast?

I heard that Jasper scaled to $45m ARR in 12 months...with a team of 8. For context, they are one of the fastest-growing companies ever. Grew from $0 to $45m ARR in 12 months (then raised $125m at a $1.5b valuation). As a fellow founder, their story is really inspiring to me (curious about what others think): In December 2020, Dave Rogenmoser and his co-founders were on the brink of shutting down their business. They'd spent 3+ years building a conversion optimization software called Proof...and it was flatlining. A few weeks prior they had to make the painful decision to let go of half their team. Competition and churn had completely eroded growth. Things were painful. 8 years of work left them with a string of startups that never quite made it: 2 failed software businesses (couldn't make money*) A SMB marketing agency (maxed out at $25k/mo*) An online course company (hard to get big*) The Pivot: In January 2021, they had an idea to use Chat GPT-3, the generative AI model released 6 months earlier, to write high-converting Facebook ads. Within 30 days, they launched the business. With the skeleton crew remaining from the last startup, they scaled the business to $45m ARR and 70,000+ customers without hiring a single new person. Soon after, they raised $125m at a $1.5b valuation. Dave Rogenmoser, CEO at Jasper, had some great one-liners in a few podcasts I listened to on the business. Here are some of his learnings: Right Skill, Wrong Vehicle: He spent 8 years building marketing businesses which gave this team the knowledge and confidence to spend $1m/mo on sales and marketing to scale the business to $45m ARR in year 1. Launch Fast & Iterate Quickly: The team agreed that if the business didn't work in 30 days, they'd shut it down. Dave says, "If you have been working on a problem for more than 18 months and haven't found Product market fit (PMF), odds are you won't...Make the hard pivot."* Ride A Big Wave: Generative AI technology is a new technology that is changing the way we work. But it's not just text. It's images, voice, etc. Identify new customer segments (e.g., Municipalities, Banks, Lawyers, etc.), learn their problems, and apply this novel technology to solve them. What other businesses have you seen scale like this? I've never seen a SaaS business grow that fast. I meet interesting founders 2x per week and share the learnings here.

Building Business Development/Sales Pipeline
reddit
LLM Vibe Score0
Human Vibe Score1
Nevoy_92This week

Building Business Development/Sales Pipeline

Hey all! Happy weekend wherever you may be! Wanting to get some advice and insight into a couple areas as mentioned in the title. Background is the following: My Partner and I started our company about in 2021. When we kicked off we were building a control and camera vision system for automating and optimizing indoor vertical farms. We got to early mvp but market was not as big and barrier to entry was high. So we pivoted early 2023 to utilize components of our technology in a wildfire detection and risk analysis platform. Happy to say we are once again at MVP but need to get PMF and pipeline going both with revenue generating clients and pilots/demos. Through this period we’ve kept the lights on by running a consulting service and digital agency. We’ve also pushed out a couple of AI tools to market. Effectively I need to build out a strong pipeline for each vertical and associated sales team. Right now spread too thin trying to conduct sales and business dev on each front. Challenges: Wildfire: Business to Gov relationships so need to build for that. Additionally early stage technology so imo relationships are critical. Additionally need to take advantage of grant funding. Target Markets: Canada, USA, Mediterranean, Northern Europe/Scandavian Countries. Consulting and Agency: Things feel dry… we have a recurring client list but we want to grow this channel exponentially, focusing on RFP’s and med to large company profiles rather then the current SMB. Our current activités are mediocre imo for outreach and connection. AI Tools: I believe these are great opportunities. TLDR 1)sales based assistant as well as 2)central AI aggregation with prompt repository. Business Dev Energy into this is basically focused on digital means. In the process of generating video content to push via ads and online social platforms. Challenge: low engagement right now users signing up but no commitments to purchase. Need to evaluate value offer and feedback on PMF. From the sales team side, effectively need to generate the sales so I can expand the team and grow accordingly. I’m a huge proponent of commission based compensation. Also open to a base salary. However anyone I onboard at this moment would have to be commission cause cash-flow. On that front, what are current commissions structures looking like for people? What’s engaging what’s worth taking a risk what is just a huge no? On the challenges for the product lines any feedback questions and even poking holes is appreciated! Thanks!

This founder was about to shut down his business and open a restaurant. He pivoted the business and grew it to $45m ARR in 12 months. What other businesses can scale like this?
reddit
LLM Vibe Score0
Human Vibe Score1
CountryPitifulThis week

This founder was about to shut down his business and open a restaurant. He pivoted the business and grew it to $45m ARR in 12 months. What other businesses can scale like this?

I heard that Jasper scaled to $45m ARR in 12 months...with a team of 8. For context, they are one of the fastest-growing companies ever. Grew from $0 to $45m ARR in 12 months (then raised $125m at a $1.5b valuation). As a fellow founder, their story is really inspiring to me (curious about what others think): In December 2020, Dave Rogenmoser and his co-founders were on the brink of shutting down their business. They'd spent 3+ years building a conversion optimization software called Proof...and it was flatlining. A few weeks prior they had to make the painful decision to let go of half their team. Competition and churn had completely eroded growth. Things were painful. 8 years of work left them with a string of startups that never quite made it: 2 failed software businesses (couldn't make money*) A SMB marketing agency (maxed out at $25k/mo*) An online course company (hard to get big*) The Pivot: In January 2021, they had an idea to use Chat GPT-3, the generative AI model released 6 months earlier, to write high-converting Facebook ads. Within 30 days, they launched the business. With the skeleton crew remaining from the last startup, they scaled the business to $45m ARR and 70,000+ customers without hiring a single new person. Soon after, they raised $125m at a $1.5b valuation. Dave Rogenmoser, CEO at Jasper, had some great one-liners in a few podcasts I listened to on the business. Here are some of his learnings: Right Skill, Wrong Vehicle: He spent 8 years building marketing businesses which gave this team the knowledge and confidence to spend $1m/mo on sales and marketing to scale the business to $45m ARR in year 1. Launch Fast & Iterate Quickly: The team agreed that if the business didn't work in 30 days, they'd shut it down. Dave says, "If you have been working on a problem for more than 18 months and haven't found Product market fit (PMF), odds are you won't...Make the hard pivot."* Ride A Big Wave: Generative AI technology is a new technology that is changing the way we work. But it's not just text. It's images, voice, etc. Identify new customer segments (e.g., Municipalities, Banks, Lawyers, etc.), learn their problems, and apply this novel technology to solve them. What other businesses have you seen scale like this? I've never seen a SaaS business grow that fast. I meet interesting founders 2x per week and share the learnings here.

How Our AI Tool Helped a Small Business Save 15% on Annual Expenses
reddit
LLM Vibe Score0
Human Vibe Score1
Medical-Wait-6960This week

How Our AI Tool Helped a Small Business Save 15% on Annual Expenses

I’m the founder of a startup that built an AI-powered tool to analyze and optimize business finances, with a special focus on small and medium-sized enterprises (SMEs). After months of development and testing, I’m pumped to share our solution with you and get your feedback. Here’s what we do, how it works, and the results we’ve seen. The Problem We Solve Managing a company’s finances, especially for an SME, is often a nightmare: forgotten subscriptions, poorly negotiated supplier contracts, invoices with errors… We’ve all been there. Our tool uses AI to automate expense analysis, spot issues, and suggest practical ways to cut costs—without you having to spend hours on it. How It Works (A Bit of Tech Talk) We built our tool on a multi-agent architecture using the CREWAI framework. Here are the main AI agents we’ve got running: Expense Analyst: Digs through your invoices and categorizes your spending. Compliance Auditor: Checks for errors, fraud, or compliance hiccups. Financial Reporter: Generates clear reports with actionable recommendations. Supplier Negotiator: Hunts down cheaper supplier options using the Serper API and offers negotiation strategies. To hook up your company’s data, we use NEEDLE, a RAG (Retrieval-Augmented Generation) system that lets our agents tap into your info in real time. Everything’s locked down in an SQLite database with end-to-end encryption. Real Results We tested the tool with 10 companies, and here’s what we found: Average cost reduction of 12% in three months. Fraud detection: For example, we flagged 5 shady invoices at one company, saving them €3,000. Supplier optimization: For an SME, we found an energy supplier 20% cheaper, saving them €8,000 a year. A real-world case: A consulting firm with 50 employees ran our tool on their SaaS subscriptions. Outcome? They ditched 3 unused subscriptions, renegotiated 2 contracts, and saved 15% on their annual expenses. Challenges We Tackled No sugarcoating here—it wasn’t a walk in the park. The biggest hurdle? Data security. We’re handling sensitive stuff, so we went all in: End-to-end encryption for everything we process. GDPR compliance with strict rules. Role-based access controls to limit who sees what. Another tough one was integrating with existing systems. We’ve already got connectors for QuickBooks, Xero, and SAP, and we’re working on more. Why It’s Different Sure, there are tools like Expensify or Ramp out there, but our multi-agent approach digs deeper. We deliver super-detailed analysis and precise recommendations. And our knack for finding cheaper suppliers in real time? That’s a game-changer for quick savings.I’m the founder of a startup that built an AI-powered tool to analyze and optimize business finances, with a special focus on small and medium-sized enterprises (SMEs). After months of development and testing, I’m pumped to share our solution with you and get your feedback. Here’s what we do, how it works, and the results we’ve seen. Ask me your technical questions, share your ideas or critiques we’re here to get better! Thanks you for reading this.

Best AI tools to help company productivity?
reddit
LLM Vibe Score0
Human Vibe Score1
Significant_Stable_7This week

Best AI tools to help company productivity?

Hey guys! I recently did a big restructuring of my production company and moving away from smaller businesses ad’s and moving up to working with larger marketing agencies. My partner and I are brainstorming ways to automate or at least improve certain parts of our business as we also start to expand our team & to improve ease of labour as our turn around times tend to have to be pretty quick. The main things we’re looking to improve is in: • Sales/out reach strategy: we are constantly reaching out to new agencies in different parts of the world. I am already used to manually making a plan for each company we reach out to but it can be very time consuming. I don’t know if there is even a tool that could help with this haha. Even if it helps with pointers! • Organizing/visualizing spreadsheets: we deal with spreadsheets on what we spend per production and how we distribute our total budget per department. If there is anyway to ease the workflow for our managers and on top of that also allow us to expand easier without having to look for someone who is very efficient on excel or spending more time and money on the training. • Scheduling: We already have so much to organize day per day, im not sure if there is any tool or ai system that could help in regards to scheduling meetings, organizing priorities or even just deadlines for certain projects. Example: we need to schedule everything from pre production deadlines (meetings with talent, agency, and crew) production deadlines, & post production deadlines. I’m sure there is other small things I am missing but those are the three main things! There is just so many things i saw on the internet that are “ai powered” or “ai improved workflow” that all claim are the best or some just use chat gpt so its essentially all the same thing. I thought id ask on here to see if anyone has actually tried and could recommend some ai tools out there! Cheers,

The 15 Best (Free to Use) AI Tools for Creating Websites, Presentations, Graphics, UIs, Photos, and more
reddit
LLM Vibe Score0
Human Vibe Score1
Tapedulema919This week

The 15 Best (Free to Use) AI Tools for Creating Websites, Presentations, Graphics, UIs, Photos, and more

While we wait for ChatGPT to roll out its own official image input+output tool, I wanted to put together a list of the best AI design tools I've seen so far. Obviously text-based tasks like writing and coding get the bulk of the attention, but I wanted to see how it’s being used in design and more visual tasks. From UI and full-on website design, to graphics and photo generation, there are a ton of interesting and free tools coming out that are worth trying and using as inspiration for your own projects. These tools cover a bunch of different use cases and can hopefully help some of you, whether you’re a professional designer looking to automate parts of your work or just someone who wants to find ways to speed up the design work for your business/side projects. All of them are free to try, but most have some kind of paid plan or limit on the number of free generations. Fair enough given it costs money to run the models, but I've tried to include notes on any that don't have permanent free plans. Let me know if you know of any tools I’ve missed so I can add them to the list! I’ve grouped them by categories, to make it easier to see what each tool is capable of, then given a bit more detail under each specific tool. AI Website, Graphic and UI Generators: Framer: Describe the website you want, and Framer will create it for you. Edit and instantly publish your site from their platform. Ironically my favorite thing about Framer isn’t its AI tool. Its real advantage is its website editor which is the best I’ve seen on any platform (and usable for free). It’s like Figma if Figma let you publish directly to the web. Microsoft Designer: Generates designs based on user input for social media posts, logos, and business graphics. It’s free to use with a Microsoft account, and fairly impressive if not always consistent. If you pay a lot or spend a ton of time on design/social media content, Designer is definitely worth checking out. UIzard: Transforms text and images into design mockups, wireframes, and full user interfaces. It’s an ambitious concept, but very cool. While Framer was better for generating websites from text prompts, UIZard offers something none of the others did: taking a sketch drawing and turning it into a UI and/or wireframing. Visualizations, Graphics and Illustrations: Taskade: AI powered productivity tool to visualize your notes, projects, and tasks. Taskade lets you easily generate mind maps and other visualizations of your work, and makes use of AI in a bunch of cool ways. For example, you can generate a mind map to help you brainstorm and then ask it to expand on a certain point or even research it for you with the internet. Bing Image Creator: Generate images from natural text descriptions, powered by DALL-E. Whether you’re looking for blog illustrations, images for your site’s pages or any other purpose, it’s worth trying. AutoDraw: Autodraw is a Google Project that lets you draw something freehand with your cursor, and AutoDraw uses AI to transform it into a refined image with icons and predrawn designs, all for free in your browser. AI Presentations and Slides: Plus AI for Google Slides: AI generated slides and full-on presentations, all within Google Slides. I liked how Plus AI worked within Google Slides and made it easy to make changes to the presentation (as lets be real, no AI tool is going to generate exactly* the content and formatting you need for a serious presentation). SlidesGo: Generate slides with illustrations, images, and icons chosen by AI. SlidesGo also has their own editor to let you edit and refine the AI generated presentation. Tome: Tell Tome what you want to say to your audience, and it will create a presentation that effectively communicates it clearly and effectively. Tome actually goes beyond just presentations and has a few cool formats worth checking out that I could see being useful for salespeople and anyone who needs to pitch an idea or product at work or to clients. Product Photography: These are all fairly similar so I’ve kept the descriptions short, but it’s genuinely a pretty useful category if you run any kind of business or side hustle that needs product photos. These photos establish the professionalism of your store/brand, and all the ones I tried had genuinely impressive results that seemed much better than what I could do myself. Pebblely: AI image generator for product images in various styles and settings. 40 free images, paid after that. Booth.ai: Generates professional-quality product photos using AI, focused on furniture, fashion, and packaged goods. Stylized.ai: Generates product photos integrated into ecommerce platforms like Shopify. Miscellaneous Tools: Fronty: Converts uploaded images or drawings into HTML and CSS code using AI. It’s a bit clunky, but a cool concept nonetheless. LetsEnhance: Uses AI to enhance the resolution of images and photographs. Generally works pretty well from my experience, and gives you 10 free credits with signup. Unfortunately beyond that it is a paid product. Remove.bg: Specializes in recognizing and removing image backgrounds effectively. Doesn’t promise much, but it does the job and doesn’t require you to sign up. TL;DR/Overall favorites: These are the ones I've found the most use for in my day-to-day work. Framer: responsive website design with a full-featured editor to edit and publish your site all in one place. Free + paid plans. Taskade: visualize and automate your workflows, projects, mind maps, and more with AI powered templates. Free + paid plans. Microsoft Designer: generate social media and other marketing graphics with AI. Free to use. Plus AI: plugin for Google Slides to generate slide content, designs, and make tweaks with AI. Free + paid plans. Pebblely: professional-quality product photos in various settings and backgrounds, free to generate up to 40 images* (through you can always sign up for another account…)

I run an AI automation agency (AAA). My honest overview and review of this new business model
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

I run an AI automation agency (AAA). My honest overview and review of this new business model

I started an AI tools directory in February, and then branched off that to start an AI automation agency (AAA) in June. So far I've come across a lot of unsustainable "ideas" to make money with AI, but at the same time a few diamonds in the rough that aren't fully tapped into yet- especially the AAA model. Thought I'd share this post to shine light into this new business model and share some ways you could potentially start your own agency, or at the very least know who you are dealing with and how to pick and choose when you (inevitably) get bombarded with cold emails from them down the line. Foreword Running an AAA does NOT involve using AI tools directly to generate and sell content directly. That ship has sailed, and unless you are happy with $5 from Fiverr every month or so, it is not a real business model. Cry me a river but generating generic art with AI and slapping it onto a T-shirt to sell on Etsy won't make you a dime. At the same time, the AAA model will NOT require you to have a deep theoretical knowledge of AI, or any academic degree, as we are more so dealing with the practical applications of generative AI and how we can implement these into different workflows and tech-stacks, rather than building AI models from the ground up. Regardless of all that, common sense and a willingness to learn will help (a shit ton), as with anything. Keep in mind - this WILL involve work and motivation as well. The mindset that AI somehow means everything can be done for you on autopilot is not the right way to approach things. The common theme of businesses I've seen who have successfully implemented AI into their operations is the willingess to work with AI in a way that augments their existing operations, rather than flat out replace a worker or team. And this is exactly the train of thought you need when working with AI as a business model. However, as the field is relatively unsaturated and hype surrounding AI is still fresh for enterprises, right now is the prime time to start something new if generative AI interests you at all. With that being said, I'll be going over three of the most successful AI-adjacent businesses I've seen over this past year, in addition to some tips and resources to point you in the right direction. so.. WTF is an AI Automation Agency? The AI automation agency (or as some YouTubers have coined it, the AAA model) at its core involves creating custom AI solutions for businesses. I have over 1500 AI tools listed in my directory, however the feedback I've received from some enterprise users is that ready-made SaaS tools are too generic to meet their specific needs. Combine this with the fact virtually no smaller companies have the time or skills required to develop custom solutions right off the bat, and you have yourself real demand. I would say in practice, the AAA model is quite similar to Wordpress and even web dev agencies, with the major difference being all solutions you develop will incorporate key aspects of AI AND automation. Which brings me to my second point- JUST AI IS NOT ENOUGH. Rather than reducing the amount of time required to complete certain tasks, I've seen many AI agencies make the mistake of recommending and (trying to) sell solutions that more likely than not increase the workload of their clients. For example, if you were to make an internal tool that has AI answer questions based on their knowledge base, but this knowledge base has to be updated manually, this is creating unnecessary work. As such I think one of the key components of building successful AI solutions is incorporating the new (Generative AI/LLMs) with the old (programmtic automation- think Zapier, APIs, etc.). Finally, for this business model to be successful, ideally you should target a niche in which you have already worked and understand pain points and needs. Not only does this make it much easier to get calls booked with prospects, the solutions you build will have much greater value to your clients (meaning you get paid more). A mistake I've seen many AAA operators make (and I blame this on the "Get Rich Quick" YouTubers) is focusing too much on a specific productized service, rather than really understanding the needs of businesses. The former is much done via a SaaS model, but when going the agency route the only thing that makes sense is building custom solutions. This is why I always take a consultant-first approach. You can only build once you understand what they actually need and how certain solutions may impact their operations, workflows, and bottom-line. Basics of How to Get Started Pick a niche. As I mentioned previously, preferably one that you've worked in before. Niches I know of that are actively being bombarded with cold emails include real estate, e-commerce, auto-dealerships, lawyers, and medical offices. There is a reason for this, but I will tell you straight up this business model works well if you target any white-collar service business (internal tools approach) or high volume businesses (customer facing tools approach). Setup your toolbox. If you wanted to start a pressure washing business, you would need a pressure-washer. This is no different. For those without programming knowledge, I've seen two common ways AAA get setup to build- one is having a network of on-call web developers, whether its personal contacts or simply going to Upwork or any talent sourcing agency. The second is having an arsenal of no-code tools. I'll get to this more in a second, but this works beecause at its core, when we are dealing with the practical applications of AI, the code is quite simple, simply put. Start cold sales. Unless you have a network already, this is not a step you can skip. You've already picked a niche, so all you have to do is find the right message. Keep cold emails short, sweet, but enticing- and it will help a lot if you did step 1 correctly and intimately understand who your audience is. I'll be touching base later about how you can leverage AI yourself to help you with outreach and closing. The beauty of gen AI and the AAA model You don't need to be a seasoned web developer to make this business model work. The large majority of solutions that SME clients want is best done using an API for an LLM for the actual AI aspect. The value we create with the solutions we build comes with the conceptual framework and design that not only does what they need it to but integrates smoothly with their existing tech-stack and workflow. The actual implementation is quite straightforward once you understand the high level design and know which tools you are going to use. To give you a sense, even if you plan to build out these apps yourself (say in Python) the large majority of the nitty gritty technical work has already been done for you, especially if you leverage Python libraries and packages that offer high level abstraction for LLM-related functions. For instance, calling GPT can be as little as a single line of code. (And there are no-code tools where these functions are simply an icon on a GUI). Aside from understanding the capabilities and limitations of these tools and frameworks, the only thing that matters is being able to put them in a way that makes sense for what you want to build. Which is why outsourcing and no-code tools both work in our case. Okay... but how TF am I suppposed to actually build out these solutions? Now the fun part. I highly recommend getting familiar with Langchain and LlamaIndex. Both are Python libraires that help a lot with the high-level LLM abstraction I mentioned previously. The two most important aspects include being able to integrate internal data sources/knowledge bases with LLMs, and have LLMs perform autonomous actions. The two most common methods respectively are RAG and output parsing. RAG (retrieval augmented Generation) If you've ever seen a tool that seemingly "trains" GPT on your own data, and wonder how it all works- well I have an answer from you. At a high level, the user query is first being fed to what's called a vector database to run vector search. Vector search basically lets you do semantic search where you are searching data based on meaning. The vector databases then retrieves the most relevant sections of text as it relates to the user query, and this text gets APPENDED to your GPT prompt to provide extra context to the AI. Further, with prompt engineering, you can limit GPT to only generate an answer if it can be found within this extra context, greatly limiting the chance of hallucination (this is where AI makes random shit up). Aside from vector databases, we can also implement RAG with other data sources and retrieval methods, for example SQL databses (via parsing the outputs of LLM's- more on this later). Autonomous Agents via Output Parsing A common need of clients has been having AI actually perform tasks, rather than simply spitting out text. For example, with autonomous agents, we can have an e-commerce chatbot do the work of a basic customer service rep (i.e. look into orders, refunds, shipping). At a high level, what's going on is that the response of the LLM is being used programmtically to determine which API to call. Keeping on with the e-commerce example, if I wanted a chatbot to check shipping status, I could have a LLM response within my app (not shown to the user) with a prompt that outputs a random hash or string, and programmatically I can determine which API call to make based on this hash/string. And using the same fundamental concept as with RAG, I can append the the API response to a final prompt that would spit out the answer for the user. How No Code Tools Can Fit In (With some example solutions you can build) With that being said, you don't necessarily need to do all of the above by coding yourself, with Python libraries or otherwise. However, I will say that having that high level overview will help IMMENSELY when it comes to using no-code tools to do the actual work for you. Regardless, here are a few common solutions you might build for clients as well as some no-code tools you can use to build them out. Ex. Solution 1: AI Chatbots for SMEs (Small and Medium Enterprises) This involves creating chatbots that handle user queries, lead gen, and so forth with AI, and will use the principles of RAG at heart. After getting the required data from your client (i.e. product catalogues, previous support tickets, FAQ, internal documentation), you upload this into your knowledge base and write a prompt that makes sense for your use case. One no-code tool that does this well is MyAskAI. The beauty of it especially for building external chatbots is the ability to quickly ingest entire websites into your knowledge base via a sitemap, and bulk uploading files. Essentially, they've covered the entire grunt work required to do this manually. Finally, you can create a inline or chat widget on your client's website with a few lines of HTML, or altneratively integrate it with a Slack/Teams chatbot (if you are going for an internal Q&A chatbot approach). Other tools you could use include Botpress and Voiceflow, however these are less for RAG and more for building out complete chatbot flows that may or may not incorporate LLMs. Both apps are essentially GUIs that eliminate the pain and tears and trying to implement complex flows manually, and both natively incoporate AI intents and a knowledge base feature. Ex. Solution 2: Internal Apps Similar to the first example, except we go beyond making just chatbots but tools such as report generation and really any sort of internal tool or automations that may incorporate LLM's. For instance, you can have a tool that automatically generates replies to inbound emails based on your client's knowledge base. Or an automation that does the same thing but for replies to Instagram comments. Another example could be a tool that generates a description and screeenshot based on a URL (useful for directory sites, made one for my own :P). Getting into more advanced implementations of LLMs, we can have tools that can generate entire drafts of reports (think 80+ pages), based not only on data from a knowledge base but also the writing style, format, and author voice of previous reports. One good tool to create content generation panels for your clients would be MindStudio. You can train LLM's via prompt engineering in a structured way with your own data to essentially fine tune them for whatever text you need it to generate. Furthermore, it has a GUI where you can dictate the entire AI flow. You can also upload data sources via multiple formats, including PDF, CSV, and Docx. For automations that require interactions between multiple apps, I recommend the OG zapier/make.com if you want a no-code solution. For instance, for the automatic email reply generator, I can have a trigger such that when an email is received, a custom AI reply is generated by MyAskAI, and finally a draft is created in my email client. Or, for an automation where I can create a social media posts on multiple platforms based on a RSS feed (news feed), I can implement this directly in Zapier with their native GPT action (see screenshot) As for more complex LLM flows that may require multiple layers of LLMs, data sources, and APIs working together to generate a single response i.e. a long form 100 page report, I would recommend tools such as Stack AI or Flowise (open-source alternative) to build these solutions out. Essentially, you get most of the functions and features of Python packages such as Langchain and LlamaIndex in a GUI. See screenshot for an example of a flow How the hell are you supposed to find clients? With all that being said, none of this matters if you can't find anyone to sell to. You will have to do cold sales, one way or the other, especially if you are brand new to the game. And what better way to sell your AI services than with AI itself? If we want to integrate AI into the cold outreach process, first we must identify what it's good at doing, and that's obviously writing a bunch of text, in a short amount of time. Similar to the solutions that an AAA can build for its clients, we can take advantage of the same principles in our own sales processes. How to do outreach Once you've identified your niche and their pain points/opportunities for automation, you want to craft a compelling message in which you can send via cold email and cold calls to get prospects booked on demos/consultations. I won't get into too much detail in terms of exactly how to write emails or calling scripts, as there are millions of resources to help with this, but I will tell you a few key points you want to keep in mind when doing outreach for your AAA. First, you want to keep in mind that many businesses are still hesitant about AI and may not understand what it really is or how it can benefit their operations. However, we can take advantage of how mass media has been reporting on AI this past year- at the very least people are AWARE that sooner or later they may have to implement AI into their businesses to stay competitive. We want to frame our message in a way that introduces generative AI as a technology that can have a direct, tangible, and positive impact on their business. Although it may be hard to quantify, I like to include estimates of man-hours saved or costs saved at least in my final proposals to prospects. Times are TOUGH right now, and money is expensive, so you need to have a compelling reason for businesses to get on board. Once you've gotten your messaging down, you will want to create a list of prospects to contact. Tools you can use to find prospects include Apollo.io, reply.io, zoominfo (expensive af), and Linkedin Sales Navigator. What specific job titles, etc. to target will depend on your niche but for smaller companies this will tend to be the owner. For white collar niches, i.e. law, the professional that will be directly benefiting from the tool (i.e. partners) may be better to contact. And for larger organizations you may want to target business improvement and digital transformation leads/directors- these are the people directly in charge of projects like what you may be proposing. Okay- so you have your message, and your list, and now all it comes down to is getting the good word out. I won't be going into the details of how to send these out, a quick Google search will give you hundreds of resources for cold outreach methods. However, personalization is key and beyond simple dynamic variables you want to make sure you can either personalize your email campaigns directly with AI (SmartWriter.ai is an example of a tool that can do this), or at the very least have the ability to import email messages programmatically. Alternatively, ask ChatGPT to make you a Python Script that can take in a list of emails, scrape info based on their linkedin URL or website, and all pass this onto a GPT prompt that specifies your messaging to generate an email. From there, send away. How tf do I close? Once you've got some prospects booked in on your meetings, you will need to close deals with them to turn them into clients. Call #1: Consultation Tying back to when I mentioned you want to take a consultant-first appraoch, you will want to listen closely to their goals and needs and understand their pain points. This would be the first call, and typically I would provide a high level overview of different solutions we could build to tacke these. It really helps to have a presentation available, so you can graphically demonstrate key points and key technologies. I like to use Plus AI for this, it's basically a Google Slides add-on that can generate slide decks for you. I copy and paste my default company messaging, add some key points for the presentation, and it comes out with pretty decent slides. Call #2: Demo The second call would involve a demo of one of these solutions, and typically I'll quickly prototype it with boilerplate code I already have, otherwise I'll cook something up in a no-code tool. If you have a niche where one type of solution is commonly demanded, it helps to have a general demo set up to be able to handle a larger volume of calls, so you aren't burning yourself out. I'll also elaborate on how the final product would look like in comparison to the demo. Call #3 and Beyond: Once the initial consultation and demo is complete, you will want to alleviate any remaining concerns from your prospects and work with them to reach a final work proposal. It's crucial you lay out exactly what you will be building (in writing) and ensure the prospect understands this. Furthermore, be clear and transparent with timelines and communication methods for the project. In terms of pricing, you want to take this from a value-based approach. The same solution may be worth a lot more to client A than client B. Furthermore, you can create "add-ons" such as monthly maintenance/upgrade packages, training sessions for employeees, and so forth, separate from the initial setup fee you would charge. How you can incorporate AI into marketing your businesses Beyond cold sales, I highly recommend creating a funnel to capture warm leads. For instance, I do this currently with my AI tools directory, which links directly to my AI agency and has consistent branding throughout. Warm leads are much more likely to close (and honestly, much nicer to deal with). However, even without an AI-related website, at the very least you will want to create a presence on social media and the web in general. As with any agency, you will want basic a professional presence. A professional virtual address helps, in addition to a Google Business Profile (GBP) and TrustPilot. a GBP (especially for local SEO) and Trustpilot page also helps improve the looks of your search results immensely. For GBP, I recommend using ProfilePro, which is a chrome extension you can use to automate SEO work for your GBP. Aside from SEO optimzied business descriptions based on your business, it can handle Q/A answers, responses, updates, and service descriptions based on local keywords. Privacy and Legal Concerns of the AAA Model Aside from typical concerns for agencies relating to service contracts, there are a few issues (especially when using no-code tools) that will need to be addressed to run a successful AAA. Most of these surround privacy concerns when working with proprietary data. In your terms with your client, you will want to clearly define hosting providers and any third party tools you will be using to build their solution, and a DPA with these third parties listed as subprocessors if necessary. In addition, you will want to implement best practices like redacting private information from data being used for building solutions. In terms of addressing concerns directly from clients, it helps if you host your solutions on their own servers (not possible with AI tools), and address the fact only ChatGPT queries in the web app, not OpenAI API calls, will be used to train OpenAI's models (as reported by mainstream media). The key here is to be open and transparent with your clients about ALL the tools you are using, where there data will be going, and make sure to get this all in writing. have fun, and keep an open mind Before I finish this post, I just want to reiterate the fact that this is NOT an easy way to make money. Running an AI agency will require hours and hours of dedication and work, and constantly rearranging your schedule to meet prospect and client needs. However, if you are looking for a new business to run, and have a knack for understanding business operations and are genuinely interested in the pracitcal applications of generative AI, then I say go for it. The time is ticking before AAA becomes the new dropshipping or SMMA, and I've a firm believer that those who set foot first and establish themselves in this field will come out top. And remember, while 100 thousand people may read this post, only 2 may actually take initiative and start.

As a soloproneur, here is how I'm scaling with AI and GPT-based tools
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

As a soloproneur, here is how I'm scaling with AI and GPT-based tools

Being a solopreneur has its fair share of challenges. Currently I've got businesses in ecommerce, agency work, and affiliate marketing, and one undeniable truth remains: to truly scale by yourself, you need more than just sheer will. That's where I feel technology, especially AI, steps in. As such, I wanted some AI tools that have genuinely made a difference in my own work as a solo business operator. No fluff, just tried-and-true tools and platforms that have worked for me. The ability for me to scale alone with AI tools that take advantage of GPT in one way, or another has been significant and really changed my game over the past year. They bring in an element of adaptability and intelligence and work right alongside “traditional automation”. Whether you're new to this or looking to optimize your current setup, I hope this post helps. FYI I used multiple prompts with GPT-4 to draft this using my personal notes. Plus AI (add-on for google slides/docs) I handle a lot of sales calls and demos for my AI automation agency. As I’m providing a custom service rather than a product, every client has different pain points and as such I need to make a new slide deck each time. And making slides used to be a huge PITA and pretty much the bane of my existence until slide deck generators using GPT came out. My favorite so far has been PlusAI, which works as a plugin for Google Slides. You pretty much give it a rough idea, or some key points and it creates some slides right within Google Slides. For me, I’ve been pasting the website copy or any information on my client, then telling PlusAI the service I want to propose. After the slides are made, you have a lot of leeway to edit the slides again with AI, compared to other slide generators out there. With 'Remix', I can switch up layouts if something feels off, and 'Rewrite' is there to gently nudge the AI in a different direction if I ever need it to. It's definitely given me a bit of breathing space in a schedule that often feels suffocating. echo.win (web-based app) As a solopreneur, I'm constantly juggling roles. Managing incoming calls can be particularly challenging. Echo.win, a modern call management platform, has become a game-changer for my business. It's like having a 24/7 personal assistant. Its advanced AI understands and responds to queries in a remarkably human way, freeing up my time. A standout feature is the Scenario Builder, allowing me to create personalized conversation flows. Live transcripts and in-depth analytics help me make data-driven decisions. The platform is scalable, handling multiple simultaneous calls and improving customer satisfaction. Automatic contact updates ensure I never miss an important call. Echo.win's pricing is reasonable, offering a personalized business number, AI agents, unlimited scenarios, live transcripts, and 100 answered call minutes per month. Extra minutes are available at a nominal cost. Echo.win has revolutionized my call management. It's a comprehensive, no-code platform that ensures my customers are always heard and never missed MindStudio by YouAi (web app/GUI) I work with numerous clients in my AI agency, and a recurring task is creating chatbots and demo apps tailored to their specific needs and connected to their knowledge base/data sources. Typically, I would make production builds from scratch with libraries such as LangChain/LlamaIndex, however it’s quite cumbersome to do this for free demos. As each client has unique requirements, it means I'm often creating something from scratch. For this, I’ve been using MindStudio (by YouAi) to quickly come up with the first iteration of my app. It supports multiple AI models (GPT, Claude, Llama), let’s you upload custom data sources via multiple formats (PDF, CSV, Excel, TXT, Docx, and HTML), allows for custom flows and rules, and lets you to quickly publish your apps. If you are in their developer program, YouAi has built-in payment infrastructure to charge your users for using your app. Unlike many of the other AI builders I’ve tried, MindStudio basically lets me dictate every step of the AI interaction at a high level, while at the same time simplifying the behind-the-scenes work. Just like how you'd sketch an outline or jot down main points, you start with a scaffold or decide to "remix" an existing AI, and it will open up the IDE. I often find myself importing client data or specific project details, and then laying out the kind of app or chatbot I'm looking to prototype. And once you've got your prototype you can customize the app as much as you want. LLamaIndex (Python framework) As mentioned before, in my AI agency, I frequently create chatbots and apps for clients, tailored to their specific needs and connected to their data sources. LlamaIndex, a data framework for LLM applications, has been a game-changer in this process. It allows me to ingest, structure, and access private or domain-specific data. The major difference over LangChain is I feel like LlamaIndex does high level abstraction much better.. Where LangChain unnecessarily abstracts the simplest logic, LlamaIndex actually has clear benefits when it comes to integrating your data with LLMs- it comes with data connectors that ingest data from various sources and formats, data indexes that structure data for easy consumption by LLMs, and engines that provide natural language access to data. It also includes data agents, LLM-powered knowledge workers augmented by tools, and application integrations that tie LlamaIndex back into the rest of the ecosystem. LlamaIndex is user-friendly, allowing beginners to use it with just five lines of code, while advanced users can customize and extend any module to fit their needs. To be completely honest, to me it’s more than a tool- at its heart it’s a framework that ensures seamless integration of LLMs with data sources while allowing for complete flexibility compared to no-code tools. GoCharlie (web app) GoCharlie, the first AI Agent product for content creation, has been a game-changer for my business. Powered by a proprietary LLM called Charlie, it's capable of handling multi-input/multi-output tasks. GoCharlie's capabilities are vast, including content repurposing, image generation in 4K and 8K for various aspect ratios, SEO-optimized blog creation, fact-checking, web research, and stock photo and GIF pull-ins. It also offers audio transcriptions for uploaded audio/video files and YouTube URLs, web scraping capabilities, and translation. One standout feature is its multiple input capability, where I can attach a file (like a brand brief from a client) and instruct it to create a social media campaign using brand guidelines. It considers the file, prompt, and website, and produces multiple outputs for each channel, each of which can be edited separately. Its multi-output feature allows me to write a prompt and receive a response, which can then be edited further using AI. Overall, very satisfied with GoCharlie and in my opinion it really presents itself as an effective alternative to GPT based tools. ProfilePro (chrome extension) As someone overseeing multiple Google Business Profiles (GBPs) for my various businesses, I’ve been using ProfilePro by Merchynt. This tool stood out with its ability to auto-generate SEO-optimized content like review responses and business updates based on minimal business input. It works as a Chrome extension, and offers suggestions for responses automatically on your GBP, with multiple options for the tone it will write in. As a plus, it can generate AI images for Google posts, and offer suggestions for services and service/product descriptions. While it streamlines many GBP tasks, it still allows room for personal adjustments and refinements, offering a balance between automation and individual touch. And if you are like me and don't have dedicated SEO experience, it can handle ongoing optimization tasks to help boost visibility and drive more customers to profiles through Google Maps and Search

How to increase the sales of my book
reddit
LLM Vibe Score0
Human Vibe Score1
danonino80This week

How to increase the sales of my book

In just 3 months, it generated over $100 in revenue. I wanted to share my journey for two reasons: to potentially assist others in self-publishing their own books and to receive feedback to enhance my marketing strategy. I envision that there are others facing similar challenges. Let's dive into the financials, time spent, Key takeaways and the Challenges to address behind this product. Finances First, let's take a look at the financial overview. 💳 Expenses 🔹 E-book creation: · Book cover: $ 0. I used Adobe Express with 30 days of free trial. · ChatGPT: 20 $ a month. I leveraged AI to generate the chapters of the book, ensuring that no critical topics were overlooked during the content creation process and to refine the English, as it's not my native language. I also used to help me with copywriting of the web. If anyone is interested, I can share my Python code for outlining the chapters calling the API, but you can also directly ask chatgpt. · Kindle KDP (Kindle Direct Publishing): order author copies: 10 $. 🔹 Web creation: Domain: I got a com) / .org /.net domain for just 1 $ the first year. Carrd.co subscription: 19 $ (1 year) 🔹 Marketing: Promoted post on reddit: $30 Paid ads with google ads: $30 💰 Revenue 🔸 Sales: $102 💸 Net Profit: \~- $ 18 I initially thought the sales for this e-book would be quite modest, maybe only 3 or 4 books. However, the fact that I've sold more than that so far is a pleasant surprise. Even though the overall numbers may still be considered "peanuts" in the grand scheme of book sales, it suggests there could be more demand for content on digital asset custody than I had originally anticipated. This is a good learning experience, and I'll look to refine my marketing approach to see if I can reach a wider audience interested in this topic 🔹 Time Spent Next, let's review the time invested. 📖 Writing the e-book: 40 hours 🌍 Website + Stripe integration: 10 hours 📣 Creating promotional content: 10 hours ⏱️ Additional marketing efforts: 5 hours Total time spent: 65 hours As you can see, I dedicated more time to writing the e-book itself than to marketing and distribution. I spent relevant time to marketing because I though that a successful product launch requires a robust marketing effort. Many e-book authors overlook this crucial aspect! I utilized three sales channels: · Amazon: I found that there were no books specifically about digital asset custody, resulting in strong positioning in Amazon searches. Additionally, my book immediately secured the top position in Google searches for "digital asset custody book." However, despite achieving 50% of sales in the UK, I have not received any reviews globally. Sales distribution for this channel: 20% physical book, 80% ebook. · Twitter: Daniel\_ZZ80. With only 46 followers, the performance on this platform has not been optimal. I am beginning to write posts related to digital assets to increase visibility. · Gumroad: Lockeyyy.gumroad.com. I offered a discounted version of the ebook, but have not yet made any sales through this channel. Key takeaways: · The process of creating this e-book was extremely fulfilling, and while it has garnered overwhelmingly positive feedback from friends and colleagues (not considered as sales), it has yet to receive any Amazon reviews ☹. · Kindle KDP proved to be ideal for a rapid go-to-market strategy. · AI is an excellent tool for generating ideas and providing access to global audiences with perfect grammar. Otherwise, I would need to hire a translator, which can be very expensive. · Despite offering a full 30-day money-back guarantee, leading me to believe that the quality of the content is indeed good. · I have gained valuable insights for future technical books. · Although the current financial balance may be negative, I anticipate reaching the break-even point within one month, and this has now become a passive income stream. However, I recognize the need to regularly update the content due to the rapidly changing nature of this field. Challenges to address: · Is the timing for launching this book appropriate? In other words, is the world of digital asset custody a trendy and interesting topic for the audience? · What is causing the lack of sales through Gumroad? · Should I seek assistance as my marketing efforts have not yielded results? · Why are there no reviews on Amazon? · Why are sales primarily concentrated in the EU with only one sale in the US, which is my main target market? Feedback is appreciated. If you're interested in learning more about my approach, feel free to send me a direct message. A bit about my background: After dedicating my entire career to the banking industry, I explored various side projects. As an IT professional, I have now transitioned into the digital asset realm. After three years of intensive study, I recently published my first book on digital asset custody. I hope you found this post informative. Cheers! P.S.: I'm currently in the process of launching two more books using this system. 😊

AI SaaS: A website to fine-tune LLM model according to your requirements
reddit
LLM Vibe Score0
Human Vibe Score1
Dangerous_Ferret3362This week

AI SaaS: A website to fine-tune LLM model according to your requirements

Hey fellow entrepreneurs and AI enthusiasts! I'm exploring a business idea and would love your thoughts and feedback. The concept is a SaaS platform that allows users to easily fine-tune large language models (LLMs) on their own datasets without needing deep technical expertise. Here's the gist: The Problem: Many businesses and researchers want to leverage LLMs for specific use cases, but fine-tuning these models requires significant technical knowledge and resources. The Solution: A user-friendly web platform where users can: Choose from popular LLM architectures Upload their own dataset or input text Configure fine-tuning parameters through an intuitive interface Automatically fine-tune the model on our GPU infrastructure Download the fine-tuned model or use it via API Key Features: No coding required Scalable cloud infrastructure Support for various fine-tuning techniques (prompt tuning, adapter tuning, full fine-tuning) Job monitoring and results visualization API access for integrated use in applications Target Market: Researchers without extensive ML engineering resources Startups building AI-powered products Enterprises looking to customize LLMs for internal use Monetization: Tiered subscription model based on usage (compute time, model size, etc.) + potential enterprise contracts for high-volume users. I'd really appreciate your thoughts on: Is this solving a real pain point? Would you use a service like this? Why or why not? What features would make this a must-have for you? Any foreseeable obstacles or considerations I'm missing? Suggestions for go-to-market strategy? Thank you!

Please rate my business idea aimed at emerging markets
reddit
LLM Vibe Score0
Human Vibe Score1
Whole_Ad_9002This week

Please rate my business idea aimed at emerging markets

Imagine a single platform that transforms your business operations effortlessly. Think of it like a Swiss Army knife for your business. First Steps (Every Business Needs These): Our website builder is your first step online. Just pick what you want, drag it where you need it, and you've got a professional website or online store. No tech skills needed. Perfect for getting your business visible to customers. Our security tool comes next because every business needs protection. It's like a security guard for your files and data, keeping hackers out and making sure you never lose important work. Think of it as insurance for your digital business. Cybersecurity and backups combined. Growing Business Needs: As your team grows, our digital office keeps everyone connected. Email, chat, share files, and track projects in one place. It's like having everyone in the same room, even when working remotely. This becomes essential once you have more than a few people. When paperwork starts piling up, our AI helper steps in. It's like having a smart assistant who learns your business and helps everyone get more done - handling routine tasks, summarizing meetings, and suggesting better ways to work. Scaling Up (For Established Businesses): Once you're handling lots of data, our eco-friendly cloud storage becomes crucial. Your business gets fast, reliable storage while helping the environment. We use servers powered by renewable energy, so you can grow sustainably. Our data transfer tool becomes important when you're working with multiple systems and need to move files around. It's like having a digital moving service that safely carries your files between different cloud services. With 20+ services to connect to cloud to cloud or on prem to cloud, its easy and fast. Advanced Needs: When your business processes get complex, our workflow automation tool helps you set up automatic systems for repetitive tasks. It's like training a robot to handle your routine work while your team focuses on growth. Simple no-code automation when you need it. Everything works on its own, but they're designed to work even better together. And since it's all in one place, you only need one password and one monthly bill. Most businesses save about half their tech costs by switching to our platform. Cost Benefits: Single subscription replaces multiple vendor contracts Reduces IT overhead by 40-60% on average No need for expensive technical specialists Predictable monthly costs Scales pricing with your usage

Interview with founder of ReadyPlayerMe (raised $70M+ from a16z)
reddit
LLM Vibe Score0
Human Vibe Score1
Due_Cryptographer461This week

Interview with founder of ReadyPlayerMe (raised $70M+ from a16z)

Thanks to everyone who replied to my previous post with the questions you had for Rainer, I added some of them into this interview. I’m Nikita of Databas3 , and that’s my first interview in a series where I’m learning more about the journey of the best tech and web3 founders. Would appreciate your feedback and suggestions for the next guest! Nikita: Let’s begin with a brief introduction. Can you share a bit about yourself and how the business started? Rainer: I’m Rainer, the CTO of ReadyPlayerMe. Our journey began in 2013 with four co-founders. Over the years, our focus has shifted mainly around our product’s evolution, but our core idea always revolved around virtual actors or virtual people. Our initial venture was into hardware. We created the first full-body scanner in the Nordics, a significant step in photogrammetry. This led us to develop the Luna Scanner, a three-meter tall structure designed to capture facial features and likenesses. When Facebook acquired Oculus in 2014, we foresaw the potential of VR and virtual worlds, especially in social experiences. Nikita: Interesting. How did you move on from there? Rainer: Recognizing the limitations of hardware, we transitioned into software. Our early scanner designs had limitations in scalability. For example, our three-meter tall scanner wasn’t a feasible solution for scanning millions of people. So, we leveraged the datasets from our initial projects and designed a mobile version, making facial scanning as easy as using your phone. Around 2015, this was a new territory, as facial scanning wasn’t a mainstream application. Nikita: What were the early applications of these scanned models? Rainer: In the beginning, we focused on 3D printed figurines from full-body scans. However, as we shifted to facial scanning, we licensed our technology to gaming companies, collaborating with giants like Wargaming and Tencent. We even ventured into virtual fittings with H&M. Each collaboration was custom-tailored, blending our technology with their systems. This model made us cash flow positive. Nikita: So this was the beginning of your foray into the gaming industry? Rainer: Precisely. The demand from gaming companies was substantial. As we built custom solutions for these enterprises, we saw a bigger potential. While our cash flow was positive, we realized the challenge of scaling through exclusive enterprise deals. We envisioned our avatar creation tech reaching indie games and beyond. Nikita: And that led to the birth of ReadyPlayerMe? Rainer: Exactly. Once we understood our market direction, we quickly developed the first iteration of ReadyPlayerMe as a web-based experience, emphasizing easy integration for game developers. The initial version was a character builder, allowing users to personalize their avatars, which many adopted for their social media profiles. Our goal was to create avatars that users could connect with and use across various platforms. Instead of licensing our technology, we offered it for free to everyone. As ReadyPlayerMe gained traction, especially in VR applications, we secured funding to further our mission. Nikita: Your growth seems swift and organic. Were there any challenges? Rainer: Our focus on easy integration significantly fueled our adoption. Pairing that with personalized avatars resonated well with our audience. But like any venture, we’ve faced our share of challenges and have always aimed to evolve and better our offerings. The rapid growth in Web3 projects and virtual worlds made personalization and customization more important. With the NFT boom, you could add utility by allowing access to selected collections. This played into web-based games and metaverse applications. The shift towards Web3 and personalization provided a significant tailwind for us. Many used our characters as profile pictures on social media. Nikita: I’ve heard from other founders that a16z really values viral marketing. Was this one reason they wanted to invest in your project? How was the process with them? Rainer: When a16z reached out, it felt like a natural fit. We wanted investors who understood the gaming space. Our main market is Web3, but we’re exploring the top games market. Their expertise in gaming was invaluable. They’ve been very supportive throughout. We were fortunate to be on their radar. Nikita: So your early growth and organic traction played a role in attracting investors? Rainer: Definitely. Early product growth and the potential future trajectory were essential in our discussions. Nikita: As the CTO, you must have faced challenges. Can you speak about the tech side and its evolution? Rainer: The early version of our platform was built by in-house engineers. As we grew, we had to adapt to increasing complexities and ensure we had the right team to execute our vision. My role often shifted between product management and tech, depending on the need. Nikita: It sounds like the startup environment remains strong within your company. Rainer: Absolutely. We’re all committed, hands-on, and working towards building the best product. Nikita: You mentioned the team earlier. How many people are in your team now? Rainer: We have 70 people, with about half in product and engineering. Nikita: And did you hire the tech team? Rainer: We brought on a head of engineering at the beginning of this year. He’s been instrumental in scaling the engineering organization, from increasing the headcount to refining engineering processes. We’ve recently reorganized into domain-specific teams. As the team grows, regular reorganization ensures we focus on delivering specific customer value. Every stage requires attention to the team’s composition to ensure efficient delivery. Nikita: Any advice for founders just starting with their first startup? Rainer: Focus on customer value, no matter how niche it might seem initially. Begin with a specific problem and solution, then expand from there. You don’t need a massive project right away. Begin small, prove the concept, and scale from there. Nikita: You’ve mentioned your love for books and podcasts. Any recommendations? Rainer: For startups, “High Growth Handbook” and “Lean Startup” are must-reads. “Working Backwards” offers insights into Amazon’s customer-centric approach. For podcasts, I listen to “Rework,” “Lenny’s Podcast,” and “Huberman Lab.” Nikita: All of us have some side project ideas from time to time. How do you handle these when managing a big project? Rainer: Over the years, I’ve built various side projects. Some are small applications to solve immediate problems, like a menu bar app for AirPods which made it to No. 1 on Product Hunt, and was nominated for Golden Kitty Award. I sometimes delve into 3D and AI, merging them for technical demos. I keep a list of ideas and pick from them as the urge arises. Nikita: Any final thoughts or advice? Rainer: As you scale, do so with clarity. Avoid scaling just for external appeal. Always hire when there’s genuine need, not just for the sake of expansion. It helps in staying lean and focused.

How to increase the sales of my book
reddit
LLM Vibe Score0
Human Vibe Score1
danonino80This week

How to increase the sales of my book

In just 3 months, it generated over $100 in revenue. I wanted to share my journey for two reasons: to potentially assist others in self-publishing their own books and to receive feedback to enhance my marketing strategy. I envision that there are others facing similar challenges. Let's dive into the financials, time spent, Key takeaways and the Challenges to address behind this product. Finances First, let's take a look at the financial overview. 💳 Expenses 🔹 E-book creation: · Book cover: $ 0. I used Adobe Express with 30 days of free trial. · ChatGPT: 20 $ a month. I leveraged AI to generate the chapters of the book, ensuring that no critical topics were overlooked during the content creation process and to refine the English, as it's not my native language. I also used to help me with copywriting of the web. If anyone is interested, I can share my Python code for outlining the chapters calling the API, but you can also directly ask chatgpt. · Kindle KDP (Kindle Direct Publishing): order author copies: 10 $. 🔹 Web creation: Domain: I got a com) / .org /.net domain for just 1 $ the first year. Carrd.co subscription: 19 $ (1 year) 🔹 Marketing: Promoted post on reddit: $30 Paid ads with google ads: $30 💰 Revenue 🔸 Sales: $102 💸 Net Profit: \~- $ 18 I initially thought the sales for this e-book would be quite modest, maybe only 3 or 4 books. However, the fact that I've sold more than that so far is a pleasant surprise. Even though the overall numbers may still be considered "peanuts" in the grand scheme of book sales, it suggests there could be more demand for content on digital asset custody than I had originally anticipated. This is a good learning experience, and I'll look to refine my marketing approach to see if I can reach a wider audience interested in this topic 🔹 Time Spent Next, let's review the time invested. 📖 Writing the e-book: 40 hours 🌍 Website + Stripe integration: 10 hours 📣 Creating promotional content: 10 hours ⏱️ Additional marketing efforts: 5 hours Total time spent: 65 hours As you can see, I dedicated more time to writing the e-book itself than to marketing and distribution. I spent relevant time to marketing because I though that a successful product launch requires a robust marketing effort. Many e-book authors overlook this crucial aspect! I utilized three sales channels: · Amazon: I found that there were no books specifically about digital asset custody, resulting in strong positioning in Amazon searches. Additionally, my book immediately secured the top position in Google searches for "digital asset custody book." However, despite achieving 50% of sales in the UK, I have not received any reviews globally. Sales distribution for this channel: 20% physical book, 80% ebook. · Twitter: Daniel\_ZZ80. With only 46 followers, the performance on this platform has not been optimal. I am beginning to write posts related to digital assets to increase visibility. · Gumroad: Lockeyyy.gumroad.com. I offered a discounted version of the ebook, but have not yet made any sales through this channel. Key takeaways: · The process of creating this e-book was extremely fulfilling, and while it has garnered overwhelmingly positive feedback from friends and colleagues (not considered as sales), it has yet to receive any Amazon reviews ☹. · Kindle KDP proved to be ideal for a rapid go-to-market strategy. · AI is an excellent tool for generating ideas and providing access to global audiences with perfect grammar. Otherwise, I would need to hire a translator, which can be very expensive. · Despite offering a full 30-day money-back guarantee, leading me to believe that the quality of the content is indeed good. · I have gained valuable insights for future technical books. · Although the current financial balance may be negative, I anticipate reaching the break-even point within one month, and this has now become a passive income stream. However, I recognize the need to regularly update the content due to the rapidly changing nature of this field. Challenges to address: · Is the timing for launching this book appropriate? In other words, is the world of digital asset custody a trendy and interesting topic for the audience? · What is causing the lack of sales through Gumroad? · Should I seek assistance as my marketing efforts have not yielded results? · Why are there no reviews on Amazon? · Why are sales primarily concentrated in the EU with only one sale in the US, which is my main target market? Feedback is appreciated. If you're interested in learning more about my approach, feel free to send me a direct message. A bit about my background: After dedicating my entire career to the banking industry, I explored various side projects. As an IT professional, I have now transitioned into the digital asset realm. After three years of intensive study, I recently published my first book on digital asset custody. I hope you found this post informative. Cheers! P.S.: I'm currently in the process of launching two more books using this system. 😊

Writing a exercise based TTRPG rulebook for a system where your real world fitness is tied to character progression
reddit
LLM Vibe Score0
Human Vibe Score1
BezboznyThis week

Writing a exercise based TTRPG rulebook for a system where your real world fitness is tied to character progression

My dad was a star athlete when he was young, and my mom was a huge sci-fi/fantasy nerd, so I got both ends of the stick as it were. Love gaming and nerd culture, but also love to exercise and self improvement. Sometimes exercise can feel boring though compared to daydreaming about fantastic fictional worlds, so for a long time I've been kicking around the idea of how to "Gamify" fitness. and recently I've been working on this passion project of a Table Top RPG (Like D&D) where the stats of your character are related to your own fitness, so if you want your character in game to improve, you have to improve in the real world. Below is a rough draft you can look through that details the settings and mechanics of the game I've come up with so far. I'd love to eventually get a full book published and sell it online. maybe even starting a whole brand of "Gamified fitness": REP-SET: GAINSZ In the war torn future of 24th century… There are no rest days… In the futuristic setting of "REP-SET: GAINSZ," the "War of Gains" casts a long shadow over the Sol System as the various factions vie for territory and resources. However, war has evolved. Unmanned drones and long-range strikes have faded into obsolescence. Battles, both planet-side and in the depths of space, are now fought by soldiers piloting REP-SETs: Reactive Exoskeletal Platform - Symbiotic Evolution Trainer Massive, humanoid combat mechs. Powered by mysterious “EV” energy, these mechanical marvels amplify, and are in turn amplified by, the fitness and mental acuity of their pilots. The amplification is exponential, leading pilots into a life of constant training in order for their combat prowess to be bolstered by every incremental gain in their level of fitness. With top pilots having lifting capacity measured in tons, and reaction times measured by their Mach number, REP-SET enhanced infantry now dominate the battlefield. The Factions: The Federated Isometocracy of Terra (FIT): Quote: "The strength of the body is the strength of the spirit. Together, we will lift humanity to its destined greatness. But ask not the federation to lift for you. Ask yourself: Do you even lift for the Federation?" Description: An idealistic but authoritarian faction founded on the principle of maximizing the potential of all individuals. FIT citizens believe in relentless striving for physical and mental perfection, leading to collective excellence. Their goal is the unification of humankind under a rule guided by this doctrine, which sometimes comes at the cost of individual liberties. Mech Concept: REP-SET mechs. Versatile humanoid designs focusing on strength, endurance, and adaptability. By connecting to the AI spirit within their REP-SETs core, each pilot enhances the performance of their machine through personal willpower and peak physical training. Some high-rank REP-SETS include features customized to the pilot's strengths, visually signifying their dedication and discipline. The Dominion of Organo-Mechanical Supremacy (DOMS): Quote: "Without pain, there is no gain. Become the machine. Embrace the burn.” Description: A fanatical collective ideologically obsessed with "Ascendency through suffering" by merging their bodies with technology that not only transcends biological limitations, but also acts to constantly induce pain in it's users. Driven by a sense of ideological superiority and a thirst for domination, DOMS seek to bring the painful blessings of their deity "The lord of the Burn" to the rest of the solar system. Their conquest could turn them into a significant threat to humanity. Mech Concept: Hybrid mechs, where the distinction between the pilot and the machine is blurred. The cockpit functions as a life-support system for the pilot, heavily modified with augmentations. Mechs themselves are often modular, allowing for adaptation and assimilation of enemy technology. Some DOMS mechs might display disturbing elements of twisted flesh alongside cold, mechanical parts. The Tren: Quote: "Grow... bigger... feast... protein..." Description: A ravenous conglomeration of biochemically engineered muscular monstrosities, united only by a shared insatiable hunger for "More". Existing mostly in deep space, they seek organic matter to consume and assimilate. They progress in power not due to any form of training or technology, but from a constant regimen of ravenous consumption and chemically induced muscle growth, all exponentially enhanced by EV energies. While some have been known to possess a certain level of intellect and civility, their relentless hunger makes them incredibly mentally volatile. When not consuming others, the strong consume the weak within their own faction. Mech Concept: Bio-Organic horrors. While they do have massive war machines, some are living vessels built around immense creatures. These machines resemble grotesque fleshy designs that prioritize rapid mutation and growth over sleek aesthetics. Often unsettling to behold. Synthetic Intelligence Theocracy (SIT): Quote: "Failure is an unacceptable data point.” Description: A society ruled by a vast and interconnected artificial intelligence network. The SIT governs with seemingly emotionless rationality, striving for efficiency and maximum productivity. This leads to a cold, but arguably prosperous society, unless you challenge the logic of the collective AI. Their goals? Difficult to predict, as it hinges on how the AI calculates what's "optimal" for the continuation or "evolution" of existence. Mech Concept: Sleek, almost featureless robotic creations with a focus on efficient movement and energy management. Often drone-like or modular, piloted through direct mind-machine linking rather than traditional cockpits. Their aesthetic suggests cold and impersonal perfection. The Way Isolate(TWI): Quote: "The body unblemished, the mind unwavering. That is the path to true strength. That and a healthy diet of Aster-Pea proteins." Description: Known by some as "The asteroid farmers", The Way Isolate is a proud and enigmatic faction that stands apart from the other powers in the Sol System. A fiercely independent tribe bound by oaths of honor, loyalty, and hard work. Wandering the asteroid belt in their vast arc ships, their unparalleled mastery in asteroidal-agricultural engineering, ensuring they have no need to colonize planets for nutritional needs, has allowed them to abstain from the pursuit of territorial expansion in “The War of Gains”, instead focusing on inward perfection, both spiritual and physical. They eschew all technological bodily enhancements deemed unnatural, believing that true power can only be cultivated through the relentless pursuit of personal strength achieved through sheer will and bodily perfection. The Way Isolate views biohacking, genetic manipulation, and even advanced cybernetics as corruptions of the human spirit, diluting the sacredness of individual willpower. Mech Concept: Way Isolate mechs are built with maneuverability and precision in mind rather than flashy augmentations. Their REP-SETs are streamlined, favoring lean designs that mirror the athleticism of their pilots. Excelling in low to zero G environments, their mechs lack bulky armor, relying on evasion and maneuverability rather than brute force endurance. Weaponry leans towards traditional kinetic based armaments, perhaps employing archaic but reliable weapon styles such as blades or axes as symbols of their purity of purpose. These mechs reflect the individual prowess of their pilots, where victory is determined by focus, technique, and the raw power of honed physical ability. Base Player Character Example: You are a young, idealistic FIT soldier, barely out of training and working as a junior REP-SET mechanic on the Europa Ring World. The Miazaki district, a landscape of towering mountains and gleaming cities, houses a sprawling mountainside factory – a veritable hive of Gen 5 REP-SET construction. Here, the lines between military and civilian blur within a self-sufficient society dependent on this relentless industry. Beneath the surface, you harbor a secret. In a forgotten workshop, the ghost of a REP-SET takes shape – a unique machine built around an abandoned, enigmatic AI core. Ever since you salvaged it as a child from the wreckage of your hometown, scarred by a brutal Tren attack, you've dedicated yourself to its restoration. A lingering injury from that fateful battle mocks your progress, a constant reminder of the fitness exams you cannot pass. Yet, you train relentlessly, dreaming of the day you'll stand as a true REP-SET pilot. A hidden truth lies at the heart of the REP-SETS: as a pilot's abilities grow, their mech develops unique, almost mystical powers – a manifestation of the bond between the human spirit and the REP-SET's AI. The ache in your old wound serves as a grim prophecy. This cold war cannot last. The drums of battle grow louder with each passing day. GAME MECHANICS: The TTRPG setting of “REP-SET: GAINSZ” is marked by a unique set of rules, by which the players real world capabilities and fitness will reflect and affect the capabilities, progression, and success of their REP-SET pilot character in-game. ABILITY SCORES: Pilots' capabilities will be defined by 6 “Ability scores”: Grace, Agility, Iron, Nourishment, Strength, and Zen. Each of the 6 ability scores will duel represent both a specific area of exercise/athleticism and a specific brand of healthy habits. The definitions of these ability scores are as follows: Grace (GRC): "You are an artist, and your body is your canvas; the way you move is your paint and brush." This ability score, the domain of dancers and martial artists, represents a person's ability to move with organic, flowing control and to bring beauty to the world. Skill challenges may be called upon when the player character needs to act with poise and control, whether socially or physically. Real-world skill checks may involve martial arts drills, dancing to music, or balance exercises. Bonuses may be granted if the player has recently done something artistically creative or kind, and penalties may apply if they have recently lost their temper. This ability score affects how much NPCs like your character in game. Agility (AGI): "Your true potential is locked away, and speed is the key to unlocking it." The domain of sprinters, this ability score represents not only a person's absolute speed and reaction time but also their capacity to finish work early and avoid procrastination. Skill challenges may be called upon when the player character needs to make a split-second choice, move fast, or deftly dodge something dangerous. Real-world skill checks may involve acts of speed such as sprinting or punching/kicking at a steadily increasing tempo. Bonuses may apply if the player has finished work early, and penalties may apply if they are procrastinating. This ability score affects moving speed and turn order in game. Iron (IRN): "Not money, nor genetics, nor the world's greatest trainers... it is your resolve, your will to better yourself, that will make you great." Required by all athletes regardless of focus, this ability score represents a player's willpower and their capacity to push through pain, distraction, or anything else to achieve their goals. Skill challenges may be called upon when the player character needs to push through fear, doubt, or mental manipulation. Real-world skill checks may involve feats of athletic perseverance, such as planking or dead hangs from a pull-up bar. Bonuses may apply when the player maintains or creates scheduled daily routines of exercise, self-improvement, and work completion, and penalties may apply when they falter in those routines. This ability score affects the max "Dynamic exercise bonus” that can be applied to skill checks in game (a base max of +3 when Iron = 10, with an additional +1 for every 2 points of iron. So if every 20 pushups gives you +1 on a “Strength” skill check, then doing 80 pushups will only give you +4 if you have at least 12 iron). Nourishment (NRS): "A properly nourished body will last longer than a famished one." This ability score, focused on by long-distance runners, represents a player's endurance and level of nutrition. Skill challenges may be called upon when making checks that involve the player character's stamina or health. Real-world skill checks may involve endurance exercises like long-distance running. Bonuses may apply if the player has eaten healthily or consumed enough water, and penalties may apply if they have eaten junk food. This ability score affects your HP (Health points), which determines how much damage you can take before you are incapacitated. Strength (STR): "When I get down on my hands, I'm not doing pushups, I'm bench-pressing the planet." The domain of powerlifters and strongmen, this ability score represents raw physical might and the ability to overcome obstacles. Skill challenges may be called upon when the player character needs to lift, push, or break something. Real-world skill checks might involve weightlifting exercises, feats of grip strength, or core stability tests. Bonuses may apply for consuming protein-rich foods or getting a good night's sleep, and penalties may apply after staying up late or indulging in excessive stimulants. This ability score affects your carrying capacity and base attack damage in game. Zen (ZEN): "Clarity of mind reflects clarity of purpose. Still the waters within to act decisively without." This ability score, prized by meditators and yogis, represents mental focus, clarity, and inner peace. Skill challenges may be called upon when the player character needs to resist distractions, see through illusions, or make difficult decisions under pressure. Real-world skill checks may involve meditation, breathing exercises, or mindfulness activities. Bonuses may apply after attending a yoga class, spending time in nature, or creating a calm and organized living space. Penalties may apply after experiencing significant stress, emotional turmoil, or having an unclean or unorganized living space. This ability score affects your amount of ZP in game (Zen Points: your pool of energy you pull from to use mystical abilities) Determining initial player ability scores: Initially, “Ability scores” are decided during character creation by giving the player a list of 6 fitness tests to gauge their level of fitness in each category. Running each test through a specific calculation will output an ability score. A score of 10 represents the average person, a score of 20 represents a peak athlete in their category. The tests are: Grace: Timed balancing on one leg with eyes closed (10 seconds is average, 60 is peak) Agility: Mile run time in minutes and second (10:00 minutes:seconds is average, 3:47 is peak) Iron: Timed dead-hang from a pull-up bar (30 seconds is average, 160 is peak) Nourishment: Miles run in an hour (4 is average, 12 is peak) Strength: Pushups in 2 minute (34 is average, 100 is peak) Zen: Leg stretch in degrees (80 is average, and 180 aka "The splits" is peak) Initial Score Calculation Formula: Ability Score = 10 + (Player Test Score - Average Score) / (Peak Score - Average\_Score) \* 10 Example: if the player does 58 pushups in 2 minutes, their strength would be: 10 plus (58 - 34) divided by (100-34) multiplied by 10 = 10 + (24)/(66)\* 10 = 10 + 3.6363... = 13.6363 rounded to nearest whole number = Strength (STR): 14 SKILLS AND SKILL CHALLENGES: The core mechanic of the game will be in how skill challenges are resolved. All “Skill challenges” will have a numerical challenge rating that must be met or beaten by the sum of a 10 sided dice roll and your score in the pertinent skill. Skill scores are determined by 2 factors: Ability Score Bonus: Every 2 points above 10 gives +1 bonus point. (EX. 12 = +1, 14 = +2, etc.) This also means that if you have less than 10 in an ability score, you will get negative points. Personal Best Bonus: Each skill has its own unique associated exercise that can be measured (Time, speed, distance, amount of reps, etc). A higher record means a higher bonus. EX: Authority skill checks are associated with a timed “Lateral raise hold”. Every 30 seconds of the hold added onto your personal best single attempt offers a +1 bonus. So if you can do a lateral hold for 90 seconds, that’s a +3 to your authority check! So if you have a 16 in Iron, and your Personal Best lateral raise hold is 90 seconds, that would give you an Authority score of +6 (T-Pose for dominance!) Dynamic Exercise Bonus: This is where the unique mechanics of the game kick in. At any time during a skill challenge (even after your roll) you can add an additional modifier to the skill check by completing the exercise during gameplay! Did you roll just below the threshold for success? Crank out another 20 pushups, squats, or curls to push yourself just over the edge into success! There are 18 skills total, each with its own associated ability score and unique exercise: Grace (GRC): \-Kinesthesia (Timed: Blind single leg stand time) \-Precision (Scored: Basket throws) \-Charm (Timed reps: Standing repeated forward dumbell chest press and thrust) \-Stealth (Timed distance: Leopard Crawl) Agility (AGI): \-acrobatics (timed reps: high kicks) \-Computers (Word per minute: Typing test) \-Speed (Time: 100 meter sprint) Iron (IRN): \-Authority (Timed: Lateral raise hold) \-Resist (Timed: Plank) \-Persist (Timed:Pull-up bar dead hang) Nourishment(NRS): \-Recovery (TBD) \-Stim crafting (TBD) \-Survival (TBD) Strength(STR): \-Mechanics (Timed reps: Alternating curls) \-Might (Timed reps: pushups) Zen(ZEN): \-Perceive (TBD) \-Empathy (TBD) \-Harmony (TBD) \-Lore (TBD) Healthy Habits Bonus: Being able to demonstrate that you have conducted healthy habits during gameplay can also add one time bonuses per skill challenge “Drank a glass of water +1 to Nourishment check”, “Cleaned your room, +3 on Zen check”. But watch out, if you’re caught in unhealthy Habits, the GM can throw in penalties, “Ate junk food, -1 to Nourishment check”, etc. Bonuses/penalties from in-game items, equipment, buffs, debuffs, etc., helping players to immerse into the mechanics of the world of REP-SET for the thrill of constantly finding ways to improve their player. Gradient success: Result of skill challenges can be pass or fail, but can also be on a sliding scale of success. Are you racing to the battlefield? Depending on your Speed check, you might arrive early and have a tactical advantage, just in time for an even fight, or maybe far too late and some of your favorite allied NPCs have paid the price… So you’re often encouraged to stack on those dynamic exercise bonuses when you can to get the most fortuitous outcomes available to you. Gameplay sample: GM: Your REP-SET is a phantom, a streak of light against the vast hull of the warship. Enemy fighters buzz angrily, but you weaves and dodges with uncanny precision. The energy wave might be losing effectiveness, but your agility and connection to the machine have never been stronger. Then, it happens. A gap in the defenses. A vulnerable seam in the warship's armor. Your coms agents keen eye spots it instantly. "Lower power junction, starboard side! You have an opening!" This is your chance to strike the decisive blow. But how? It'll take a perfect combination of skill and strategy, drawing upon your various strengths. Here are your options: Option 1: Brute Strength: Channel all remaining power into a single, overwhelming blast from the core. High-risk, high-reward. It could overload the REP-SET if you fail, but it might also cripple the warship. (Strength-focused, Might sub-skill) Option 2: Calculated Strike: With surgical precision, target the power junction with a pinpoint burst of destabilizing energy. Less flashy and ultimately less damaging, but potentially more effective in temporarily disabling the ship. (Agility-focused, Precision sub-skill) Option 3: Harmonic Disruption: Attempt to harmonize with your REP-SET's AI spirit for help in connecting to the digital systems of the Warship. Can you generate an internal energy resonance within the warship, causing it to malfunction from within? (Zen-focused, Harmony sub-skill) Player: I'll take option 1, brute strength! GM: Ok, This will be a "Might" check. The CR is going to be very high on this one. I'm setting it at a 20. What's your Might bonus? Player: Dang, a 20?? That's literally impossible. My Might is 15 and I've got a PB of 65 pushups in 2 minutes, that sets me at a +5. Even if I roll a 10 and do 60 pushups for the DE I'll only get 18 max. GM: Hey I told you it was high risk. You want to choose another option? Player: No, no. This is what my character would do. I'm a real hot-blooded meathead for sure. GM: Ok then, roll a D10 and add your bonus. Player: \Rolls\ a 9! not bad, actually that's a really good roll. So +5, that's a 14. GM: Alright, would you like to add a dynamic exercise bonus? Player: Duh, it's not like I can do 120 pushups I'd need to beat the CR, but I can at least do better than 14. Alright, here goes. \the player gets down to do pushups and the 2 minute time begins. After some time...\ Player: 65....... 66! GM: Times up. Player: Ow... my arms... GM: so with 66, that's an extra +3, and its a new PB, so that's a +1. That sets your roll to 18. Player: Ow... Frack... still not 20... for a second there i really believed I could do 120 pushups... well I did my best... Ow... 20 CR is just too impossible you jerk... GM: Hmm... Tell me, what did you eat for lunch today? Player: Me? I made some vegetable and pork soup, and a protein shake. I recorded it all in my diet app. GM: And how did you sleep last night? Player: Like a baby, went to sleep early, woke up at 6. GM: in that case, you can add a +1 "Protein bonus" and +1 "Healthy rest" bonus to any strength related check for the day if you'd like, including this one. Player: Really?? Heck yes! add it to the roll! GM: With those extra bonuses, your roll reaches 20. How do you want to do this? Player: I roar "For Terra!" and pour every last ounce of my strength into the REP-SET. GM: "For Terra!" you roar, your cry echoing through coms systems of the REP-SET. The core flares blindingly bright. The surge of power dwarfs anything the REP-SET has unleashed before. With a titanic shriek that cracks the very fabric of space, the REP-SET slams into the vulnerable power junction. Raw energy explodes outwards, tendrils of light arcing across the warship's massive hull. The impact is staggering. The leviathan-like warship buckles, its sleek form rippling with shockwaves. Sparks shower like rain, secondary explosions erupt as critical systems overload. Then…silence. The warship goes dark. Power flickers within the REP-SET itself, then steadies. Alarms fade, replaced by the eerie quiet of damaged but functional systems. "We…did it?" The coms agents voice is incredulous, tinged with relief. She's awaiting your reply. Player: "I guess so." I say, and I smile and laugh. And then I slump back... and fall unconscious. \to the other players\ I'm not doing any more skill checks for a while guys, come pick me up please. \teammates cheer\ &#x200B;

AI SaaS: A website to fine-tune LLM model according to your requirements
reddit
LLM Vibe Score0
Human Vibe Score1
Dangerous_Ferret3362This week

AI SaaS: A website to fine-tune LLM model according to your requirements

Hey fellow entrepreneurs and AI enthusiasts! I'm exploring a business idea and would love your thoughts and feedback. The concept is a SaaS platform that allows users to easily fine-tune large language models (LLMs) on their own datasets without needing deep technical expertise. Here's the gist: The Problem: Many businesses and researchers want to leverage LLMs for specific use cases, but fine-tuning these models requires significant technical knowledge and resources. The Solution: A user-friendly web platform where users can: Choose from popular LLM architectures Upload their own dataset or input text Configure fine-tuning parameters through an intuitive interface Automatically fine-tune the model on our GPU infrastructure Download the fine-tuned model or use it via API Key Features: No coding required Scalable cloud infrastructure Support for various fine-tuning techniques (prompt tuning, adapter tuning, full fine-tuning) Job monitoring and results visualization API access for integrated use in applications Target Market: Researchers without extensive ML engineering resources Startups building AI-powered products Enterprises looking to customize LLMs for internal use Monetization: Tiered subscription model based on usage (compute time, model size, etc.) + potential enterprise contracts for high-volume users. I'd really appreciate your thoughts on: Is this solving a real pain point? Would you use a service like this? Why or why not? What features would make this a must-have for you? Any foreseeable obstacles or considerations I'm missing? Suggestions for go-to-market strategy? Thank you!

Founder Pitch: AI Agent for Simplifying Public Cloud Management
reddit
LLM Vibe Score0
Human Vibe Score1
rasvi786This week

Founder Pitch: AI Agent for Simplifying Public Cloud Management

Video to understand : https://youtu.be/9ocUjlUrU\w?si=S0ETDbKSdJqlVDyg Are You Ready to Redefine Cloud Management with AI? Imagine an intelligent AI agent that transforms the complexity of managing public cloud infrastructure into simple, natural language commands. No more navigating through endless configurations or deciphering technical documentation—our AI agent is here to revolutionize the way organizations interact with cloud platforms. About the Project We’re building an AI-powered agent designed to handle public cloud management tasks seamlessly. Whether you’re setting up your organization’s cloud foundation or deploying complex workloads, this AI agent makes it as easy as having a conversation. What Can the AI Agent Do? Cloud Foundation Setup: Example: “Please set up a cloud foundation blueprint for my organization on Google Cloud.”* The AI agent will ask key questions (e.g., organization ID) and guide you through authentication. Once authorized, it sets up the foundation using GCP APIs. Workload Deployment: Example: “Spin up a GKE cluster for me.”* The agent will ask for necessary details (e.g., number of nodes, VPC info), authenticate, and deploy the cluster in minutes. Security and Compliance Validation: Example: “Validate my organization’s cloud setup and check for security vulnerabilities.”* The agent audits your setup, identifies potential risks, and provides actionable insights. Current Progress We’ve developed a working prototype that integrates with major cloud providers like Google Cloud. The AI agent can already: Authenticate with cloud APIs Execute foundational tasks such as setting up organizations and spinning up clusters Perform initial security validations Who I’m Looking For I’m searching for a co-founder with enterprise sales experience and a strategic vision to grow our user base. You will be instrumental in helping us: Build relationships with companies willing to pilot our product Develop go-to-market strategies for enterprise adoption Identify opportunities for partnerships with cloud service providers Your Role As a co-founder, you’ll lead efforts to: Secure Pilot Programs: Identify and onboard enterprises for product trials to gather feedback and refine the solution. Drive Growth: Develop scalable strategies to grow our user base across industries. Market Positioning: Work with me to define our unique value proposition and establish thought leadership in the cloud management space. My Background I bring over a decade of experience in tech, with a strong focus on software engineering and infrastructure. My contributions so far include: Developing the core AI engine and cloud integrations Designing workflows that simplify complex cloud tasks Why Join This Project? Revolutionize Cloud Management: Be part of a project that will redefine how organizations interact with public clouds. Tackle Challenging Problems: Work at the cutting edge of AI and cloud computing. High Growth Potential: Join an industry projected to grow exponentially as enterprises embrace AI-driven automation. Build a Company from Scratch: Shape the product, team, and culture as we grow together. What’s Next? Our immediate priorities include: Expanding the AI agent’s capabilities to support multi-cloud setups. Conducting pilot programs with enterprise clients. Iterating on the product based on real-world feedback. What We Need to Succeed Expertise in enterprise sales and partnerships A deep understanding of enterprise challenges and cloud adoption trends A shared passion for leveraging AI to solve complex problems Let’s work together to build the future of cloud management. If you’re excited about this vision and bring the expertise we need, I’d love to connect and discuss how we can take this project to the next level.

Please, help me to narrow down the list of ideas to pursuit
reddit
LLM Vibe Score0
Human Vibe Score-1
SpiritedSecond4791This week

Please, help me to narrow down the list of ideas to pursuit

Hi guys, I need help to narrow down the possible problems to solve. How do you do it? What do you think about these ideas? All came from real-life problems. Break-It-Down Problem-Solving Assistant Problem: Large, complex projects can feel overwhelming and difficult to tackle. Solution: An AI-guided assistant that analyzes your project goals and automatically breaks them into smaller, manageable tasks. It provides suggested resources and real-time collaboration with team members for smoother task delegation. Personalized Sleep Solutions Problem: Poor sleep quality affects health, productivity, and overall well-being. Solution: An adaptive app that tracks sleep patterns through wearable data and adjusts sleep routines, room settings, and audio cues based on real-time sleep stages for optimal rest. Skill Analysis & Development Tool Problem: It’s challenging to identify valuable skills for career growth and keep up with future demands. Solution: AI-driven skill analysis with a personalized career roadmap that maps out high-demand skills for your specific industry, combined with real-time market trend analysis to suggest learning resources and certifications. Innovator’s Problem Discovery Platform Problem: Innovators struggle to identify real industry problems that need innovative solutions. Solution: An AI-powered platform that gathers and analyzes challenges from different industries, crowdsources ideas, and uses machine learning to highlight innovation opportunities tailored to your skills and interests. High-Earning Career Strategy Platform Problem: Many professionals face challenges in maximizing their earning potential and advancing their careers. Solution: A dynamic career advancement platform that analyzes your skill set, tracks job market trends, and offers personalized mentorship sessions with high-earning professionals in your field, along with salary benchmarking and negotiation tips.

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey
reddit
LLM Vibe Score0
Human Vibe Score0.778
benfromwhereThis week

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey

Update on February 22th: I changed my AI influencer's names because it caused some problems on my business. One year, two AI-powered influencers, and $250K in revenue. Sounds unreal? It’s not. Today, I’m pulling back the curtain on the strategies, tools, and hard-won lessons that took me from concept to a six-figure success story in the AI influencer space. Hey, I'm Ben—a 32-year-old designer who spent the past year navigating the world of AI influencers. Let me clear up any confusion right from the start: I’m not here to sell you anything. This is purely a case study to share what worked, what didn’t, and what I’ve learned along the way. I’ll also make sure to answer all your questions in the comments for free whenever I can, so don’t hesitate to ask. Links to Past Topics: If you're curious about some of the groundwork I covered, check out a few of my earlier posts here: How I Make $10,000 Monthly | AI Influencer Management How I Earned $7000+ in 15 Days | AI Influencer Business Update These earlier posts cover a lot of the backstory, so feel free to explore them before diving into this one. So if you're ready, here is the full story: \---- The idea of creating an AI influencer was one of those “what if” moments that wouldn’t leave my mind. At first, it sounded futuristic—even a bit too ambitious. It all started when I stumbled upon an AI influencer on Instagram with the handle AnnaMaes2000. Her content blew me away—the quality, the detail, and just how real everything looked. I was instantly hooked and ended up going through every post, just trying to figure out how she was pulling this off. That’s when I knew I had to learn how this was done. The next step? YouTube. I dived into videos on Stable Diffusion, soaking up everything I could about creating AI-generated images. Those tutorials taught me the basics and got me up to speed. Then, I created my first AI influencer, let's call her Mel for now. Right after that, to complete the storyline and boost engagement, I introduced Mel's “mother,” Jess. Adding Jess gave the whole project depth and a narrative that drew people in, creating a unique family dynamic that instantly elevated traffic and interest. After thousands of bad photos, hundreds of deleted posts, and months of trial and error, you can now see the quality that defines my current accounts. Here’s a rundown of the tools and checkpoints I’ve used from day one, in order: Fooocus on RunDiffusion — Juggernaut V8 Fooocus on RunDiffusion — Juggernaut V9 Fooocus on PC (locally) — Juggernaut V9 Fooocus on PC (locally) —Lyuyang Mix + Juggernaut V9 Flux on PC (couple of photos only since it's so slow even on RTX 4090) Flux on Fal.ai. \---- There’s no magic Instagram hack that guarantees success, despite what everyone thinks and keeps asking me. Quality content, consistent uploads, and solid craftsmanship are what actually help your photos hit trends and show up on the Explore page. Unlike 95% of low-quality AI accounts out there, I don’t rely on faceswap videos, spam Reels, or go around liking comments on other accounts. My approach is fully organic, focused solely on creating my own unique content. By following Instagram's guidelines to the letter, I've managed to direct some of Mel and Jess' fans over to Patreon and Fanvue. There, for a small subscription fee, fans can access exclusive lingerie content. For those looking for more, higher-tier subscriptions give access to even more premium content. Some possible questions and their answers: No, you can't share hardcore NSFW content on Patreon. You can do that on Fanvue. Yes, you can create AI creators on Fanvue — OnlyFans doesn't allow it. Yes, you can use your own ID to get KYC. Yes, we're telling both Mel and Jess is (or use) AI to generate content. And yes, some people leave and some people still have fun with chatting, having a good time and get perfect content for their needs. And yes, we have a chatter team to work on these accounts. \---- This journey wasn’t all smooth sailing. I faced unexpected roadblocks, like platform restrictions that limited certain types of content, and managing fan expectations was more challenging than anticipated. Staying within guidelines while keeping fans engaged required constant adaptation. These hurdles forced me to get creative, adjust my approach, and learn fast. Once I saw Mel and Jess gaining traction, I knew it was time to scale up. Expanding meant finding new ways to keep content fresh, creating deeper narratives, and considering how to bring even more followers into the fold. My focus turned to building a sustainable model that could grow without sacrificing quality or authenticity. If you’re thinking about diving into AI content creation, here’s my advice: patience, consistency, and a focus on quality are key. Don’t cut corners or rely on quick-fix hacks. Invest time in learning the right tools, creating engaging stories, and building an audience that values what you bring to the table. This approach took me from zero to six figures, and it’s what makes the journey worth it. \---- And finally, here’s the income breakdown that everyone’s curious about: Mel on Fanvue: $82,331.58 (Gross earnings because we have chatter cuts like 15%) Mel on Patreon: $50,865.98 (Net earnings) Jess on Fanvue: $89,068.26 (Gross earnings because we have chatter cuts like 15%) Jess on Patreon: $39,040.70 And thanks to Reddit and my old posts, I got a perfect investor like after 5 months, so this is a "payback" for that. Like I said, I'll answer every question in the comments — take care and let me know.

What role will tech play in sustainability for businesses?
reddit
LLM Vibe Score0
Human Vibe Score0
brycetychsenThis week

What role will tech play in sustainability for businesses?

Have you ever wanted to know how technology is shifting the business sector towards a greener future? Well, wonder no more! In this post, we'll explore the stupendous ways technology is exerting a pivotal role in promoting sustainability within businesses. Smart Energy Management Solutions Gone are the days of wasting energy and money on ineffective practices. With the advent of smart energy management systems, businesses can now optimize their energy usage in real-time. From smart thermostats to AI-powered energy analytics, these solutions help reduce carbon footprints while saving on utility bills. It's a win-win situation for both the environment and the bottom line! Renewable Energy Integration Due to advancements in technology, businesses can now easily integrate solar, wind, and other renewable energy sources into their operations. Not only does this reduce greenhouse gas emissions, but it also shields businesses from the volatility of traditional energy markets. Supply Chain Transparency Ever wondered where your products come from and how they're made? With blockchain technology, businesses can now provide unprecedented transparency throughout their supply chains. From sourcing raw materials to manufacturing processes, consumers can trace the journey of products, ensuring ethical and sustainable practices every step of the way. Data-Driven Sustainability Strategies In the age of big data, knowledge is power, especially when it comes to sustainability. By harnessing the power of data analytics, businesses can identify areas for improvement and implement targeted sustainability strategies. Whether it's optimizing transportation routes or minimizing waste generation, data-driven insights enable businesses to make smarter, greener decisions. Who knew numbers could be so eco-friendly? Eco-Friendly Innovation Last but not least, technology is driving innovation in eco-friendly products and services. From biodegradable packaging to electric vehicles, businesses are constantly pushing the boundaries of sustainability. By embracing these innovations, companies not only reduce their environmental impact but also appeal to eco-conscious consumers. That is it for now, people! From energy management to supply chain transparency, technology is paving the way for a more sustainable future in business. Let's continue to embrace these innovations and work together towards a greener tomorrow.

How to get that big idea for your next business? Use trends!
reddit
LLM Vibe Score0
Human Vibe Score1
IRemember123This week

How to get that big idea for your next business? Use trends!

Hello entrepreneurs and aspiring business owners, I am Mikael and I want to share a post about how to spot business ideas. If you're wondering who the owl is, it's Agent O, my sidekick (please bear with him... or me, if you can). Let's get on to it. So, there are basically two ways of getting ideas for your new business: Find a service, product or experience that's already working. Identify and ride a trend. 🦉 : Third, have a rich relative pass you their business and sip margaritas by the sea while scrolling Reddit for the rest of your life! 🕵️ : Refrain yourself, I just got started ffs, I don't want to get banned! So, what are trends? Trends are patterns of adoption of a product, service or experience by people who want to satisfy a common need. Cool, huh? How trends start Trends emerge and evolve as temporary or permanent solutions to human needs. All products, services and experiences are the expression of human needs manifested through a perceived lack, which we humans interpret as problems. Let me make this more clear. Humans have needs: from basic (food, shelter, safety) to advanced (community, knowledge) to evolved (self actualization, spirituality) and everything in between. Don’t see this as a hierarchy, as it’s usually depicted with Maslow’s pyramid. See it as cycles with different degrees of impact on humans that vary in time and intensity. 🦉 : WHAT!?? 🕵️ : Hear me out… How Trends Affect Society Human needs are physical, emotional, intellectual and spiritual. Every day we feel the impact of those needs with different degrees of required fulfillment. You can’t go on without air for more than a few minutes. You can’t live without food and water for more than a few days. So, when it comes to the needs of the body, these have a shorter timeframe in which they need to be addressed. 🦉 : Ahh, I see what you did there… \\🕵️ \\: Thanks! But you can also live with an unfulfilled need for love or friends for a long time. You can live with a decaying health as well. And you also can live your entire life without finding out if there is a God or not. Humans perceive needs as something they lack within, which in turn is expressed as a problem on the outside. I lack food or water, this will create a problem for my survival. So I need to find food and water in my environment. This lack creates a behavior seeking a product, service or experience to fulfill that need. Makes sense? 🦉 : I just went out and got me a “Mice à la Forest” dinner! 🕵️: Bon appétit! See, Agent O fulfilled a bodily need. That’s what animals do, as they’re driven by instinct and are governed by natural laws (survive, reproduce, sleep, repeat). Humans are driven by more complex needs, as our intellect and emotions allow us to override those basic primary instincts. Why Trends Are Important What an entrepreneur does is to shift the perspective: instead of seeing a lack, he/she sees an opportunity by asking the question: how can I fulfill this need? Or, even better put: how can I help people by solving their problem? That’s the first step to solving a problem: asking a question. That is why the best products are actually problems solved by entrepreneurs who work to solve their own need for a product, service or experience. They then provide it to other people for a cost. Easy, right? That’s what entrepreneurship is: solving a problem. The bigger the problem, the bigger the impact. The bigger the impact, the higher the revenue. It’s easier to understand trends now, isn’t it? You can see that trends are nothing more than the initial adoption of a product, service or experience by a group of people who are looking for a solution to their common need. 🦉 : Did you get that from a book? 🕵️ : You snore when you sleep… ¯\\(\ツ)/\¯ 🦉 : $@#&\*! Hooman! Needs are the foundation on which the modern world is built. Once you understand needs, you fundamentally change your perception of problems into opportunities. This mental shift is the entrepreneurial mindset: where others see problems, you see solutions. Where Do Trends Start So, to recap: human needs are translated into problems. Founders understand the root of the problem (the need) and create products, services, experiences as solutions to those needs. They offer the solution to the public through startups and companies, which belong to a specific niche in a particular industry. 🦉 : Aaah, so that’s why it’s called venture capital? 🕵️ : Yeah, because you’re venturing into a new endeavor to let people know about your solution to their (and ideally your) problem. 🦉 : So if you use ads to market your venture, it’s an adventure? 🕵️ : I see what you did there… If the need behind the adoption is strong and real enough, that trend will translate into a niche within an industry. If the adoption isn’t driven by strong fundamental needs, it will turn into a fad and disappear from the perception of the public, no matter how much marketing money is thrown at it. This happens because the solution (product/service/experience) to the need didn’t create the physical, intellectual or emotional response required to create a recurring behavior around it. Remember this: Problem (why) -> Behavior (how) -> Solution (what) Understand this: there are multiple types of trends. There are product or service trends. There are industry driven trends. There are tendency driven trends, like the emergence of a new paradigm that improves a lot of industries (yes, I’m looking at you, AI). Where Do Trends Come From So now you can see that trends are patterns of adoption related to a specific human need that is addressed through one or multiple products or services. This is a bottom up direction coming from evolution. Multiple trends in different industries also emerge from a theme, which is a bigger vision of a human effort to address a high level problem. This is a top down direction, coming from implementation (by governments, different organizations or other interested parties with the power to influence changes at mass level). Conclusion Now you have a better understanding of trends by looking at them through the lens of human needs. Also, you might also understand time better because you realize that human needs have different degrees of impact in time and intensity. So you now see that trends don’t only relate to individuals, but also to groups of people, from the smallest community to countries and even global needs. That is the reason you’ll sometimes hear some say that time is a flat circle: because clothes change, but humans are quite the same. Needs don’t change a lot in time, just the way we address and solve them. Here’s an interesting game for you: take a look at some behaviors in your life. Which of them are driven by a bodily need, which by an intellectual or emotional one? Which ones are completely automated and you had no idea you were doing? How are these behaviors controlling parts of your life that you were unaware of until now? If you made it this far, thank you for taking the time to read this. I hope you enjoyed it, found it useful and entertaining. Ofc, I value your opinion and welcome it in the comments. Thank you!

How to get that big idea for your next business? Use trends!
reddit
LLM Vibe Score0
Human Vibe Score1
IRemember123This week

How to get that big idea for your next business? Use trends!

Hello entrepreneurs and aspiring business owners, I am Mikael and I want to share a post about how to spot business ideas. If you're wondering who the owl is, it's Agent O, my sidekick (please bear with him... or me, if you can). Let's get on to it. So, there are basically two ways of getting ideas for your new business: Find a service, product or experience that's already working. Identify and ride a trend. 🦉 : Third, have a rich relative pass you their business and sip margaritas by the sea while scrolling Reddit for the rest of your life! 🕵️ : Refrain yourself, I just got started ffs, I don't want to get banned! So, what are trends? Trends are patterns of adoption of a product, service or experience by people who want to satisfy a common need. Cool, huh? How trends start Trends emerge and evolve as temporary or permanent solutions to human needs. All products, services and experiences are the expression of human needs manifested through a perceived lack, which we humans interpret as problems. Let me make this more clear. Humans have needs: from basic (food, shelter, safety) to advanced (community, knowledge) to evolved (self actualization, spirituality) and everything in between. Don’t see this as a hierarchy, as it’s usually depicted with Maslow’s pyramid. See it as cycles with different degrees of impact on humans that vary in time and intensity. 🦉 : WHAT!?? 🕵️ : Hear me out… How Trends Affect Society Human needs are physical, emotional, intellectual and spiritual. Every day we feel the impact of those needs with different degrees of required fulfillment. You can’t go on without air for more than a few minutes. You can’t live without food and water for more than a few days. So, when it comes to the needs of the body, these have a shorter timeframe in which they need to be addressed. 🦉 : Ahh, I see what you did there… \\🕵️ \\: Thanks! But you can also live with an unfulfilled need for love or friends for a long time. You can live with a decaying health as well. And you also can live your entire life without finding out if there is a God or not. Humans perceive needs as something they lack within, which in turn is expressed as a problem on the outside. I lack food or water, this will create a problem for my survival. So I need to find food and water in my environment. This lack creates a behavior seeking a product, service or experience to fulfill that need. Makes sense? 🦉 : I just went out and got me a “Mice à la Forest” dinner! 🕵️: Bon appétit! See, Agent O fulfilled a bodily need. That’s what animals do, as they’re driven by instinct and are governed by natural laws (survive, reproduce, sleep, repeat). Humans are driven by more complex needs, as our intellect and emotions allow us to override those basic primary instincts. Why Trends Are Important What an entrepreneur does is to shift the perspective: instead of seeing a lack, he/she sees an opportunity by asking the question: how can I fulfill this need? Or, even better put: how can I help people by solving their problem? That’s the first step to solving a problem: asking a question. That is why the best products are actually problems solved by entrepreneurs who work to solve their own need for a product, service or experience. They then provide it to other people for a cost. Easy, right? That’s what entrepreneurship is: solving a problem. The bigger the problem, the bigger the impact. The bigger the impact, the higher the revenue. It’s easier to understand trends now, isn’t it? You can see that trends are nothing more than the initial adoption of a product, service or experience by a group of people who are looking for a solution to their common need. 🦉 : Did you get that from a book? 🕵️ : You snore when you sleep… ¯\\(\ツ)/\¯ 🦉 : $@#&\*! Hooman! Needs are the foundation on which the modern world is built. Once you understand needs, you fundamentally change your perception of problems into opportunities. This mental shift is the entrepreneurial mindset: where others see problems, you see solutions. Where Do Trends Start So, to recap: human needs are translated into problems. Founders understand the root of the problem (the need) and create products, services, experiences as solutions to those needs. They offer the solution to the public through startups and companies, which belong to a specific niche in a particular industry. 🦉 : Aaah, so that’s why it’s called venture capital? 🕵️ : Yeah, because you’re venturing into a new endeavor to let people know about your solution to their (and ideally your) problem. 🦉 : So if you use ads to market your venture, it’s an adventure? 🕵️ : I see what you did there… If the need behind the adoption is strong and real enough, that trend will translate into a niche within an industry. If the adoption isn’t driven by strong fundamental needs, it will turn into a fad and disappear from the perception of the public, no matter how much marketing money is thrown at it. This happens because the solution (product/service/experience) to the need didn’t create the physical, intellectual or emotional response required to create a recurring behavior around it. Remember this: Problem (why) -> Behavior (how) -> Solution (what) Understand this: there are multiple types of trends. There are product or service trends. There are industry driven trends. There are tendency driven trends, like the emergence of a new paradigm that improves a lot of industries (yes, I’m looking at you, AI). Where Do Trends Come From So now you can see that trends are patterns of adoption related to a specific human need that is addressed through one or multiple products or services. This is a bottom up direction coming from evolution. Multiple trends in different industries also emerge from a theme, which is a bigger vision of a human effort to address a high level problem. This is a top down direction, coming from implementation (by governments, different organizations or other interested parties with the power to influence changes at mass level). Conclusion Now you have a better understanding of trends by looking at them through the lens of human needs. Also, you might also understand time better because you realize that human needs have different degrees of impact in time and intensity. So you now see that trends don’t only relate to individuals, but also to groups of people, from the smallest community to countries and even global needs. That is the reason you’ll sometimes hear some say that time is a flat circle: because clothes change, but humans are quite the same. Needs don’t change a lot in time, just the way we address and solve them. Here’s an interesting game for you: take a look at some behaviors in your life. Which of them are driven by a bodily need, which by an intellectual or emotional one? Which ones are completely automated and you had no idea you were doing? How are these behaviors controlling parts of your life that you were unaware of until now? If you made it this far, thank you for taking the time to read this. I hope you enjoyed it, found it useful and entertaining. Ofc, I value your opinion and welcome it in the comments. Thank you!

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey
reddit
LLM Vibe Score0
Human Vibe Score0.778
benfromwhereThis week

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey

Update on February 22th: I changed my AI influencer's names because it caused some problems on my business. One year, two AI-powered influencers, and $250K in revenue. Sounds unreal? It’s not. Today, I’m pulling back the curtain on the strategies, tools, and hard-won lessons that took me from concept to a six-figure success story in the AI influencer space. Hey, I'm Ben—a 32-year-old designer who spent the past year navigating the world of AI influencers. Let me clear up any confusion right from the start: I’m not here to sell you anything. This is purely a case study to share what worked, what didn’t, and what I’ve learned along the way. I’ll also make sure to answer all your questions in the comments for free whenever I can, so don’t hesitate to ask. Links to Past Topics: If you're curious about some of the groundwork I covered, check out a few of my earlier posts here: How I Make $10,000 Monthly | AI Influencer Management How I Earned $7000+ in 15 Days | AI Influencer Business Update These earlier posts cover a lot of the backstory, so feel free to explore them before diving into this one. So if you're ready, here is the full story: \---- The idea of creating an AI influencer was one of those “what if” moments that wouldn’t leave my mind. At first, it sounded futuristic—even a bit too ambitious. It all started when I stumbled upon an AI influencer on Instagram with the handle AnnaMaes2000. Her content blew me away—the quality, the detail, and just how real everything looked. I was instantly hooked and ended up going through every post, just trying to figure out how she was pulling this off. That’s when I knew I had to learn how this was done. The next step? YouTube. I dived into videos on Stable Diffusion, soaking up everything I could about creating AI-generated images. Those tutorials taught me the basics and got me up to speed. Then, I created my first AI influencer, let's call her Mel for now. Right after that, to complete the storyline and boost engagement, I introduced Mel's “mother,” Jess. Adding Jess gave the whole project depth and a narrative that drew people in, creating a unique family dynamic that instantly elevated traffic and interest. After thousands of bad photos, hundreds of deleted posts, and months of trial and error, you can now see the quality that defines my current accounts. Here’s a rundown of the tools and checkpoints I’ve used from day one, in order: Fooocus on RunDiffusion — Juggernaut V8 Fooocus on RunDiffusion — Juggernaut V9 Fooocus on PC (locally) — Juggernaut V9 Fooocus on PC (locally) —Lyuyang Mix + Juggernaut V9 Flux on PC (couple of photos only since it's so slow even on RTX 4090) Flux on Fal.ai. \---- There’s no magic Instagram hack that guarantees success, despite what everyone thinks and keeps asking me. Quality content, consistent uploads, and solid craftsmanship are what actually help your photos hit trends and show up on the Explore page. Unlike 95% of low-quality AI accounts out there, I don’t rely on faceswap videos, spam Reels, or go around liking comments on other accounts. My approach is fully organic, focused solely on creating my own unique content. By following Instagram's guidelines to the letter, I've managed to direct some of Mel and Jess' fans over to Patreon and Fanvue. There, for a small subscription fee, fans can access exclusive lingerie content. For those looking for more, higher-tier subscriptions give access to even more premium content. Some possible questions and their answers: No, you can't share hardcore NSFW content on Patreon. You can do that on Fanvue. Yes, you can create AI creators on Fanvue — OnlyFans doesn't allow it. Yes, you can use your own ID to get KYC. Yes, we're telling both Mel and Jess is (or use) AI to generate content. And yes, some people leave and some people still have fun with chatting, having a good time and get perfect content for their needs. And yes, we have a chatter team to work on these accounts. \---- This journey wasn’t all smooth sailing. I faced unexpected roadblocks, like platform restrictions that limited certain types of content, and managing fan expectations was more challenging than anticipated. Staying within guidelines while keeping fans engaged required constant adaptation. These hurdles forced me to get creative, adjust my approach, and learn fast. Once I saw Mel and Jess gaining traction, I knew it was time to scale up. Expanding meant finding new ways to keep content fresh, creating deeper narratives, and considering how to bring even more followers into the fold. My focus turned to building a sustainable model that could grow without sacrificing quality or authenticity. If you’re thinking about diving into AI content creation, here’s my advice: patience, consistency, and a focus on quality are key. Don’t cut corners or rely on quick-fix hacks. Invest time in learning the right tools, creating engaging stories, and building an audience that values what you bring to the table. This approach took me from zero to six figures, and it’s what makes the journey worth it. \---- And finally, here’s the income breakdown that everyone’s curious about: Mel on Fanvue: $82,331.58 (Gross earnings because we have chatter cuts like 15%) Mel on Patreon: $50,865.98 (Net earnings) Jess on Fanvue: $89,068.26 (Gross earnings because we have chatter cuts like 15%) Jess on Patreon: $39,040.70 And thanks to Reddit and my old posts, I got a perfect investor like after 5 months, so this is a "payback" for that. Like I said, I'll answer every question in the comments — take care and let me know.

AI SaaS: A website to fine-tune LLM model according to your requirements
reddit
LLM Vibe Score0
Human Vibe Score1
Dangerous_Ferret3362This week

AI SaaS: A website to fine-tune LLM model according to your requirements

Hey fellow entrepreneurs and AI enthusiasts! I'm exploring a business idea and would love your thoughts and feedback. The concept is a SaaS platform that allows users to easily fine-tune large language models (LLMs) on their own datasets without needing deep technical expertise. Here's the gist: The Problem: Many businesses and researchers want to leverage LLMs for specific use cases, but fine-tuning these models requires significant technical knowledge and resources. The Solution: A user-friendly web platform where users can: Choose from popular LLM architectures Upload their own dataset or input text Configure fine-tuning parameters through an intuitive interface Automatically fine-tune the model on our GPU infrastructure Download the fine-tuned model or use it via API Key Features: No coding required Scalable cloud infrastructure Support for various fine-tuning techniques (prompt tuning, adapter tuning, full fine-tuning) Job monitoring and results visualization API access for integrated use in applications Target Market: Researchers without extensive ML engineering resources Startups building AI-powered products Enterprises looking to customize LLMs for internal use Monetization: Tiered subscription model based on usage (compute time, model size, etc.) + potential enterprise contracts for high-volume users. I'd really appreciate your thoughts on: Is this solving a real pain point? Would you use a service like this? Why or why not? What features would make this a must-have for you? Any foreseeable obstacles or considerations I'm missing? Suggestions for go-to-market strategy? Thank you!

Can AI Mentorship and Community Support Help Entrepreneurs Succeed?
reddit
LLM Vibe Score0
Human Vibe Score1
Expensive_Ad_1176This week

Can AI Mentorship and Community Support Help Entrepreneurs Succeed?

Starting a business can often feel like you're flying blind, especially without a mentor to guide you. But what if you could tap into AI-powered mentorship tools and a supportive community to get advice and feedback whenever you need it? 🚀 AI mentorship offers personalized guidance and structured frameworks, minus the need for traditional face-to-face time. And platforms like this one allow us to connect, share experiences, and learn from each other. It’s a game-changer, right? Here’s what I’m curious about: Have you tried AI mentorship tools? What was your experience? How do you currently get advice and feedback on your business? Do you think mentorship should always be face-to-face, or can online tools and communities play a big role in helping entrepreneurs succeed? Would you consider using structured learning tools (like lesson-based frameworks or step-by-step guidance) to guide your entrepreneurship journey? I’m working on Procasio, an educational entrepreneurship app designed to promote inclusivity and accessibility. It would combine AI mentorship, structured learning paths, gamified elements, and case studies, helping small business owners, teachers, students, and aspiring entrepreneurs learn effectively without overwhelming costs. 🎓💡 The app would include: Discussion posts and messaging for real-time advice. Goal setting and personalized learning recommendations. Case studies and practical scenarios to put theory into action. A low-cost, accessible approach for entrepreneurs at any stage. I’d love to hear your thoughts—do you think AI-powered mentorship and structured learning can make entrepreneurship education easier and more effective?

Presenting my fresh new ideas for Google! Warning: These might make TOO much money
reddit
LLM Vibe Score0
Human Vibe Score1
Good0timesThis week

Presenting my fresh new ideas for Google! Warning: These might make TOO much money

It's well known that Google likes its safe avenues, such as email, word processing, and file storage. One would think that it has done everything it can. I disagree! There are so many untapped markets out there to compete in and employ its digital ~~monopoly~~ advantage. behold the future of Google projects and even more wealth at your fingertips: Google Street Crime: An app that uses AI/ML to identify easy places to rob or even just people to hit. Google Jail: The perfect way to communicate with your family, lawyers, and drug dealers. Google Prostitution: Similar to the 'Uber' style of business, albeit big pimping. Google Human Trafficking: Sell your victims on a competitive global marketplace! Or organs. Google Slavery: Kind of like the same but probably with different taxes. Google Death Sentence: Automated appeal process. Free cookies with all fails! Google Cartel: Another kind of marketplace. Premium gives you a free chainsaw. Google Misery: A VR-powered world which portrays hell. It's an improvement. Google Domination: The blissful removal of this tedious experiment of democracy. Google Religion: The worship of Google as a lovecraftian but benevolent creator of all existence. All of these shall be accompanied by chirpy music, minimalist graphics, and deliriously happy animated cartoons. Well now you've got the ideas so chase that money you freaks! It's right there. Smell of money! Smell of money!

Founder Pitch: AI Agent for Simplifying Public Cloud Management
reddit
LLM Vibe Score0
Human Vibe Score1
rasvi786This week

Founder Pitch: AI Agent for Simplifying Public Cloud Management

Video to understand : https://youtu.be/9ocUjlUrU\w?si=S0ETDbKSdJqlVDyg Are You Ready to Redefine Cloud Management with AI? Imagine an intelligent AI agent that transforms the complexity of managing public cloud infrastructure into simple, natural language commands. No more navigating through endless configurations or deciphering technical documentation—our AI agent is here to revolutionize the way organizations interact with cloud platforms. About the Project We’re building an AI-powered agent designed to handle public cloud management tasks seamlessly. Whether you’re setting up your organization’s cloud foundation or deploying complex workloads, this AI agent makes it as easy as having a conversation. What Can the AI Agent Do? Cloud Foundation Setup: Example: “Please set up a cloud foundation blueprint for my organization on Google Cloud.”* The AI agent will ask key questions (e.g., organization ID) and guide you through authentication. Once authorized, it sets up the foundation using GCP APIs. Workload Deployment: Example: “Spin up a GKE cluster for me.”* The agent will ask for necessary details (e.g., number of nodes, VPC info), authenticate, and deploy the cluster in minutes. Security and Compliance Validation: Example: “Validate my organization’s cloud setup and check for security vulnerabilities.”* The agent audits your setup, identifies potential risks, and provides actionable insights. Current Progress We’ve developed a working prototype that integrates with major cloud providers like Google Cloud. The AI agent can already: Authenticate with cloud APIs Execute foundational tasks such as setting up organizations and spinning up clusters Perform initial security validations Who I’m Looking For I’m searching for a co-founder with enterprise sales experience and a strategic vision to grow our user base. You will be instrumental in helping us: Build relationships with companies willing to pilot our product Develop go-to-market strategies for enterprise adoption Identify opportunities for partnerships with cloud service providers Your Role As a co-founder, you’ll lead efforts to: Secure Pilot Programs: Identify and onboard enterprises for product trials to gather feedback and refine the solution. Drive Growth: Develop scalable strategies to grow our user base across industries. Market Positioning: Work with me to define our unique value proposition and establish thought leadership in the cloud management space. My Background I bring over a decade of experience in tech, with a strong focus on software engineering and infrastructure. My contributions so far include: Developing the core AI engine and cloud integrations Designing workflows that simplify complex cloud tasks Why Join This Project? Revolutionize Cloud Management: Be part of a project that will redefine how organizations interact with public clouds. Tackle Challenging Problems: Work at the cutting edge of AI and cloud computing. High Growth Potential: Join an industry projected to grow exponentially as enterprises embrace AI-driven automation. Build a Company from Scratch: Shape the product, team, and culture as we grow together. What’s Next? Our immediate priorities include: Expanding the AI agent’s capabilities to support multi-cloud setups. Conducting pilot programs with enterprise clients. Iterating on the product based on real-world feedback. What We Need to Succeed Expertise in enterprise sales and partnerships A deep understanding of enterprise challenges and cloud adoption trends A shared passion for leveraging AI to solve complex problems Let’s work together to build the future of cloud management. If you’re excited about this vision and bring the expertise we need, I’d love to connect and discuss how we can take this project to the next level.

Please rate my business idea aimed at emerging markets
reddit
LLM Vibe Score0
Human Vibe Score1
Whole_Ad_9002This week

Please rate my business idea aimed at emerging markets

Imagine a single platform that transforms your business operations effortlessly. Think of it like a Swiss Army knife for your business. First Steps (Every Business Needs These): Our website builder is your first step online. Just pick what you want, drag it where you need it, and you've got a professional website or online store. No tech skills needed. Perfect for getting your business visible to customers. Our security tool comes next because every business needs protection. It's like a security guard for your files and data, keeping hackers out and making sure you never lose important work. Think of it as insurance for your digital business. Cybersecurity and backups combined. Growing Business Needs: As your team grows, our digital office keeps everyone connected. Email, chat, share files, and track projects in one place. It's like having everyone in the same room, even when working remotely. This becomes essential once you have more than a few people. When paperwork starts piling up, our AI helper steps in. It's like having a smart assistant who learns your business and helps everyone get more done - handling routine tasks, summarizing meetings, and suggesting better ways to work. Scaling Up (For Established Businesses): Once you're handling lots of data, our eco-friendly cloud storage becomes crucial. Your business gets fast, reliable storage while helping the environment. We use servers powered by renewable energy, so you can grow sustainably. Our data transfer tool becomes important when you're working with multiple systems and need to move files around. It's like having a digital moving service that safely carries your files between different cloud services. With 20+ services to connect to cloud to cloud or on prem to cloud, its easy and fast. Advanced Needs: When your business processes get complex, our workflow automation tool helps you set up automatic systems for repetitive tasks. It's like training a robot to handle your routine work while your team focuses on growth. Simple no-code automation when you need it. Everything works on its own, but they're designed to work even better together. And since it's all in one place, you only need one password and one monthly bill. Most businesses save about half their tech costs by switching to our platform. Cost Benefits: Single subscription replaces multiple vendor contracts Reduces IT overhead by 40-60% on average No need for expensive technical specialists Predictable monthly costs Scales pricing with your usage

Made 60k mrr for a business by just lead nurturing. Need suggestions and validation.
reddit
LLM Vibe Score0
Human Vibe Score1
Alarmed-Argument-605This week

Made 60k mrr for a business by just lead nurturing. Need suggestions and validation.

Apart from the story I need a suggestion and validation here. It's a bit long, skip to tl;dr if you couldn't handle length. A few days ago, I saw a person on Reddit sharing his struggles that, Even after generating a lot of leads from ads of Meta and Google (even with lowest cpc cpa cpl), he was not able to convert them into sales. Out of curiosity I dm'ed him with all fancy services that I offer and expressed that as a agency I would work with him for monthly recurring fee. He suggested for one time consulting fee, I agreed. It was literally a eye opener for me. This guy is in coaching business offering courses for people. His niche was too vague. Courses were on mindset coaching, confidence and public speaking coaching, right attitude coaching, manifestation coaching and all crap shits related to this. At first I thought he was not getting sales because who will pay for all this craps. I openly discussed with him that he has to change what he offers because, if I saw this ad I wouldn't buy this for sure. He then showed me how much money people offering similar service are making . I was literally taken back. He was part of a influencer group (the main guy who encourages these guys to start coaching business, looks like some mlm shit) where people post their succes stories. Literally lot of guys were making above 150k and 200k per month. Even with very basic landing page and average offer They are still winning. Here's where it gets interesting. I tried to clone everything that the top people in this industry are doing in marketing from end to end.( like the same landing page, bonus offers around 50k, exclusive community, free 1 on 1 calls for twice a month).Nothing worked for a month and later surprisingly even the sales started dropping a bit more. I got really confused here. So to do a discovery I went and purchased the competitor course and Man I was literally taken back. Like he has automated everything from end to end. You click the ad, see vsl, you have to fill a form and join a free Skool community where he gives away free stuffs and post success stories of people who took the course. Now every part of this journey you will get a follow up mail and follow up sms. Like after filling the form. after that now if you join and don't purchase the course you will be pampered with email and sms filled with success stories. For sure anybody will be tempted to buy the course. Here is the key take away. He was able to make more sales because he was very successful in nurturing the leads with follow ups after follow ups. Even after you purchased his course he is making passive income from 1 on calls and bonus live webinars. So follow ups will be for 1 on 1 calls and webinars after the course is over. Core point is our guy even after spending 2 to 3k per month on ads was not able to bring huge sales like competitors because he failed the nuture them. Even after making the same offers and the same patterns of marketing as competitors, the sales declined because people thought this is some spam that everyone is doing because the template of the ads was very professional and similar. suprising one is people fall for basic templates thinking it's a underrated one. so what we did here is we integrated a few softwares into one and set up all same webinars, automated email and sms follow ups, ad managers for stats, launched him a free LMS platform where without any additional fees so he can uploaded unlimited courses, skool like community and add product's like Shopify ( he was selling few merchandise with his brand name on) where he can add unlimited products with connection to all payment gateway, integrated with crm with unlimited contacts, workflow and lead nurturing with calender syncing for 1 on 1 calls. But these are a bit old school, what we did was even better. integrated a conversational ai with all of his sales platforms and gave a nocode automation builder with ai for the workflow. we also set him up with a ai voice agent that's automatically calls and markets for his course and also replies for queries when called. we also set up him a dedicated afflitate manager portal with automated commissions. Though he didn't cross 100k Mark, He did a great number after this. He was struggling with 6k sales, now he has reached somewhere mid of 45k to 50k mrr. Max he hit was 61.8k. I see this a big difference.So one small thing, nurturing the lead can bring you immense sales. To set up all of this it costs around 1.2k monthly for me with all the bills. ( I know there are few free for Individual user platforms out there, but It gets very costly when you switch to their premium plans. with heavy volumes you would require more than premium they offer.) I offered him like 3k per month to work as a agency for him who takes care of all these stuffs. He declined and offered for one time set up fee stating that he will pay 1.2k directly. The one time fee was also a bit low, though I agreed since this was a learning for me. what happened next after that is, he referred me to a few other people in the same niche. But the problem is they are not interested in spending 1 to 2 k in bills for software. They requested that if, will I be able to provide the saas alone for less than 500 dollars with one time set up fee. I haven't responded yet since I have to take an enterprise plan for all the software used and pay full advance price for billings. Then to break even that I have to make minimum 50 or odd users for that. let's grantly say 100 users with all other future costs. So here's what I'm planning to do. I'm planning to offer this as saas for let's say 239 dollars per month. with may or may not one time set up fee. ( I checked the entire internet, there is no single person offering at this price point for unlimited. Also one can easily start their marketing agency with this.) The suggestion and validation that I need here is 1.are you going through the same struggles or faced these struggles? would you be interested to buy at 239 dollars per month? let's say you're from a different niche, Did the features I told were okay for you or you need something specific for your industry that you will be interested in buying? please answer in comments and if you will purchase for this price let me know in comments/dms. I will take that into account and if the response rate is above 100 queries, then will integrate this and sell for that price. (ps: If you see this post on similar subs, please bear cause I'm trying to get suggestions from different POV) tl;dr - lead nurturing can massively boost sales *I made a software integration for a client for a 1.2k per month billing and here I want to know if more than 100 people are interested so that I will make this into my own saas and sell it for like a cheap price of 239 dollars per month TIA.

What do you think of SaaS 2.0: Service-as-a-Software?
reddit
LLM Vibe Score0
Human Vibe Score1
FrenzyOfLifeThis week

What do you think of SaaS 2.0: Service-as-a-Software?

A new term has recently emerged in the business world: Service-as-a-Software a.k.a. SaaS 2.0 In general, some authors of articles promoting this term assume that the new and rapidly growing possibilities offered by AI and automation mean that problems that were previously too individual or support-intensive can now be tackled. The focus is on (human) service on the customer side and the background processes in the company are fully AI-supported and automated. Unlike traditional SaaS, no software is primarily offered here as self-use. In other words: "Service as a Software" (SaaS 2.0) is a new type of business model that mixes software automation with real human support. Unlike traditional SaaS, which provides self-service tools for users to solve problems on their own, SaaS 2.0 focuses on delivering results by combining technology with human expertise. In this model, software handles repetitive tasks like data processing, scheduling, or matching, while humans step in to provide guidance, handle exceptions, or solve complex issues. This approach is often called Human-in-the-Loop because humans are actively involved in key parts of the process, ensuring a personalized and empathetic experience for the customer. SaaS 2.0 is especially useful in industries like healthcare, education, or elderly care placement, where trust and personalization are critical. For example, a traditional SaaS might offer a tool to search for care homes, while a SaaS 2.0 solution would also provide a care consultant to help families make the best choice. In this case no traditional marketplace is needed where the supply and demand side used to be scaled simultaneously. Instead, an AI can now search for the best match for a place in a retirement home and a human in the loop can be the external face for the customer and the retirement homes and thus act as an agent. By automating routine tasks and using humans for high-value touchpoints, SaaS 2.0 delivers better outcomes, builds stronger relationships with customers, and stands out from traditional software that relies only on automation. What do you think about the potential of this concept?

Please rate my business idea aimed at emerging markets
reddit
LLM Vibe Score0
Human Vibe Score1
Whole_Ad_9002This week

Please rate my business idea aimed at emerging markets

Imagine a single platform that transforms your business operations effortlessly. Think of it like a Swiss Army knife for your business. First Steps (Every Business Needs These): Our website builder is your first step online. Just pick what you want, drag it where you need it, and you've got a professional website or online store. No tech skills needed. Perfect for getting your business visible to customers. Our security tool comes next because every business needs protection. It's like a security guard for your files and data, keeping hackers out and making sure you never lose important work. Think of it as insurance for your digital business. Cybersecurity and backups combined. Growing Business Needs: As your team grows, our digital office keeps everyone connected. Email, chat, share files, and track projects in one place. It's like having everyone in the same room, even when working remotely. This becomes essential once you have more than a few people. When paperwork starts piling up, our AI helper steps in. It's like having a smart assistant who learns your business and helps everyone get more done - handling routine tasks, summarizing meetings, and suggesting better ways to work. Scaling Up (For Established Businesses): Once you're handling lots of data, our eco-friendly cloud storage becomes crucial. Your business gets fast, reliable storage while helping the environment. We use servers powered by renewable energy, so you can grow sustainably. Our data transfer tool becomes important when you're working with multiple systems and need to move files around. It's like having a digital moving service that safely carries your files between different cloud services. With 20+ services to connect to cloud to cloud or on prem to cloud, its easy and fast. Advanced Needs: When your business processes get complex, our workflow automation tool helps you set up automatic systems for repetitive tasks. It's like training a robot to handle your routine work while your team focuses on growth. Simple no-code automation when you need it. Everything works on its own, but they're designed to work even better together. And since it's all in one place, you only need one password and one monthly bill. Most businesses save about half their tech costs by switching to our platform. Cost Benefits: Single subscription replaces multiple vendor contracts Reduces IT overhead by 40-60% on average No need for expensive technical specialists Predictable monthly costs Scales pricing with your usage

The Weekly Brief for anyone looking to incorporate AI into their business.
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Business_BriefThis week

The Weekly Brief for anyone looking to incorporate AI into their business.

Good morning and happy Sunday. Today is Sunday and you know what that means. The weekly brief. Covering all of last week’s most important AI business related stories. Here are some of the biggest stories: Claude the newest generative AI. Amazon to change up its search. AI leaders Testify. Meta Open sources its LLM. Voice Actors Struggle Growing AI innovations has led to a struggle for many voice actors. As AI powered voice technology is progressing everyday jobs are becoming more and more scarce. With many publishers already leaning towards replacing many of their voice actors for faster, cheaper, and more efficient AI voices. Meet Claude Anthropic, an AI company founded by ex-OpenAI employee released their generative AI called Claude. Some key aspects of their model is the ability to give more correct and less harmful answers, and perform similar tasks that many other generative AI’s can do. A keynote is that Google has invested 300milloion into the company, which is a direct competitor to their AI Bard. Interesting to see how that will play out. Amazon Changes to Change up Search A new job description at Amazon may have hinted towards their future plans for AI. The description under software developer read “reimagining Amazon Search with an interactive conversational experience”. This may hint towards a generative AI search experience in Amazon for customers. ChatGPT User Get More Access Premium ChatGPT users got access to Web browsing and plugins. This is a crucial step for OpenAI as they plan to pivot to a more assist type AI. While at the same time continuing to research and develop their AI models. This move puts a lot of pressure on Google to hopefully step up their game. AI Leaders Testify This Wednesday AI leaders (Sam Altman, Christina Montgomery and Gary Marcus) all testified before congress about AI regulation. They were asked many questions about AI regulation but came up with two solutions. FDA-Like Approval Processing: AI developing companies are open to safety checks, audits, licensing and risk review. Precision Approach: Develop risk rules, provide explanations and provide guidelines for risks, encourage transparency around AI companies, finally assess impact of AI technologies. Meta Open Sourcing Thursday Meta open sourced this coding for their LLM. As the company wants to see the use of its LLM to help drive innovation, inspire smaller companies, and overall develop better AI technologies. Comes as an interesting move as competitors try and keep their AI’s an insider secret.

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey
reddit
LLM Vibe Score0
Human Vibe Score0.778
benfromwhereThis week

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey

Update on February 22th: I changed my AI influencer's names because it caused some problems on my business. One year, two AI-powered influencers, and $250K in revenue. Sounds unreal? It’s not. Today, I’m pulling back the curtain on the strategies, tools, and hard-won lessons that took me from concept to a six-figure success story in the AI influencer space. Hey, I'm Ben—a 32-year-old designer who spent the past year navigating the world of AI influencers. Let me clear up any confusion right from the start: I’m not here to sell you anything. This is purely a case study to share what worked, what didn’t, and what I’ve learned along the way. I’ll also make sure to answer all your questions in the comments for free whenever I can, so don’t hesitate to ask. Links to Past Topics: If you're curious about some of the groundwork I covered, check out a few of my earlier posts here: How I Make $10,000 Monthly | AI Influencer Management How I Earned $7000+ in 15 Days | AI Influencer Business Update These earlier posts cover a lot of the backstory, so feel free to explore them before diving into this one. So if you're ready, here is the full story: \---- The idea of creating an AI influencer was one of those “what if” moments that wouldn’t leave my mind. At first, it sounded futuristic—even a bit too ambitious. It all started when I stumbled upon an AI influencer on Instagram with the handle AnnaMaes2000. Her content blew me away—the quality, the detail, and just how real everything looked. I was instantly hooked and ended up going through every post, just trying to figure out how she was pulling this off. That’s when I knew I had to learn how this was done. The next step? YouTube. I dived into videos on Stable Diffusion, soaking up everything I could about creating AI-generated images. Those tutorials taught me the basics and got me up to speed. Then, I created my first AI influencer, let's call her Mel for now. Right after that, to complete the storyline and boost engagement, I introduced Mel's “mother,” Jess. Adding Jess gave the whole project depth and a narrative that drew people in, creating a unique family dynamic that instantly elevated traffic and interest. After thousands of bad photos, hundreds of deleted posts, and months of trial and error, you can now see the quality that defines my current accounts. Here’s a rundown of the tools and checkpoints I’ve used from day one, in order: Fooocus on RunDiffusion — Juggernaut V8 Fooocus on RunDiffusion — Juggernaut V9 Fooocus on PC (locally) — Juggernaut V9 Fooocus on PC (locally) —Lyuyang Mix + Juggernaut V9 Flux on PC (couple of photos only since it's so slow even on RTX 4090) Flux on Fal.ai. \---- There’s no magic Instagram hack that guarantees success, despite what everyone thinks and keeps asking me. Quality content, consistent uploads, and solid craftsmanship are what actually help your photos hit trends and show up on the Explore page. Unlike 95% of low-quality AI accounts out there, I don’t rely on faceswap videos, spam Reels, or go around liking comments on other accounts. My approach is fully organic, focused solely on creating my own unique content. By following Instagram's guidelines to the letter, I've managed to direct some of Mel and Jess' fans over to Patreon and Fanvue. There, for a small subscription fee, fans can access exclusive lingerie content. For those looking for more, higher-tier subscriptions give access to even more premium content. Some possible questions and their answers: No, you can't share hardcore NSFW content on Patreon. You can do that on Fanvue. Yes, you can create AI creators on Fanvue — OnlyFans doesn't allow it. Yes, you can use your own ID to get KYC. Yes, we're telling both Mel and Jess is (or use) AI to generate content. And yes, some people leave and some people still have fun with chatting, having a good time and get perfect content for their needs. And yes, we have a chatter team to work on these accounts. \---- This journey wasn’t all smooth sailing. I faced unexpected roadblocks, like platform restrictions that limited certain types of content, and managing fan expectations was more challenging than anticipated. Staying within guidelines while keeping fans engaged required constant adaptation. These hurdles forced me to get creative, adjust my approach, and learn fast. Once I saw Mel and Jess gaining traction, I knew it was time to scale up. Expanding meant finding new ways to keep content fresh, creating deeper narratives, and considering how to bring even more followers into the fold. My focus turned to building a sustainable model that could grow without sacrificing quality or authenticity. If you’re thinking about diving into AI content creation, here’s my advice: patience, consistency, and a focus on quality are key. Don’t cut corners or rely on quick-fix hacks. Invest time in learning the right tools, creating engaging stories, and building an audience that values what you bring to the table. This approach took me from zero to six figures, and it’s what makes the journey worth it. \---- And finally, here’s the income breakdown that everyone’s curious about: Mel on Fanvue: $82,331.58 (Gross earnings because we have chatter cuts like 15%) Mel on Patreon: $50,865.98 (Net earnings) Jess on Fanvue: $89,068.26 (Gross earnings because we have chatter cuts like 15%) Jess on Patreon: $39,040.70 And thanks to Reddit and my old posts, I got a perfect investor like after 5 months, so this is a "payback" for that. Like I said, I'll answer every question in the comments — take care and let me know.

An honest opinion about start-up idea
reddit
LLM Vibe Score0
Human Vibe Score1
Comfortable_Mud1233This week

An honest opinion about start-up idea

You will be helpful to us especially if you have worked with a lot of data (whether in a corporation or somewhere else). We aim to develop a document library platform that aggregates data from various storage services such as Amazon S3 (AWS) and Google Cloud Storage (GCP). The platform serves as a centralized interface or "panel" where users within an organization can access and display documents stored across different sources. Key features include: Data aggregation without storage: The platform pulls data from multiple sources but does not store it locally. This approach minimizes data redundancy and storage costs. AI-powered semantic search: Utilizes artificial intelligence to perform semantic searches across files, enabling users to find documents based on context and meaning rather than just keywords. Tagging and versioning: Supports the addition of tags for better categorization and tracking of different versions of files. The solution targets companies handling large volumes of data and documents dispersed across various storage services. Strengths we found: Non-invasive integration: Eliminates the need for data migration, reducing setup time and complexity. Enhanced search capabilities: AI-driven semantic search outperforms basic keyword searches, saving time. Cross-platform functionality: Provides a level of interoperability that competitors lack. Cost efficiency: Avoids additional storage costs and reduces time spent searching for documents. Weaknesses that we see: Limited feature set compared to ECMs: May lack some advanced features like workflow automation, collaboration tools, and compliance auditing provided by ECMs. We're new: so no trust. Is this something that companies would want to integrate and pay for? Thanks a lot, it can save us a lot of time :)

What do you think of SaaS 2.0: Service-as-a-Software?
reddit
LLM Vibe Score0
Human Vibe Score1
FrenzyOfLifeThis week

What do you think of SaaS 2.0: Service-as-a-Software?

A new term has recently emerged in the business world: Service-as-a-Software a.k.a. SaaS 2.0 In general, some authors of articles promoting this term assume that the new and rapidly growing possibilities offered by AI and automation mean that problems that were previously too individual or support-intensive can now be tackled. The focus is on (human) service on the customer side and the background processes in the company are fully AI-supported and automated. Unlike traditional SaaS, no software is primarily offered here as self-use. In other words: "Service as a Software" (SaaS 2.0) is a new type of business model that mixes software automation with real human support. Unlike traditional SaaS, which provides self-service tools for users to solve problems on their own, SaaS 2.0 focuses on delivering results by combining technology with human expertise. In this model, software handles repetitive tasks like data processing, scheduling, or matching, while humans step in to provide guidance, handle exceptions, or solve complex issues. This approach is often called Human-in-the-Loop because humans are actively involved in key parts of the process, ensuring a personalized and empathetic experience for the customer. SaaS 2.0 is especially useful in industries like healthcare, education, or elderly care placement, where trust and personalization are critical. For example, a traditional SaaS might offer a tool to search for care homes, while a SaaS 2.0 solution would also provide a care consultant to help families make the best choice. In this case no traditional marketplace is needed where the supply and demand side used to be scaled simultaneously. Instead, an AI can now search for the best match for a place in a retirement home and a human in the loop can be the external face for the customer and the retirement homes and thus act as an agent. By automating routine tasks and using humans for high-value touchpoints, SaaS 2.0 delivers better outcomes, builds stronger relationships with customers, and stands out from traditional software that relies only on automation. What do you think about the potential of this concept?

Please rate my business idea aimed at emerging markets
reddit
LLM Vibe Score0
Human Vibe Score1
Whole_Ad_9002This week

Please rate my business idea aimed at emerging markets

Imagine a single platform that transforms your business operations effortlessly. Think of it like a Swiss Army knife for your business. First Steps (Every Business Needs These): Our website builder is your first step online. Just pick what you want, drag it where you need it, and you've got a professional website or online store. No tech skills needed. Perfect for getting your business visible to customers. Our security tool comes next because every business needs protection. It's like a security guard for your files and data, keeping hackers out and making sure you never lose important work. Think of it as insurance for your digital business. Cybersecurity and backups combined. Growing Business Needs: As your team grows, our digital office keeps everyone connected. Email, chat, share files, and track projects in one place. It's like having everyone in the same room, even when working remotely. This becomes essential once you have more than a few people. When paperwork starts piling up, our AI helper steps in. It's like having a smart assistant who learns your business and helps everyone get more done - handling routine tasks, summarizing meetings, and suggesting better ways to work. Scaling Up (For Established Businesses): Once you're handling lots of data, our eco-friendly cloud storage becomes crucial. Your business gets fast, reliable storage while helping the environment. We use servers powered by renewable energy, so you can grow sustainably. Our data transfer tool becomes important when you're working with multiple systems and need to move files around. It's like having a digital moving service that safely carries your files between different cloud services. With 20+ services to connect to cloud to cloud or on prem to cloud, its easy and fast. Advanced Needs: When your business processes get complex, our workflow automation tool helps you set up automatic systems for repetitive tasks. It's like training a robot to handle your routine work while your team focuses on growth. Simple no-code automation when you need it. Everything works on its own, but they're designed to work even better together. And since it's all in one place, you only need one password and one monthly bill. Most businesses save about half their tech costs by switching to our platform. Cost Benefits: Single subscription replaces multiple vendor contracts Reduces IT overhead by 40-60% on average No need for expensive technical specialists Predictable monthly costs Scales pricing with your usage

TiCs -where innovation meets intelligence
reddit
LLM Vibe Score0
Human Vibe Score1
MohammadBaisThis week

TiCs -where innovation meets intelligence

Be Part of India’s AI Revolution – Join the TiCs Movement! We are TiCs (Tuba International Cooperative Society)—India’s first global AI powerhouse. We’re not just building a company; we’re launching a movement that will redefine AI-driven healthcare, fitness, and well-being. Through our brands WellNest (AI-powered health ecosystem) and Zenova (next-gen smart wearables), we are pioneering a future where technology truly understands and enhances human health. Why Are We Calling You? We’re assembling a community of passionate minds—AI enthusiasts, developers, designers, innovators, and problem-solvers—who want to be part of something bigger. This is NOT an internship. This is NOT a job. This is a mission to build the future of health-tech. What’s in It for You? ✅ Work on groundbreaking AI & LLM projects that solve real-world healthcare problems ✅ Hands-on experience in AI, ML, IoT, and smart wearables ✅ Mentorship & learning opportunities from top AI leaders ✅ Exclusive perks like health, wellness, and gym packages ✅ Recognition & growth opportunities—top contributors will be given leadership roles as we scale ✅ Certificates & endorsements to showcase your contributions ✅ Opportunity to be part of a global AI-led revolution in healthcare & fitness ✅ Network with like-minded innovators, entrepreneurs, and industry pioneers ✅ Early access to WellNest & Zenova products and AI-driven health plans ✅ Possibility of paid roles & equity-based opportunities for the most dedicated members Who Should Join? Students & fresh graduates eager to apply their skills AI & tech enthusiasts passionate about real-world innovation Developers, designers, and creators who want to build something impactful Anyone who believes in the power of AI for good and wants to contribute This is More Than Just a Tech Project We’re building an AI-powered health revolution. If you want to be part of something that changes lives, breaks barriers, and creates real impact, this is your chance. Movements aren’t built by employees—they are led by believers. If you believe in the power of AI to transform health, join us and let’s build the future together!

Roast my resume + suggestions for my portfolio
reddit
LLM Vibe Score0
Human Vibe Score0
saasypThis week

Roast my resume + suggestions for my portfolio

Hi everyone, I'm a European (I don't know if it's important to mention) Master's AI student, and as many out there, I'm trying to break into the ML (Deep Learning more specifically) world and I am aware of the current market crisis we're going through. Therefore, I ask you to rate/roast my resume as much as you can, since I'm trying to land an internship alongside the studies. The only project I’ve worked on so far was a research project conducted as part of my university studies. Since it was primarily research-oriented, there weren’t significant opportunities to benchmark the results using standard metrics for comparison. (maybe you can suggest me how to leverage it on the resume - yes it says Jan 2024 because the deadline is in January but it’s done already) I am deeply passionate about ML/DL , and I understand the importance of having a strong portfolio to showcase my skills. However, I struggle with finding creative and impactful project ideas to put into practice. While I consider myself a beginner, my Master’s program has provided me with a solid foundation (including the maths/algebra/statistics from my bachelor) in ML and unluckily I haven't had the opportunity to land a decent internship where I could learn and apply such things yet. As far as I read from multiple posts here, I should start to work on solving a "problem" that I might face or something that I'm interested in, but as I said I've completely no idea, thus I'd highly appreciate your help also with this. Is contributing to open source project valid as well? Could you suggest any websites where I can find some? Thanks for your precious time and attention :)

I single-handedly built the world’s best AI investing platform. Here’s NexusTrade’s 2024 year in review
reddit
LLM Vibe Score0
Human Vibe Score1
No-Definition-2886This week

I single-handedly built the world’s best AI investing platform. Here’s NexusTrade’s 2024 year in review

I copy-pasted the content of this article to save you a click! I’ve been developing an AI investing platform for 4 years, and I’m blown away by all of the new features I’ve gotten done! Here’s my project’s 2024 year in review —- When someone asks me what is the best way to learn how to trade and invest, I have an unbiased answer – NexusTrade.io. I started NexusTrade to empower everybody, including beginners and non-technical investors, to learn how to make smarter investing decisions. NexusTrade is the best way for a new investor to learn algorithmic trading and financial research, and I’m not the only person to think so. Just this year alone, user growth has skyrocketed from 1,703 users to 14,319 users. This is driven by new features, better research tools, and the launch of algorithmic trading. Here’s NexusTrade’s 2024 year in review, a semi-complete list of the features I’ve launched. Summarizing this year in review TL;DR: I implemented a variety of new features to enhance NexusTrade’s algorithmic trading and financial research capabilities. This includes: Cryptocurrency support Enhanced financial research, like the AI-Powered Stock Screener Unique watchlists and daily market summaries Live-trading with Alpaca. Next year, I plan to implement features to make NexusTrade more tailored for each user’s experience, and launch several unique features including copy trading and fully automated algorithmic trading. Feature-by-feature: What have I done so far in 2024? Algorithmic Cryptocurrency Trading Picture: Algorithmic Cryptocurrency Trading I kicked off the year by adding cryptocurrency support to NexusTrade. Users can now research, design, and implement automated strategies for popular cryptocurrencies, such as Bitcoin, Dogecoin, and Ethereum. AI-Powered Stock Screener and research capabilities Picture: AI-Powered Stock Screener In tandem with cryptocurrency support, I made a huge update to Aurora, the AI Assistant in NexusTrade, by implementing a natural language stock screener. This screener makes it easy to find fundamentally strong stocks. Throughout the year, I’ve made several enhancements to it. Over time, I’ve made the screener faster, more accurate, and expanded its capabilities. Using fundamental indicators within trading strategies Picture: Using fundamental indicators Doing financial research for companies isn’t enough; we also need a way to integrate this type of research into trading strategies. Thus, I’ve expanded the NexusTrade indicators, and made it possible to create strategies using metrics like revenue, net income, free cash flow, and P/E ratio. Stock watchlists with tailored, automated daily emails Picture: Stock watchlists In addition, I didn’t want the research you may have done for a stock (or list of stocks) to be forgotten. Thus, I created the most useful watchlist page of any investing platform. This watchlist makes it easy to keep track of your favorite stocks, track them over time, and even receive curated, daily emails about them. Enhanced user profile page, Google sign-ins, and two-factor authentication Picture: Enhanced user profile Keeping in theme with adding new pages to NexusTrade, many pages, such as the profile page, got a huge revamp. The new profile page is cleaner, easier to use, and allows you to secure your account more effectively, for example, by using two-factor authentication. GPT-Reports: an AI-generated analysis of every stock in the market Picture: GPT-Reports I created GPT-Stock Reports, an AI-Generated analysis of every stock in the market. This report was generated by taking each company’s earnings data and asking GPT to analyze the stock and give it a rating. Manual and semi-automated algorithmic trading with Alpaca Picture: Manual and semi-automated trading Finally, I’ve fully launched the Alpaca integration, and enabled users to execute real trades directly in the NexusTrade app! This integration has transformed NexusTrade from a financial research app into a real, algorithmic trading platform for retail investors. Concluding Thoughts When I say that NexusTrade is the best platform for traders and investors to make more money in the stock market, you may naively think that I’m biased. I created the app, and the rose-tinted glasses is bound to make every red flag look like a regular flag, right? Wrong. NexusTrade is objectively a completely new way for investors to approach financial markets. The fact that the app is so expansive is nothing short of miraculous.

ChatGPT Full Course For 2025 | ChatGPT Tutorial For Beginnners | ChatGPT Course | Simplilearn
youtube
LLM Vibe Score0.369
Human Vibe Score0.26
SimplilearnMar 28, 2025

ChatGPT Full Course For 2025 | ChatGPT Tutorial For Beginnners | ChatGPT Course | Simplilearn

🔥Purdue - Applied Generative AI Specialization - https://www.simplilearn.com/applied-ai-course?utmcampaign=C4lBsBlloL0&utmmedium=Lives&utm_source=Youtube 🔥Professional Certificate Program in Generative AI and Machine Learning - IITG (India Only) - https://www.simplilearn.com/iitg-generative-ai-machine-learning-program?utmcampaign=C4lBsBlloL0&utmmedium=Lives&utm_source=Youtube 🔥Advanced Executive Program In Applied Generative AI - https://www.simplilearn.com/applied-generative-ai-course?utmcampaign=C4lBsBlloL0&utmmedium=Lives&utm_source=Youtube This ChatGPT Full Course 2025 by Simplilearn provides a comprehensive learning journey, starting with an introduction to ChatGPT and Generative AI, followed by insights into AI job opportunities and a comparison between ChatGPT 4.0 and 4.0 Turbo. The tutorial covers prompt engineering techniques, machine learning fundamentals, and running Llama models privately. Learners will explore ChatGPT-powered application development, its role in programming, and Excel automation. The course also dives into blogging, PowerPoint automation, customer support, and finance applications. Advanced topics like RAG vs. Prompt Tuning, prompt injection, and LangChain are included, along with discussions on OpenAI's latest innovations, including Sora and Strawberry. By the end, participants will gain a strong understanding of ChatGPT’s capabilities and monetization strategies. 🚀 Following are the topics covered in the ChatGPT Full Course 2025: 00:00:00 - Introduction to ChatGPT Full Course 2025 00:09:26 - What is ChatGPT 00:10:11 - What is Gen AI 00:26:29 - How to get Job in AI 00:27:06 - ChatGPT 40 vs ChatGPT 4 01:03:14 - Chatgpt analyse 02:13:57 - Prompt Engineering Tutorial 03:10:34 - What is Machine Learning 04:07:06 - Machine Learning Tutorial 04:08:13 - Run Lama Privately 04:23:50 - Search GPT 04:25:31 - Build App Using ChatGPT 06:31:11 - ChatGPT for Programming 06:46:08 - Prompt Formulae Chatgpt 07:58:38 - Automate Excel using Chatgpt 08:00:06 - Blogging with ChatGpt 08:27:25 - Powerpoint using Chatgpt 08:28:31 - Rag Vs Prompt Tuning 09:37:43 - Chatgpt for Customer Support 11:11:06 - ChatGPT for finance 11:17:38 - Prompt injection 11:18:38 - How to Earn Money using ChatGPT 11:41:46 - Open AI Strawberry 11:52:42 - Openai sora 11:54:57 - Langchain 12:22:19 - Open ai chatgpt o1 model ✅ Subscribe to our Channel to learn more about the top Technologies: https://bit.ly/2VT4WtH ⏩ Check out the Artificial Intelligence training videos: https://youtube.com/playlist?list=PLEiEAq2VkUULa5aOQmO_al2VVmhC-eqeI #gpt #chatgpt #chatgptforbeginners #chatgptcourse #genai #generativeai #artificialintelligence #ai #machinelearning #llm #simplilearn #2025 ➡️ About Professional Certificate Program in Generative AI and Machine Learning Dive into the future of AI with our Generative AI & Machine Learning course, in collaboration with E&ICT Academy, IIT Guwahati. Learn tools like ChatGPT, OpenAI, Hugging Face, Python, and more. Join masterclasses led by IITG faculty, engage in hands-on projects, and earn Executive Alumni Status. Key Features: ✅ Program completion certificate from E&ICT Academy, IIT Guwahati ✅ Curriculum delivered in live virtual classes by seasoned industry experts ✅ Exposure to the latest AI advancements, such as generative AI, LLMs, and prompt engineering ✅ Interactive live-virtual masterclasses delivered by esteemed IIT Guwahati faculty ✅ Opportunity to earn an 'Executive Alumni Status' from E&ICT Academy, IIT Guwahati ✅ Eligibility for a campus immersion program organized at IIT Guwahati ✅ Exclusive hackathons and “ask-me-anything” sessions by IBM ✅ Certificates for IBM courses and industry masterclasses by IBM experts ✅ Practical learning through 25+ hands-on projects and 3 industry-oriented capstone projects ✅ Access to a wide array of AI tools such as ChatGPT, Hugging Face, DALL-E 2, Midjourney and more ✅ Simplilearn's JobAssist helps you get noticed by top hiring companies Skills Covered: ✅ Generative AI ✅ Prompt Engineering ✅ Chatbot Development ✅ Supervised and Unsupervised Learning ✅ Model Training and Optimization ✅ Model Evaluation and Validation ✅ Ensemble Methods ✅ Deep Learning ✅ Natural Language Processing ✅ Computer Vision ✅ Reinforcement Learning ✅ Machine Learning Algorithms ✅ Speech Recognition ✅ Statistics Learning Path: ✅ Program Induction ✅ Programming Fundamentals ✅ Python for Data Science (IBM) ✅ Applied Data Science with Python ✅ Machine Learning ✅ Deep Learning with TensorFlow (IBM) ✅ Deep Learning Specialization ✅ Essentials of Generative AI, Prompt Engineering & ChatGPT ✅ Advanced Generative AI ✅ Capstone Electives: ✅ ADL & Computer Vision ✅ NLP and Speech Recognition ✅ Reinforcement Learning ✅ Academic Masterclass ✅ Industry Masterclass 👉 Learn More At: https://www.simplilearn.com/iitg-generative-ai-machine-learning-program?utmcampaign=C4lBsBlloL0&utmmedium=Lives&utm_source=Youtube

nine
github
LLM Vibe Score0.406
Human Vibe Score0.000678327714013925
NethermindEthMar 28, 2025

nine

NINE - Neural Interconnected Nodes Engine A flexible framework for building a distributed network of AI agents that work everywhere (STD, WASM, TEE) with a dynamic interface and hot-swappable components. One of the key concepts of the framework is a meta-layer that enables building software systems in a No-code style, where the entire integration is handled by the LLM. Documentation | Telegram | X | Discord Overview Project Structure The project is built using Rust (full-stack) and organized as a workspace consisting of two major groups: substance/ - The core components of the system, responsible for interaction. particles/ - Plugins for the system that enable additional functionalities. examples/ - Usage examples of the framework. Use cases The following cases will have a minimal implementation, and they will be used to track the progress of the framework and its flexibility in building such systems. ☑️ Chatbots - AI-driven natural language chatbots for customer support, virtual assistants, and automation. ☑️ AI-governed blockchains (ChaosChain) - Self-regulating and intelligent blockchain ecosystems with automated decision-making. ⬜ Personal AI Assistant with dynamic UI - AI that generates adaptive and context-aware user interfaces on demand. ☑️ AI-powered trading bots - Autonomous financial agents for high-frequency trading and portfolio management. ⬜ Intelligent email assistant - AI for reading, summarizing, filtering, and responding to emails autonomously. ⬜ Interactivity in home appliances - AI-powered automation for home appliances, making them responsive and adaptive. ⬜ On-demand observability and awareness in DevOps - AI-driven insights, predictive monitoring, and automated issue detection in IT systems. ⬜ AI-powered developer tools - AI agents assisting with code generation, debugging, and software optimization. ⬜ Autonomous research agent - Self-learning AI for data analysis, knowledge discovery, and hypothesis testing. Status: ⬜ Not started | ☑️ In Progress | ✅ Completed Interfaces The platform provides No-code interfaces that automatically adapt to your needs and use LLM for system management. ☑️ Stdio - A console interface that also allows interaction with models through the terminal or via scripts. ☑️ TUI - An advanced console interface with an informative dashboard and the ability to interact more comprehensively with the system. ☑️ GUI - A graphical immediate-state interface suitable for embedded systems with real-time information rendering. ⬜ WEB - The ability to interact with the system through a web browser, such as from a mobile phone. ⬜ Voice - An interface for people with disabilities or those who prefer interaction without a graphical representation (e.g., voice control). ⬜ API - On-the-fly API creation for your system, providing a formal interaction method. This includes encapsulating an entire mesh system into a simple tool for LLM. Features (goals) Built on Rust and implemented as hybrid actor-state machines. Supports various LLMs, tools, and extensibility. Hot model swapping without restarting. Real-time configuration adjustment. Distributed agents, the ability to run components on different machines. Provides a dynamic user interface (UI9) that is automatically generated for interacting with a network of agents. Usage An agent is a substance that assembles from components (particles). Connections automatically form between them, bringing the agent to life: License This project is licensed under the [MIT license]. [MIT license]: https://github.com/NethermindEth/nine/blob/trunk/LICENSE Contribution Unless you explicitly state otherwise, any contribution intentionally submitted for inclusion in this project by you, shall be licensed as MIT, without any additional terms or conditions.

n8n
github
LLM Vibe Score0.66
Human Vibe Score1
n8n-ioMar 28, 2025

n8n

!Banner image n8n - Secure Workflow Automation for Technical Teams n8n is a workflow automation platform that gives technical teams the flexibility of code with the speed of no-code. With 400+ integrations, native AI capabilities, and a fair-code license, n8n lets you build powerful automations while maintaining full control over your data and deployments. !n8n.io - Screenshot Key Capabilities Code When You Need It: Write JavaScript/Python, add npm packages, or use the visual interface AI-Native Platform: Build AI agent workflows based on LangChain with your own data and models Full Control: Self-host with our fair-code license or use our cloud offering Enterprise-Ready: Advanced permissions, SSO, and air-gapped deployments Active Community: 400+ integrations and 900+ ready-to-use templates Quick Start Try n8n instantly with npx (requires Node.js): Or deploy with Docker: Access the editor at http://localhost:5678 Resources 📚 Documentation 🔧 400+ Integrations 💡 Example Workflows 🤖 AI & LangChain Guide 👥 Community Forum 📖 Community Tutorials Support Need help? Our community forum is the place to get support and connect with other users: community.n8n.io License n8n is fair-code distributed under the Sustainable Use License and n8n Enterprise License. Source Available: Always visible source code Self-Hostable: Deploy anywhere Extensible: Add your own nodes and functionality Enterprise licenses available for additional features and support. Additional information about the license model can be found in the docs. Contributing Found a bug 🐛 or have a feature idea ✨? Check our Contributing Guide to get started. Join the Team Want to shape the future of automation? Check out our job posts and join our team! What does n8n mean? Short answer: It means "nodemation" and is pronounced as n-eight-n. Long answer: "I get that question quite often (more often than I expected) so I decided it is probably best to answer it here. While looking for a good name for the project with a free domain I realized very quickly that all the good ones I could think of were already taken. So, in the end, I chose nodemation. 'node-' in the sense that it uses a Node-View and that it uses Node.js and '-mation' for 'automation' which is what the project is supposed to help with. However, I did not like how long the name was and I could not imagine writing something that long every time in the CLI. That is when I then ended up on 'n8n'." - Jan Oberhauser, Founder and CEO, n8n.io

anything-llm
github
LLM Vibe Score0.572
Human Vibe Score0.4703504093656464
Mintplex-LabsMar 28, 2025

anything-llm

AnythingLLM: The all-in-one AI app you were looking for. Chat with your docs, use AI Agents, hyper-configurable, multi-user, & no frustrating set up required. | | Docs | Hosted Instance English · 简体中文 · 日本語 👉 AnythingLLM for desktop (Mac, Windows, & Linux)! Download Now A full-stack application that enables you to turn any document, resource, or piece of content into context that any LLM can use as references during chatting. This application allows you to pick and choose which LLM or Vector Database you want to use as well as supporting multi-user management and permissions. !Chatting Watch the demo! Product Overview AnythingLLM is a full-stack application where you can use commercial off-the-shelf LLMs or popular open source LLMs and vectorDB solutions to build a private ChatGPT with no compromises that you can run locally as well as host remotely and be able to chat intelligently with any documents you provide it. AnythingLLM divides your documents into objects called workspaces. A Workspace functions a lot like a thread, but with the addition of containerization of your documents. Workspaces can share documents, but they do not talk to each other so you can keep your context for each workspace clean. Cool features of AnythingLLM 🆕 Custom AI Agents 🆕 No-code AI Agent builder 🖼️ Multi-modal support (both closed and open-source LLMs!) 👤 Multi-user instance support and permissioning Docker version only 🦾 Agents inside your workspace (browse the web, etc) 💬 Custom Embeddable Chat widget for your website Docker version only 📖 Multiple document type support (PDF, TXT, DOCX, etc) Simple chat UI with Drag-n-Drop funcitonality and clear citations. 100% Cloud deployment ready. Works with all popular closed and open-source LLM providers. Built-in cost & time-saving measures for managing very large documents compared to any other chat UI. Full Developer API for custom integrations! Much more...install and find out! Supported LLMs, Embedder Models, Speech models, and Vector Databases Large Language Models (LLMs): Any open-source llama.cpp compatible model OpenAI OpenAI (Generic) Azure OpenAI AWS Bedrock Anthropic NVIDIA NIM (chat models) Google Gemini Pro Hugging Face (chat models) Ollama (chat models) LM Studio (all models) LocalAi (all models) Together AI (chat models) Fireworks AI (chat models) Perplexity (chat models) OpenRouter (chat models) DeepSeek (chat models) Mistral Groq Cohere KoboldCPP LiteLLM Text Generation Web UI Apipie xAI Novita AI (chat models) PPIO Embedder models: AnythingLLM Native Embedder (default) OpenAI Azure OpenAI LocalAi (all) Ollama (all) LM Studio (all) Cohere Audio Transcription models: AnythingLLM Built-in (default) OpenAI TTS (text-to-speech) support: Native Browser Built-in (default) PiperTTSLocal - runs in browser OpenAI TTS ElevenLabs Any OpenAI Compatible TTS service. STT (speech-to-text) support: Native Browser Built-in (default) Vector Databases: LanceDB (default) Astra DB Pinecone Chroma Weaviate Qdrant Milvus Zilliz Technical Overview This monorepo consists of three main sections: frontend: A viteJS + React frontend that you can run to easily create and manage all your content the LLM can use. server: A NodeJS express server to handle all the interactions and do all the vectorDB management and LLM interactions. collector: NodeJS express server that process and parses documents from the UI. docker: Docker instructions and build process + information for building from source. embed: Submodule for generation & creation of the web embed widget. browser-extension: Submodule for the chrome browser extension. 🛳 Self Hosting Mintplex Labs & the community maintain a number of deployment methods, scripts, and templates that you can use to run AnythingLLM locally. Refer to the table below to read how to deploy on your preferred environment or to automatically deploy. | Docker | AWS | GCP | Digital Ocean | Render.com | |----------------------------------------|----|-----|---------------|------------| | [![Deploy on Docker][docker-btn]][docker-deploy] | [![Deploy on AWS][aws-btn]][aws-deploy] | [![Deploy on GCP][gcp-btn]][gcp-deploy] | [![Deploy on DigitalOcean][do-btn]][do-deploy] | [![Deploy on Render.com][render-btn]][render-deploy] | | Railway | RepoCloud | Elestio | | --- | --- | --- | | [![Deploy on Railway][railway-btn]][railway-deploy] | [![Deploy on RepoCloud][repocloud-btn]][repocloud-deploy] | [![Deploy on Elestio][elestio-btn]][elestio-deploy] | or set up a production AnythingLLM instance without Docker → How to setup for development yarn setup To fill in the required .env files you'll need in each of the application sections (from root of repo). Go fill those out before proceeding. Ensure server/.env.development is filled or else things won't work right. yarn dev:server To boot the server locally (from root of repo). yarn dev:frontend To boot the frontend locally (from root of repo). yarn dev:collector To then run the document collector (from root of repo). Learn about documents Learn about vector caching External Apps & Integrations These are apps that are not maintained by Mintplex Labs, but are compatible with AnythingLLM. A listing here is not an endorsement. Midori AI Subsystem Manager - A streamlined and efficient way to deploy AI systems using Docker container technology. Coolify - Deploy AnythingLLM with a single click. GPTLocalhost for Microsoft Word - A local Word Add-in for you to use AnythingLLM in Microsoft Word. Telemetry & Privacy AnythingLLM by Mintplex Labs Inc contains a telemetry feature that collects anonymous usage information. More about Telemetry & Privacy for AnythingLLM Why? We use this information to help us understand how AnythingLLM is used, to help us prioritize work on new features and bug fixes, and to help us improve AnythingLLM's performance and stability. Opting out Set DISABLE_TELEMETRY in your server or docker .env settings to "true" to opt out of telemetry. You can also do this in-app by going to the sidebar > Privacy and disabling telemetry. What do you explicitly track? We will only track usage details that help us make product and roadmap decisions, specifically: Type of your installation (Docker or Desktop) When a document is added or removed. No information about the document. Just that the event occurred. This gives us an idea of use. Type of vector database in use. Let's us know which vector database provider is the most used to prioritize changes when updates arrive for that provider. Type of LLM in use. Let's us know the most popular choice and prioritize changes when updates arrive for that provider. Chat is sent. This is the most regular "event" and gives us an idea of the daily-activity of this project across all installations. Again, only the event is sent - we have no information on the nature or content of the chat itself. You can verify these claims by finding all locations Telemetry.sendTelemetry is called. Additionally these events are written to the output log so you can also see the specific data which was sent - if enabled. No IP or other identifying information is collected. The Telemetry provider is PostHog - an open-source telemetry collection service. View all telemetry events in source code 👋 Contributing create issue create PR with branch name format of - LGTM from core-team 🌟 Contributors 🔗 More Products [VectorAdmin][vector-admin]: An all-in-one GUI & tool-suite for managing vector databases. [OpenAI Assistant Swarm][assistant-swarm]: Turn your entire library of OpenAI assistants into one single army commanded from a single agent. [![][back-to-top]](#readme-top) Copyright © 2025 [Mintplex Labs][profile-link]. This project is MIT licensed. [back-to-top]: https://img.shields.io/badge/-BACKTOTOP-222628?style=flat-square [profile-link]: https://github.com/mintplex-labs [vector-admin]: https://github.com/mintplex-labs/vector-admin [assistant-swarm]: https://github.com/Mintplex-Labs/openai-assistant-swarm [docker-btn]: ./images/deployBtns/docker.png [docker-deploy]: ./docker/HOWTOUSE_DOCKER.md [aws-btn]: ./images/deployBtns/aws.png [aws-deploy]: ./cloud-deployments/aws/cloudformation/DEPLOY.md [gcp-btn]: https://deploy.cloud.run/button.svg [gcp-deploy]: ./cloud-deployments/gcp/deployment/DEPLOY.md [do-btn]: https://www.deploytodo.com/do-btn-blue.svg [do-deploy]: ./cloud-deployments/digitalocean/terraform/DEPLOY.md [render-btn]: https://render.com/images/deploy-to-render-button.svg [render-deploy]: https://render.com/deploy?repo=https://github.com/Mintplex-Labs/anything-llm&branch=render [render-btn]: https://render.com/images/deploy-to-render-button.svg [render-deploy]: https://render.com/deploy?repo=https://github.com/Mintplex-Labs/anything-llm&branch=render [railway-btn]: https://railway.app/button.svg [railway-deploy]: https://railway.app/template/HNSCS1?referralCode=WFgJkn [repocloud-btn]: https://d16t0pc4846x52.cloudfront.net/deploylobe.svg [repocloud-deploy]: https://repocloud.io/details/?app_id=276 [elestio-btn]: https://elest.io/images/logos/deploy-to-elestio-btn.png [elestio-deploy]: https://elest.io/open-source/anythingllm

GenAI_Agents
github
LLM Vibe Score0.563
Human Vibe Score0.24210481455988786
NirDiamantMar 28, 2025

GenAI_Agents

🌟 Support This Project: Your sponsorship fuels innovation in GenAI agent development. Become a sponsor to help maintain and expand this valuable resource! GenAI Agents: Comprehensive Repository for Development and Implementation 🚀 Welcome to one of the most extensive and dynamic collections of Generative AI (GenAI) agent tutorials and implementations available today. This repository serves as a comprehensive resource for learning, building, and sharing GenAI agents, ranging from simple conversational bots to complex, multi-agent systems. 📫 Stay Updated! 🚀Cutting-edgeUpdates 💡ExpertInsights 🎯Top 0.1%Content Join over 15,000 of AI enthusiasts getting unique cutting-edge insights and free tutorials! Plus, subscribers get exclusive early access and special 33% discounts to my book and the upcoming RAG Techniques course! Introduction Generative AI agents are at the forefront of artificial intelligence, revolutionizing the way we interact with and leverage AI technologies. This repository is designed to guide you through the development journey, from basic agent implementations to advanced, cutting-edge systems. 📚 Learn to Build Your First AI Agent Your First AI Agent: Simpler Than You Think This detailed blog post complements the repository by providing a complete A-Z walkthrough with in-depth explanations of core concepts, step-by-step implementation, and the theory behind AI agents. It's designed to be incredibly simple to follow while covering everything you need to know to build your first working agent from scratch. 💡 Plus: Subscribe to the newsletter for exclusive early access to tutorials and special discounts on upcoming courses and books! Our goal is to provide a valuable resource for everyone - from beginners taking their first steps in AI to seasoned practitioners pushing the boundaries of what's possible. By offering a range of examples from foundational to complex, we aim to facilitate learning, experimentation, and innovation in the rapidly evolving field of GenAI agents. Furthermore, this repository serves as a platform for showcasing innovative agent creations. Whether you've developed a novel agent architecture or found an innovative application for existing techniques, we encourage you to share your work with the community. Related Projects 📚 Dive into my comprehensive guide on RAG techniques to learn about integrating external knowledge into AI systems, enhancing their capabilities with up-to-date and relevant information retrieval. 🖋️ Explore my Prompt Engineering Techniques guide for an extensive collection of prompting strategies, from fundamental concepts to advanced methods, improving your ability to communicate effectively with AI language models. A Community-Driven Knowledge Hub This repository grows stronger with your contributions! Join our vibrant Discord community — the central hub for shaping and advancing this project together 🤝 GenAI Agents Discord Community Whether you're a novice eager to learn or an expert ready to share your knowledge, your insights can shape the future of GenAI agents. Join us to propose ideas, get feedback, and collaborate on innovative implementations. For contribution guidelines, please refer to our CONTRIBUTING.md file. Let's advance GenAI agent technology together! 🔗 For discussions on GenAI, agents, or to explore knowledge-sharing opportunities, feel free to connect on LinkedIn. Key Features 🎓 Learn to build GenAI agents from beginner to advanced levels 🧠 Explore a wide range of agent architectures and applications 📚 Step-by-step tutorials and comprehensive documentation 🛠️ Practical, ready-to-use agent implementations 🌟 Regular updates with the latest advancements in GenAI 🤝 Share your own agent creations with the community GenAI Agent Implementations Explore our extensive list of GenAI agent implementations, sorted by categories: 🌱 Beginner-Friendly Agents Simple Conversational Agent LangChain PydanticAI Overview 🔎 A context-aware conversational AI maintains information across interactions, enabling more natural dialogues. Implementation 🛠️ Integrates a language model, prompt template, and history manager to generate contextual responses and track conversation sessions. Simple Question Answering Agent Overview 🔎 Answering (QA) agent using LangChain and OpenAI's language model understands user queries and provides relevant, concise answers. Implementation 🛠️ Combines OpenAI's GPT model, a prompt template, and an LLMChain to process user questions and generate AI-driven responses in a streamlined manner. Simple Data Analysis Agent LangChain PydanticAI Overview 🔎 An AI-powered data analysis agent interprets and answers questions about datasets using natural language, combining language models with data manipulation tools for intuitive data exploration. Implementation 🛠️ Integrates a language model, data manipulation framework, and agent framework to process natural language queries and perform data analysis on a synthetic dataset, enabling accessible insights for non-technical users. 🔧 Framework Tutorial: LangGraph Introduction to LangGraph: Building Modular AI Workflows Overview 🔎 This tutorial introduces LangGraph, a powerful framework for creating modular, graph-based AI workflows. Learn how to leverage LangGraph to build more complex and flexible AI agents that can handle multi-step processes efficiently. Implementation 🛠️ Step-by-step guide on using LangGraph to create a StateGraph workflow. The tutorial covers key concepts such as state management, node creation, and graph compilation. It demonstrates these principles by constructing a simple text analysis pipeline, serving as a foundation for more advanced agent architectures. Additional Resources 📚 Blog Post 🎓 Educational and Research Agents ATLAS: Academic Task and Learning Agent System Overview 🔎 ATLAS demonstrates how to build an intelligent multi-agent system that transforms academic support through AI-powered assistance. The system leverages LangGraph's workflow framework to coordinate multiple specialized agents that provide personalized academic planning, note-taking, and advisory support. Implementation 🛠️ Implements a state-managed multi-agent architecture using four specialized agents (Coordinator, Planner, Notewriter, and Advisor) working in concert through LangGraph's workflow framework. The system features sophisticated workflows for profile analysis and academic support, with continuous adaptation based on student performance and feedback. Additional Resources 📚 YouTube Explanation Blog Post Scientific Paper Agent - Literature Review Overview 🔎 An intelligent research assistant that helps users navigate, understand, and analyze scientific literature through an orchestrated workflow. The system combines academic APIs with sophisticated paper processing techniques to automate literature review tasks, enabling researchers to efficiently extract insights from academic papers while maintaining research rigor and quality control. Implementation 🛠️ Leverages LangGraph to create a five-node workflow system including decision making, planning, tool execution, and quality validation nodes. The system integrates the CORE API for paper access, PDFplumber for document processing, and advanced language models for analysis. Key features include a retry mechanism for robust paper downloads, structured data handling through Pydantic models, and quality-focused improvement cycles with human-in-the-loop validation options. Additional Resources 📚 YouTube Explanation Blog Post Chiron - A Feynman-Enhanced Learning Agent Overview 🔎 An adaptive learning agent that guides users through educational content using a structured checkpoint system and Feynman-style teaching. The system processes learning materials (either user-provided or web-retrieved), verifies understanding through interactive checkpoints, and provides simplified explanations when needed, creating a personalized learning experience that mimics one-on-one tutoring. Implementation 🛠️ Uses LangGraph to orchestrate a learning workflow that includes checkpoint definition, context building, understanding verification, and Feynman teaching nodes. The system integrates web search for dynamic content retrieval, employs semantic chunking for context processing, and manages embeddings for relevant information retrieval. Key features include a 70% understanding threshold for progression, interactive human-in-the-loop validation, and structured output through Pydantic models for consistent data handling. Additional Resources 📚 YouTube Explanation 💼 Business and Professional Agents Customer Support Agent (LangGraph) Overview 🔎 An intelligent customer support agent using LangGraph categorizes queries, analyzes sentiment, and provides appropriate responses or escalates issues. Implementation 🛠️ Utilizes LangGraph to create a workflow combining state management, query categorization, sentiment analysis, and response generation. Essay Grading Agent (LangGraph) Overview 🔎 An automated essay grading system using LangGraph and an LLM model evaluates essays based on relevance, grammar, structure, and depth of analysis. Implementation 🛠️ Utilizes a state graph to define the grading workflow, incorporating separate grading functions for each criterion. Travel Planning Agent (LangGraph) Overview 🔎 A Travel Planner using LangGraph demonstrates how to build a stateful, multi-step conversational AI application that collects user input and generates personalized travel itineraries. Implementation 🛠️ Utilizes StateGraph to define the application flow, incorporates custom PlannerState for process management. GenAI Career Assistant Agent Overview 🔎 The GenAI Career Assistant demonstrates how to create a multi-agent system that provides personalized guidance for careers in Generative AI. Using LangGraph and Gemini LLM, the system delivers customized learning paths, resume assistance, interview preparation, and job search support. Implementation 🛠️ Leverages a multi-agent architecture using LangGraph to coordinate specialized agents (Learning, Resume, Interview, Job Search) through TypedDict-based state management. The system employs sophisticated query categorization and routing while integrating with external tools like DuckDuckGo for job searches and dynamic content generation. Additional Resources 📚 YouTube Explanation Project Manager Assistant Agent Overview 🔎 An AI agent designed to assist in project management tasks by automating the process of creating actionable tasks from project descriptions, identifying dependencies, scheduling work, and assigning tasks to team members based on expertise. The system includes risk assessment and self-reflection capabilities to optimize project plans through multiple iterations, aiming to minimize overall project risk. Implementation 🛠️ Leverages LangGraph to orchestrate a workflow of specialized nodes including task generation, dependency mapping, scheduling, allocation, and risk assessment. Each node uses GPT-4o-mini for structured outputs following Pydantic models. The system implements a feedback loop for self-improvement, where risk scores trigger reflection cycles that generate insights to optimize the project plan. Visualization tools display Gantt charts of the generated schedules across iterations. Additional Resources 📚 YouTube Explanation Contract Analysis Assistant (ClauseAI) Overview 🔎 ClauseAI demonstrates how to build an AI-powered contract analysis system using a multi-agent approach. The system employs specialized AI agents for different aspects of contract review, from clause analysis to compliance checking, and leverages LangGraph for workflow orchestration and Pinecone for efficient clause retrieval and comparison. Implementation 🛠️ Implements a sophisticated state-based workflow using LangGraph to coordinate multiple AI agents through contract analysis stages. The system features Pydantic models for data validation, vector storage with Pinecone for clause comparison, and LLM-based analysis for generating comprehensive contract reports. The implementation includes parallel processing capabilities and customizable report generation based on user requirements. Additional Resources 📚 YouTube Explanation E2E Testing Agent Overview 🔎 The E2E Testing Agent demonstrates how to build an AI-powered system that converts natural language test instructions into executable end-to-end web tests. Using LangGraph for workflow orchestration and Playwright for browser automation, the system enables users to specify test cases in plain English while handling the complexity of test generation and execution. Implementation 🛠️ Implements a structured workflow using LangGraph to coordinate test generation, validation, and execution. The system features TypedDict state management, integration with Playwright for browser automation, and LLM-based code generation for converting natural language instructions into executable test scripts. The implementation includes DOM state analysis, error handling, and comprehensive test reporting. Additional Resources 📚 YouTube Explanation 🎨 Creative and Content Generation Agents GIF Animation Generator Agent (LangGraph) Overview 🔎 A GIF animation generator that integrates LangGraph for workflow management, GPT-4 for text generation, and DALL-E for image creation, producing custom animations from user prompts. Implementation 🛠️ Utilizes LangGraph to orchestrate a workflow that generates character descriptions, plots, and image prompts using GPT-4, creates images with DALL-E 3, and assembles them into GIFs using PIL. Employs asynchronous programming for efficient parallel processing. TTS Poem Generator Agent (LangGraph) Overview 🔎 An advanced text-to-speech (TTS) agent using LangGraph and OpenAI's APIs classifies input text, processes it based on content type, and generates corresponding speech output. Implementation 🛠️ Utilizes LangGraph to orchestrate a workflow that classifies input text using GPT models, applies content-specific processing, and converts the processed text to speech using OpenAI's TTS API. The system adapts its output based on the identified content type (general, poem, news, or joke). Music Compositor Agent (LangGraph) Overview 🔎 An AI Music Compositor using LangGraph and OpenAI's language models generates custom musical compositions based on user input. The system processes the input through specialized components, each contributing to the final musical piece, which is then converted to a playable MIDI file. Implementation 🛠️ LangGraph orchestrates a workflow that transforms user input into a musical composition, using ChatOpenAI (GPT-4) to generate melody, harmony, and rhythm, which are then style-adapted. The final AI-generated composition is converted to a MIDI file using music21 and can be played back using pygame. Content Intelligence: Multi-Platform Content Generation Agent Overview 🔎 Content Intelligence demonstrates how to build an advanced content generation system that transforms input text into platform-optimized content across multiple social media channels. The system employs LangGraph for workflow orchestration to analyze content, conduct research, and generate tailored content while maintaining brand consistency across different platforms. Implementation 🛠️ Implements a sophisticated workflow using LangGraph to coordinate multiple specialized nodes (Summary, Research, Platform-Specific) through the content generation process. The system features TypedDict and Pydantic models for state management, integration with Tavily Search for research enhancement, and platform-specific content generation using GPT-4. The implementation includes parallel processing for multiple platforms and customizable content templates. Additional Resources 📚 YouTube Explanation Business Meme Generator Using LangGraph and Memegen.link Overview 🔎 The Business Meme Generator demonstrates how to create an AI-powered system that generates contextually relevant memes based on company website analysis. Using LangGraph for workflow orchestration, the system combines Groq's Llama model for text analysis and the Memegen.link API to automatically produce brand-aligned memes for digital marketing. Implementation 🛠️ Implements a state-managed workflow using LangGraph to coordinate website content analysis, meme concept generation, and image creation. The system features Pydantic models for data validation, asynchronous processing with aiohttp, and integration with external APIs (Groq, Memegen.link) to create a complete meme generation pipeline with customizable templates. Additional Resources 📚 YouTube Explanation Murder Mystery Game with LLM Agents Overview 🔎 A text-based detective game that utilizes autonomous LLM agents as interactive characters in a procedurally generated murder mystery. Drawing inspiration from the UNBOUNDED paper, the system creates unique scenarios each time, with players taking on the role of Sherlock Holmes to solve the case through character interviews and deductive reasoning. Implementation 🛠️ Leverages two LangGraph workflows - a main game loop for story/character generation and game progression, and a conversation sub-graph for character interactions. The system uses a combination of LLM-powered narrative generation, character AI, and structured game mechanics to create an immersive investigative experience with replayable storylines. Additional Resources 📚 YouTube Explanation 📊 Analysis and Information Processing Agents Memory-Enhanced Conversational Agent Overview 🔎 A memory-enhanced conversational AI agent incorporates short-term and long-term memory systems to maintain context within conversations and across multiple sessions, improving interaction quality and personalization. Implementation 🛠️ Integrates a language model with separate short-term and long-term memory stores, utilizes a prompt template incorporating both memory types, and employs a memory manager for storage and retrieval. The system includes an interaction loop that updates and utilizes memories for each response. Multi-Agent Collaboration System Overview 🔎 A multi-agent collaboration system combining historical research with data analysis, leveraging large language models to simulate specialized agents working together to answer complex historical questions. Implementation 🛠️ Utilizes a base Agent class to create specialized HistoryResearchAgent and DataAnalysisAgent, orchestrated by a HistoryDataCollaborationSystem. The system follows a five-step process: historical context provision, data needs identification, historical data provision, data analysis, and final synthesis. Self-Improving Agent Overview 🔎 A Self-Improving Agent using LangChain engages in conversations, learns from interactions, and continuously improves its performance over time through reflection and adaptation. Implementation 🛠️ Integrates a language model with chat history management, response generation, and a reflection mechanism. The system employs a learning system that incorporates insights from reflection to enhance future performance, creating a continuous improvement loop. Task-Oriented Agent Overview 🔎 A language model application using LangChain that summarizes text and translates the summary to Spanish, combining custom functions, structured tools, and an agent for efficient text processing. Implementation 🛠️ Utilizes custom functions for summarization and translation, wrapped as structured tools. Employs a prompt template to guide the agent, which orchestrates the use of tools. An agent executor manages the process, taking input text and producing both an English summary and its Spanish translation. Internet Search and Summarize Agent Overview 🔎 An intelligent web research assistant that combines web search capabilities with AI-powered summarization, automating the process of gathering information from the internet and distilling it into concise, relevant summaries. Implementation 🛠️ Integrates a web search module using DuckDuckGo's API, a result parser, and a text summarization engine leveraging OpenAI's language models. The system performs site-specific or general searches, extracts relevant content, generates concise summaries, and compiles attributed results for efficient information retrieval and synthesis. Multi agent research team - Autogen Overview 🔎 This technique explores a multi-agent system for collaborative research using the AutoGen library. It employs agents to solve tasks collaboratively, focusing on efficient execution and quality assurance. The system enhances research by distributing tasks among specialized agents. Implementation 🛠️ Agents are configured with specific roles using the GPT-4 model, including admin, developer, planner, executor, and quality assurance. Interaction management ensures orderly communication with defined transitions. Task execution involves collaborative planning, coding, execution, and quality checking, demonstrating a scalable framework for various domains. Additional Resources 📚 comprehensive solution with UI Blogpost Sales Call Analyzer Overview 🔎 An intelligent system that automates the analysis of sales call recordings by combining audio transcription with advanced natural language processing. The analyzer transcribes audio using OpenAI's Whisper, processes the text using NLP techniques, and generates comprehensive reports including sentiment analysis, key phrases, pain points, and actionable recommendations to improve sales performance. Implementation 🛠️ Utilizes multiple components in a structured workflow: OpenAI Whisper for audio transcription, CrewAI for task automation and agent management, and LangChain for orchestrating the analysis pipeline. The system processes audio through a series of steps from transcription to detailed analysis, leveraging custom agents and tasks to generate structured JSON reports containing insights about customer sentiment, sales opportunities, and recommended improvements. Additional Resources 📚 YouTube Explanation Weather Emergency & Response System Overview 🔎 A comprehensive system demonstrating two agent graph implementations for weather emergency response: a real-time graph processing live weather data, and a hybrid graph combining real and simulated data for testing high-severity scenarios. The system handles complete workflow from data gathering through emergency plan generation, with automated notifications and human verification steps. Implementation 🛠️ Utilizes LangGraph for orchestrating complex workflows with state management, integrating OpenWeatherMap API for real-time data, and Gemini for analysis and response generation. The system incorporates email notifications, social media monitoring simulation, and severity-based routing with configurable human verification for low/medium severity events. Additional Resources 📚 YouTube Explanation Self-Healing Codebase System Overview 🔎 An intelligent system that automatically detects, diagnoses, and fixes runtime code errors using LangGraph workflow orchestration and ChromaDB vector storage. The system maintains a memory of encountered bugs and their fixes through vector embeddings, enabling pattern recognition for similar errors across the codebase. Implementation 🛠️ Utilizes a state-based graph workflow that processes function definitions and runtime arguments through specialized nodes for error detection, code analysis, and fix generation. Incorporates ChromaDB for vector-based storage of bug patterns and fixes, with automated search and retrieval capabilities for similar error patterns, while maintaining code execution safety through structured validation steps. Additional Resources 📚 YouTube Explanation DataScribe: AI-Powered Schema Explorer Overview 🔎 An intelligent agent system that enables intuitive exploration and querying of relational databases through natural language interactions. The system utilizes a fleet of specialized agents, coordinated by a stateful Supervisor, to handle schema discovery, query planning, and data analysis tasks while maintaining contextual understanding through vector-based relationship graphs. Implementation 🛠️ Leverages LangGraph for orchestrating a multi-agent workflow including discovery, inference, and planning agents, with NetworkX for relationship graph visualization and management. The system incorporates dynamic state management through TypedDict classes, maintains database context between sessions using a db_graph attribute, and includes safety measures to prevent unauthorized database modifications. Memory-Enhanced Email Agent (LangGraph & LangMem) Overview 🔎 An intelligent email assistant that combines three types of memory (semantic, episodic, and procedural) to create a system that improves over time. The agent can triage incoming emails, draft contextually appropriate responses using stored knowledge, and enhance its performance based on user feedback. Implementation 🛠️ Leverages LangGraph for workflow orchestration and LangMem for sophisticated memory management across multiple memory types. The system implements a triage workflow with memory-enhanced decision making, specialized tools for email composition and calendar management, and a self-improvement mechanism that updates its own prompts based on feedback and past performance. Additional Resources 📚 Blog Post 📰 News and Information Agents News TL;DR using LangGraph Overview 🔎 A news summarization system that generates concise TL;DR summaries of current events based on user queries. The system leverages large language models for decision making and summarization while integrating with news APIs to access up-to-date content, allowing users to quickly catch up on topics of interest through generated bullet-point summaries. Implementation 🛠️ Utilizes LangGraph to orchestrate a workflow combining multiple components: GPT-4o-mini for generating search terms and article summaries, NewsAPI for retrieving article metadata, BeautifulSoup for web scraping article content, and Asyncio for concurrent processing. The system follows a structured pipeline from query processing through article selection and summarization, managing the flow between components to produce relevant TL;DRs of current news articles. Additional Resources 📚 YouTube Explanation Blog Post AInsight: AI/ML Weekly News Reporter Overview 🔎 AInsight demonstrates how to build an intelligent news aggregation and summarization system using a multi-agent architecture. The system employs three specialized agents (NewsSearcher, Summarizer, Publisher) to automatically collect, process and summarize AI/ML news for general audiences through LangGraph-based workflow orchestration. Implementation 🛠️ Implements a state-managed multi-agent system using LangGraph to coordinate the news collection (Tavily API), technical content summarization (GPT-4), and report generation processes. The system features modular architecture with TypedDict-based state management, external API integration, and markdown report generation with customizable templates. Additional Resources 📚 YouTube Explanation Journalism-Focused AI Assistant Overview 🔎 A specialized AI assistant that helps journalists tackle modern journalistic challenges like misinformation, bias, and information overload. The system integrates fact-checking, tone analysis, summarization, and grammar review tools to enhance the accuracy and efficiency of journalistic work while maintaining ethical reporting standards. Implementation 🛠️ Leverages LangGraph to orchestrate a workflow of specialized components including language models for analysis and generation, web search integration via DuckDuckGo's API, document parsing tools like PyMuPDFLoader and WebBaseLoader, text splitting with RecursiveCharacterTextSplitter, and structured JSON outputs. Each component works together through a unified workflow to analyze content, verify facts, detect bias, extract quotes, and generate comprehensive reports. Blog Writer (Open AI Swarm) Overview 🔎 A multi-agent system for collaborative blog post creation using OpenAI's Swarm package. It leverages specialized agents to perform research, planning, writing, and editing tasks efficiently. Implementation 🛠️ Utilizes OpenAI's Swarm Package to manage agent interactions. Includes an admin, researcher, planner, writer, and editor, each with specific roles. The system follows a structured workflow: topic setting, outlining, research, drafting, and editing. This approach enhances content creation through task distribution, specialization, and collaborative problem-solving. Additional Resources 📚 Swarm Repo Podcast Internet Search and Generate Agent 🎙️ Overview 🔎 A two step agent that first searches the internet for a given topic and then generates a podcast on the topic found. The search step uses a search agent and search function to find the most relevant information. The second step uses a podcast generation agent and generation function to create a podcast on the topic found. Implementation 🛠️ Utilizes LangGraph to orchestrate a two-step workflow. The first step involves a search agent and function to gather information from the internet. The second step uses a podcast generation agent and function to create a podcast based on the gathered information. 🛍️ Shopping and Product Analysis Agents ShopGenie - Redefining Online Shopping Customer Experience Overview 🔎 An AI-powered shopping assistant that helps customers make informed purchasing decisions even without domain expertise. The system analyzes product information from multiple sources, compares specifications and reviews, identifies the best option based on user needs, and delivers recommendations through email with supporting video reviews, creating a comprehensive shopping experience. Implementation 🛠️ Uses LangGraph to orchestrate a workflow combining Tavily for web search, Llama-3.1-70B for structured data analysis and product comparison, and YouTube API for review video retrieval. The system processes search results through multiple nodes including schema mapping, product comparison, review identification, and email generation. Key features include structured Pydantic models for consistent data handling, retry mechanisms for robust API interactions, and email delivery through SMTP for sharing recommendations. Additional Resources 📚 YouTube Explanation Car Buyer AI Agent Overview 🔎 The Smart Product Buyer AI Agent demonstrates how to build an intelligent system that assists users in making informed purchasing decisions. Using LangGraph and LLM-based intelligence, the system processes user requirements, scrapes product listings from websites like AutoTrader, and provides detailed analysis and recommendations for car purchases. Implementation 🛠️ Implements a state-based workflow using LangGraph to coordinate user interaction, web scraping, and decision support. The system features TypedDict state management, async web scraping with Playwright, and integrates with external APIs for comprehensive product analysis. The implementation includes a Gradio interface for real-time chat interaction and modular scraper architecture for easy extension to additional product categories. Additional Resources 📚 YouTube Explanation 🎯 Task Management and Productivity Agents Taskifier - Intelligent Task Allocation & Management Overview 🔎 An intelligent task management system that analyzes user work styles and creates personalized task breakdown strategies, born from the observation that procrastination often stems from task ambiguity among students and early-career professionals. The system evaluates historical work patterns, gathers relevant task information through web search, and generates customized step-by-step approaches to optimize productivity and reduce workflow paralysis. Implementation 🛠️ Leverages LangGraph for orchestrating a multi-step workflow including work style analysis, information gathering via Tavily API, and customized plan generation. The system maintains state through the process, integrating historical work pattern data with fresh task research to output detailed, personalized task execution plans aligned with the user's natural working style. Additional Resources 📚 YouTube Explanation Grocery Management Agents System Overview 🔎 A multi-agent system built with CrewAI that automates grocery management tasks including receipt interpretation, expiration date tracking, inventory management, and recipe recommendations. The system uses specialized agents to extract data from receipts, estimate product shelf life, track consumption, and suggest recipes to minimize food waste. Implementation 🛠️ Implements four specialized agents using CrewAI - a Receipt Interpreter that extracts item details from receipts, an Expiration Date Estimator that determines shelf life using online sources, a Grocery Tracker that maintains inventory based on consumption, and a Recipe Recommender that suggests meals using available ingredients. Each agent has specific tools and tasks orchestrated through a crew workflow. Additional Resources 📚 YouTube Explanation 🔍 Quality Assurance and Testing Agents LangGraph-Based Systems Inspector Overview 🔎 A comprehensive testing and validation tool for LangGraph-based applications that automatically analyzes system architecture, generates test cases, and identifies potential vulnerabilities through multi-agent inspection. The inspector employs specialized AI testers to evaluate different aspects of the system, from basic functionality to security concerns and edge cases. Implementation 🛠️ Integrates LangGraph for workflow orchestration, multiple LLM-powered testing agents, and a structured evaluation pipeline that includes static analysis, test case generation, and results verification. The system uses Pydantic for data validation, NetworkX for graph representation, and implements a modular architecture that allows for parallel test execution and comprehensive result analysis. Additional Resources 📚 YouTube Explanation Blog Post EU Green Deal FAQ Bot Overview 🔎 The EU Green Deal FAQ Bot demonstrates how to build a RAG-based AI agent that helps businesses understand EU green deal policies. The system processes complex regulatory documents into manageable chunks and provides instant, accurate answers to common questions about environmental compliance, emissions reporting, and waste management requirements. Implementation 🛠️ Implements a sophisticated RAG pipeline using FAISS vectorstore for document storage, semantic chunking for preprocessing, and multiple specialized agents (Retriever, Summarizer, Evaluator) for query processing. The system features query rephrasing for improved accuracy, cross-reference with gold Q&A datasets for answer validation, and comprehensive evaluation metrics to ensure response quality and relevance. Additional Resources 📚 YouTube Explanation Systematic Review Automation System + Paper Draft Creation Overview 🔎 A comprehensive system for automating academic systematic reviews using a directed graph architecture and LangChain components. The system generates complete, publication-ready systematic review papers, automatically processing everything from literature search through final draft generation with multiple revision cycles. Implementation 🛠️ Utilizes a state-based graph workflow that handles paper search and selection (up to 3 papers), PDF processing, and generates a complete academic paper with all standard sections (abstract, introduction, methods, results, conclusions, references). The system incorporates multiple revision cycles with automated critique and improvement phases, all orchestrated through LangGraph state management. Additional Resources 📚 YouTube Explanation 🌟 Special Advanced Technique 🌟 Sophisticated Controllable Agent for Complex RAG Tasks 🤖 Overview 🔎 An advanced RAG solution designed to tackle complex questions that simple semantic similarity-based retrieval cannot solve. This approach uses a sophisticated deterministic graph as the "brain" 🧠 of a highly controllable autonomous agent, capable of answering non-trivial questions from your own data. Implementation 🛠️ • Implement a multi-step process involving question anonymization, high-level planning, task breakdown, adaptive information retrieval and question answering, continuous re-planning, and rigorous answer verification to ensure grounded and accurate responses. Getting Started To begin exploring and building GenAI agents: Clone this repository: Navigate to the technique you're interested in: Follow the detailed implementation guide in each technique's notebook. Contributing We welcome contributions from the community! If you have a new technique or improvement to suggest: Fork the repository Create your feature branch: git checkout -b feature/AmazingFeature Commit your changes: git commit -m 'Add some AmazingFeature' Push to the branch: git push origin feature/AmazingFeature Open a pull request Contributors License This project is licensed under a custom non-commercial license - see the LICENSE file for details. ⭐️ If you find this repository helpful, please consider giving it a star! Keywords: GenAI, Generative AI, Agents, NLP, AI, Machine Learning, Natural Language Processing, LLM, Conversational AI, Task-Oriented AI

AITreasureBox
github
LLM Vibe Score0.447
Human Vibe Score0.1014145151561518
superiorluMar 28, 2025

AITreasureBox

AI TreasureBox English | 中文 Collect practical AI repos, tools, websites, papers and tutorials on AI. Translated from ChatGPT, picture from Midjourney. Catalog Repos Tools Websites Report&Paper Tutorials Repos updated repos and stars every 2 hours and re-ranking automatically. | No. | Repos | Description | | ----:|:-----------------------------------------|:------------------------------------------------------------------------------------------------------| | 1|🔥codecrafters-io/build-your-own-x !2025-03-28364681428|Master programming by recreating your favorite technologies from scratch.| | 2|sindresorhus/awesome !2025-03-28353614145|😎 Awesome lists about all kinds of interesting topics| | 3|public-apis/public-apis !2025-03-28334299125|A collective list of free APIs| | 4|kamranahmedse/developer-roadmap !2025-03-2831269540|Interactive roadmaps, guides and other educational content to help developers grow in their careers.| | 5|vinta/awesome-python !2025-03-28238581114|A curated list of awesome Python frameworks, libraries, software and resources| | 6|practical-tutorials/project-based-learning !2025-03-28222661124|Curated list of project-based tutorials| | 7|tensorflow/tensorflow !2025-03-281888714|An Open Source Machine Learning Framework for Everyone| | 8|Significant-Gravitas/AutoGPT !2025-03-2817391338|An experimental open-source attempt to make GPT-4 fully autonomous.| | 9|jackfrued/Python-100-Days !2025-03-2816305141|Python - 100天从新手到大师| | 10|AUTOMATIC1111/stable-diffusion-webui !2025-03-2815011553|Stable Diffusion web UI| | 11|huggingface/transformers !2025-03-2814207850|🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.| | 12|ollama/ollama !2025-03-28135166151|Get up and running with Llama 2, Mistral, Gemma, and other large language models.| | 13|f/awesome-chatgpt-prompts !2025-03-2812212738 |This repo includes ChatGPT prompt curation to use ChatGPT better.| | 14|justjavac/free-programming-books-zhCN !2025-03-2811316119|📚 免费的计算机编程类中文书籍,欢迎投稿| | 15|krahets/hello-algo !2025-03-2811107930|《Hello 算法》:动画图解、一键运行的数据结构与算法教程。支持 Python, Java, C++, C, C#, JS, Go, Swift, Rust, Ruby, Kotlin, TS, Dart 代码。简体版和繁体版同步更新,English version ongoing| | 16|yt-dlp/yt-dlp !2025-03-28105801114|A feature-rich command-line audio/video downloader| | 17|langchain-ai/langchain !2025-03-2810449479|⚡ Building applications with LLMs through composability ⚡| | 18|goldbergyoni/nodebestpractices !2025-03-281021629|✅ The Node.js best practices list (July 2024)| | 19|puppeteer/puppeteer !2025-03-289018212|JavaScript API for Chrome and Firefox| | 20|pytorch/pytorch !2025-03-288833938|Tensors and Dynamic neural networks in Python with strong GPU acceleration| | 21|neovim/neovim !2025-03-288781482|Vim-fork focused on extensibility and usability| | 22|🔥🔥langgenius/dify !2025-03-2887342639 |One API for plugins and datasets, one interface for prompt engineering and visual operation, all for creating powerful AI applications.| | 23|mtdvio/every-programmer-should-know !2025-03-28867069|A collection of (mostly) technical things every software developer should know about| | 24|open-webui/open-webui !2025-03-2886025159|User-friendly WebUI for LLMs (Formerly Ollama WebUI)| | 25|ChatGPTNextWeb/NextChat !2025-03-288231521|✨ Light and Fast AI Assistant. Support: Web | | 26|supabase/supabase !2025-03-287990956|The open source Firebase alternative.| | 27|openai/whisper !2025-03-287905542|Robust Speech Recognition via Large-Scale Weak Supervision| | 28|home-assistant/core !2025-03-287773219|🏡 Open source home automation that puts local control and privacy first.| | 29|tensorflow/models !2025-03-28774694|Models and examples built with TensorFlow| | 30| ggerganov/llama.cpp !2025-03-287731836 | Port of Facebook's LLaMA model in C/C++ | | 31|3b1b/manim !2025-03-287641918|Animation engine for explanatory math videos| | 32|microsoft/generative-ai-for-beginners !2025-03-287623860|12 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/| | 33|nomic-ai/gpt4all !2025-03-28729285 |gpt4all: an ecosystem of open-source chatbots trained on a massive collection of clean assistant data including code, stories and dialogue| | 34|comfyanonymous/ComfyUI !2025-03-2872635111|The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.| | 35|bregman-arie/devops-exercises !2025-03-2872225209|Linux, Jenkins, AWS, SRE, Prometheus, Docker, Python, Ansible, Git, Kubernetes, Terraform, OpenStack, SQL, NoSQL, Azure, GCP, DNS, Elastic, Network, Virtualization. DevOps Interview Questions| | 36|elastic/elasticsearch !2025-03-28721419|Free and Open, Distributed, RESTful Search Engine| | 37|🔥n8n-io/n8n !2025-03-2872093495|Free and source-available fair-code licensed workflow automation tool. Easily automate tasks across different services.| | 38|fighting41love/funNLP !2025-03-287200422|The Most Powerful NLP-Weapon Arsenal| | 39|hoppscotch/hoppscotch !2025-03-287060134|Open source API development ecosystem - https://hoppscotch.io (open-source alternative to Postman, Insomnia)| | 40|abi/screenshot-to-code !2025-03-286932817|Drop in a screenshot and convert it to clean HTML/Tailwind/JS code| | 41|binary-husky/gptacademic !2025-03-28680374|Academic Optimization of GPT| | 42|d2l-ai/d2l-zh !2025-03-286774142|Targeting Chinese readers, functional and open for discussion. The Chinese and English versions are used for teaching in over 400 universities across more than 60 countries| | 43|josephmisiti/awesome-machine-learning !2025-03-286739215|A curated list of awesome Machine Learning frameworks, libraries and software.| | 44|grafana/grafana !2025-03-286725414|The open and composable observability and data visualization platform. Visualize metrics, logs, and traces from multiple sources like Prometheus, Loki, Elasticsearch, InfluxDB, Postgres and many more.| | 45|python/cpython !2025-03-286602218|The Python programming language| | 46|apache/superset !2025-03-286519020|Apache Superset is a Data Visualization and Data Exploration Platform| | 47|xtekky/gpt4free !2025-03-28639391 |decentralizing the Ai Industry, free gpt-4/3.5 scripts through several reverse engineered API's ( poe.com, phind.com, chat.openai.com etc...)| | 48|sherlock-project/sherlock !2025-03-286332536|Hunt down social media accounts by username across social networks| | 49|twitter/the-algorithm !2025-03-28630586 |Source code for Twitter's Recommendation Algorithm| | 50|keras-team/keras !2025-03-28627835|Deep Learning for humans| | 51|openai/openai-cookbook !2025-03-28625136 |Examples and guides for using the OpenAI API| | 52|immich-app/immich !2025-03-286238670|High performance self-hosted photo and video management solution.| | 53|AppFlowy-IO/AppFlowy !2025-03-286173528|Bring projects, wikis, and teams together with AI. AppFlowy is an AI collaborative workspace where you achieve more without losing control of your data. The best open source alternative to Notion.| | 54|scikit-learn/scikit-learn !2025-03-286158212|scikit-learn: machine learning in Python| | 55|binhnguyennus/awesome-scalability !2025-03-286117021|The Patterns of Scalable, Reliable, and Performant Large-Scale Systems| | 56|labmlai/annotateddeeplearningpaperimplementations !2025-03-285951726|🧑‍🏫 59 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠| | 57|OpenInterpreter/open-interpreter !2025-03-285894710|A natural language interface for computers| | 58|lobehub/lobe-chat !2025-03-285832054|🤖 Lobe Chat - an open-source, extensible (Function Calling), high-performance chatbot framework. It supports one-click free deployment of your private ChatGPT/LLM web application.| | 59|meta-llama/llama !2025-03-28579536|Inference code for Llama models| | 60|nuxt/nuxt !2025-03-28566437|The Intuitive Vue Framework.| | 61|imartinez/privateGPT !2025-03-28555192|Interact with your documents using the power of GPT, 100% privately, no data leaks| | 62|Stirling-Tools/Stirling-PDF !2025-03-285500846|#1 Locally hosted web application that allows you to perform various operations on PDF files| | 63|PlexPt/awesome-chatgpt-prompts-zh !2025-03-285459720|ChatGPT Chinese Training Guide. Guidelines for various scenarios. Learn how to make it listen to you| | 64|dair-ai/Prompt-Engineering-Guide !2025-03-285451025 |🐙 Guides, papers, lecture, notebooks and resources for prompt engineering| | 65|ageitgey/facerecognition !2025-03-28544382|The world's simplest facial recognition api for Python and the command line| | 66|CorentinJ/Real-Time-Voice-Cloning !2025-03-285384814|Clone a voice in 5 seconds to generate arbitrary speech in real-time| | 67|geekan/MetaGPT !2025-03-285375376|The Multi-Agent Meta Programming Framework: Given one line Requirement, return PRD, Design, Tasks, Repo | | 68|gpt-engineer-org/gpt-engineer !2025-03-285367419|Specify what you want it to build, the AI asks for clarification, and then builds it.| | 69|lencx/ChatGPT !2025-03-2853653-3|🔮 ChatGPT Desktop Application (Mac, Windows and Linux)| | 70|deepfakes/faceswap !2025-03-28535672|Deepfakes Software For All| | 71|langflow-ai/langflow !2025-03-285319584|Langflow is a low-code app builder for RAG and multi-agent AI applications. It’s Python-based and agnostic to any model, API, or database.| | 72|commaai/openpilot !2025-03-28529759|openpilot is an operating system for robotics. Currently, it upgrades the driver assistance system on 275+ supported cars.| | 73|clash-verge-rev/clash-verge-rev !2025-03-2852848124|Continuation of Clash Verge - A Clash Meta GUI based on Tauri (Windows, MacOS, Linux)| | 74|All-Hands-AI/OpenHands !2025-03-285150675|🙌 OpenHands: Code Less, Make More| | 75|xai-org/grok-1 !2025-03-28502504|Grok open release| | 76|meilisearch/meilisearch !2025-03-284999122|A lightning-fast search API that fits effortlessly into your apps, websites, and workflow| | 77|🔥browser-use/browser-use !2025-03-2849910294|Make websites accessible for AI agents| | 78|jgthms/bulma !2025-03-28496783|Modern CSS framework based on Flexbox| | 79|facebookresearch/segment-anything !2025-03-284947116|The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.| |!green-up-arrow.svg 80|hacksider/Deep-Live-Cam !2025-03-2848612146|real time face swap and one-click video deepfake with only a single image (uncensored)| |!red-down-arrow 81|mlabonne/llm-course !2025-03-284860934|Course with a roadmap and notebooks to get into Large Language Models (LLMs).| | 82|PaddlePaddle/PaddleOCR !2025-03-284785530|Awesome multilingual OCR toolkits based on PaddlePaddle (practical ultra lightweight OCR system, support 80+ languages recognition, provide data annotation and synthesis tools, support training and deployment among server, mobile, embedded and IoT devices)| | 83|alist-org/alist !2025-03-284732618|🗂️A file list/WebDAV program that supports multiple storages, powered by Gin and Solidjs. / 一个支持多存储的文件列表/WebDAV程序,使用 Gin 和 Solidjs。| | 84|infiniflow/ragflow !2025-03-2847027129|RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.| | 85|Avik-Jain/100-Days-Of-ML-Code !2025-03-284679312|100 Days of ML Coding| | 86|v2ray/v2ray-core !2025-03-28458706|A platform for building proxies to bypass network restrictions.| | 87|hiyouga/LLaMA-Factory !2025-03-284555881|Easy-to-use LLM fine-tuning framework (LLaMA, BLOOM, Mistral, Baichuan, Qwen, ChatGLM)| | 88|Asabeneh/30-Days-Of-Python !2025-03-284544930|30 days of Python programming challenge is a step-by-step guide to learn the Python programming language in 30 days. This challenge may take more than100 days, follow your own pace. These videos may help too: https://www.youtube.com/channel/UC7PNRuno1rzYPb1xLa4yktw| | 89|type-challenges/type-challenges !2025-03-284488511|Collection of TypeScript type challenges with online judge| | 90|lllyasviel/Fooocus !2025-03-284402716|Focus on prompting and generating| | 91|RVC-Boss/GPT-SoVITS !2025-03-284327738|1 min voice data can also be used to train a good TTS model! (few shot voice cloning)| | 92|rasbt/LLMs-from-scratch !2025-03-284320667|Implementing a ChatGPT-like LLM from scratch, step by step| | 93|oobabooga/text-generation-webui !2025-03-284302012 |A gradio web UI for running Large Language Models like LLaMA, llama.cpp, GPT-J, OPT, and GALACTICA.| | 94|vllm-project/vllm !2025-03-2842982102|A high-throughput and memory-efficient inference and serving engine for LLMs| | 95|dani-garcia/vaultwarden !2025-03-284297121|Unofficial Bitwarden compatible server written in Rust, formerly known as bitwarden_rs| | 96|microsoft/autogen !2025-03-284233049|Enable Next-Gen Large Language Model Applications. Join our Discord: https://discord.gg/pAbnFJrkgZ| | 97|jeecgboot/JeecgBoot !2025-03-284205920|🔥「企业级低代码平台」前后端分离架构SpringBoot 2.x/3.x,SpringCloud,Ant Design&Vue3,Mybatis,Shiro,JWT。强大的代码生成器让前后端代码一键生成,无需写任何代码! 引领新的开发模式OnlineCoding->代码生成->手工MERGE,帮助Java项目解决70%重复工作,让开发更关注业务,既能快速提高效率,帮助公司节省成本,同时又不失灵活性。| | 98|Mintplex-Labs/anything-llm !2025-03-284186955|A full-stack application that turns any documents into an intelligent chatbot with a sleek UI and easier way to manage your workspaces.| | 99|THUDM/ChatGLM-6B !2025-03-28410192 |ChatGLM-6B: An Open Bilingual Dialogue Language Model| | 100|hpcaitech/ColossalAI !2025-03-28406902|Making large AI models cheaper, faster and more accessible| | 101|Stability-AI/stablediffusion !2025-03-28406337|High-Resolution Image Synthesis with Latent Diffusion Models| | 102|mingrammer/diagrams !2025-03-28405063|🎨 Diagram as Code for prototyping cloud system architectures| | 103|Kong/kong !2025-03-28404616|🦍 The Cloud-Native API Gateway and AI Gateway.| | 104|getsentry/sentry !2025-03-284040913|Developer-first error tracking and performance monitoring| | 105| karpathy/nanoGPT !2025-03-284034613 |The simplest, fastest repository for training/finetuning medium-sized GPTs| | 106|fastlane/fastlane !2025-03-2840014-1|🚀 The easiest way to automate building and releasing your iOS and Android apps| | 107|psf/black !2025-03-28399765|The uncompromising Python code formatter| | 108|OpenBB-finance/OpenBBTerminal !2025-03-283972074 |Investment Research for Everyone, Anywhere.| | 109|2dust/v2rayNG !2025-03-283943415|A V2Ray client for Android, support Xray core and v2fly core| | 110|apache/airflow !2025-03-283937314|Apache Airflow - A platform to programmatically author, schedule, and monitor workflows| | 111|KRTirtho/spotube !2025-03-283902746|🎧 Open source Spotify client that doesn't require Premium nor uses Electron! Available for both desktop & mobile!| | 112|coqui-ai/TTS !2025-03-283889719 |🐸💬 - a deep learning toolkit for Text-to-Speech, battle-tested in research and production| | 113|ggerganov/whisper.cpp !2025-03-283882116|Port of OpenAI's Whisper model in C/C++| | 114|ultralytics/ultralytics !2025-03-283866951|NEW - YOLOv8 🚀 in PyTorch > ONNX > OpenVINO > CoreML > TFLite| | 115|typst/typst !2025-03-283863914|A new markup-based typesetting system that is powerful and easy to learn.| | 116|streamlit/streamlit !2025-03-283845828|Streamlit — A faster way to build and share data apps.| | 117|LC044/WeChatMsg !2025-03-283836931|提取微信聊天记录,将其导出成HTML、Word、Excel文档永久保存,对聊天记录进行分析生成年度聊天报告,用聊天数据训练专属于个人的AI聊天助手| | 118|lm-sys/FastChat !2025-03-283822112 |An open platform for training, serving, and evaluating large languages. Release repo for Vicuna and FastChat-T5.| | 119|NaiboWang/EasySpider !2025-03-283819013|A visual no-code/code-free web crawler/spider易采集:一个可视化浏览器自动化测试/数据采集/爬虫软件,可以无代码图形化的设计和执行爬虫任务。别名:ServiceWrapper面向Web应用的智能化服务封装系统。| | 120|microsoft/DeepSpeed !2025-03-283765816 |A deep learning optimization library that makes distributed training and inference easy, efficient, and effective| | 121|QuivrHQ/quivr !2025-03-28376067|Your GenAI Second Brain 🧠 A personal productivity assistant (RAG) ⚡️🤖 Chat with your docs (PDF, CSV, ...) & apps using Langchain, GPT 3.5 / 4 turbo, Private, Anthropic, VertexAI, Ollama, LLMs, that you can share with users ! Local & Private alternative to OpenAI GPTs & ChatGPT powered by retrieval-augmented generation.| | 122|freqtrade/freqtrade !2025-03-283757817 |Free, open source crypto trading bot| | 123|suno-ai/bark !2025-03-28373178 |🔊 Text-Prompted Generative Audio Model| | 124|🔥cline/cline !2025-03-2837307282|Autonomous coding agent right in your IDE, capable of creating/editing files, executing commands, and more with your permission every step of the way.| | 125|LAION-AI/Open-Assistant !2025-03-28372712 |OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.| | 126|penpot/penpot !2025-03-283716217|Penpot: The open-source design tool for design and code collaboration| | 127|gradio-app/gradio !2025-03-283713320|Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work!| | 128|FlowiseAI/Flowise !2025-03-283667135 |Drag & drop UI to build your customized LLM flow using LangchainJS| | 129|SimplifyJobs/Summer2025-Internships !2025-03-28366506|Collection of Summer 2025 tech internships!| | 130|TencentARC/GFPGAN !2025-03-28365027 |GFPGAN aims at developing Practical Algorithms for Real-world Face Restoration.| | 131|ray-project/ray !2025-03-283626819|Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a toolkit of libraries (Ray AIR) for accelerating ML workloads.| | 132|babysor/MockingBird !2025-03-28360498|🚀AI拟声: 5秒内克隆您的声音并生成任意语音内容 Clone a voice in 5 seconds to generate arbitrary speech in real-time| | 133|unslothai/unsloth !2025-03-283603691|5X faster 50% less memory LLM finetuning| | 134|zhayujie/chatgpt-on-wechat !2025-03-283600124 |Wechat robot based on ChatGPT, which uses OpenAI api and itchat library| | 135|upscayl/upscayl !2025-03-283599824|🆙 Upscayl - Free and Open Source AI Image Upscaler for Linux, MacOS and Windows built with Linux-First philosophy.| | 136|freeCodeCamp/devdocs !2025-03-28359738|API Documentation Browser| | 137|XingangPan/DragGAN !2025-03-28359043 |Code for DragGAN (SIGGRAPH 2023)| | 138|2noise/ChatTTS !2025-03-283543922|ChatTTS is a generative speech model for daily dialogue.| | 139|google-research/google-research !2025-03-28352207 |Google Research| | 140|karanpratapsingh/system-design !2025-03-28351003|Learn how to design systems at scale and prepare for system design interviews| | 141|lapce/lapce !2025-03-28350855|Lightning-fast and Powerful Code Editor written in Rust| | 142| microsoft/TaskMatrix !2025-03-2834500-3 | Talking, Drawing and Editing with Visual Foundation Models| | 143|chatchat-space/Langchain-Chatchat !2025-03-283442020|Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM) QA app with langchain| | 144|unclecode/crawl4ai !2025-03-283434163|🔥🕷️ Crawl4AI: Open-source LLM Friendly Web Crawler & Scrapper| | 145|Bin-Huang/chatbox !2025-03-283374733 |A desktop app for GPT-4 / GPT-3.5 (OpenAI API) that supports Windows, Mac & Linux| | 146|milvus-io/milvus !2025-03-283366525 |A cloud-native vector database, storage for next generation AI applications| | 147|mendableai/firecrawl !2025-03-2833297128|🔥 Turn entire websites into LLM-ready markdown| | 148|pola-rs/polars !2025-03-283269320|Fast multi-threaded, hybrid-out-of-core query engine focussing on DataFrame front-ends| | 149|Pythagora-io/gpt-pilot !2025-03-28325321|PoC for a scalable dev tool that writes entire apps from scratch while the developer oversees the implementation| | 150|hashicorp/vault !2025-03-28320797|A tool for secrets management, encryption as a service, and privileged access management| | 151|shardeum/shardeum !2025-03-28319580|Shardeum is an EVM based autoscaling blockchain| | 152|Chanzhaoyu/chatgpt-web !2025-03-28319242 |A demonstration website built with Express and Vue3 called ChatGPT| | 153|lllyasviel/ControlNet !2025-03-283186413 |Let us control diffusion models!| | 154|google/jax !2025-03-28317727|Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more| | 155|facebookresearch/detectron2 !2025-03-28315987|Detectron2 is a platform for object detection, segmentation and other visual recognition tasks.| | 156|myshell-ai/OpenVoice !2025-03-28315233|Instant voice cloning by MyShell| | 157|TheAlgorithms/C-Plus-Plus !2025-03-283151411|Collection of various algorithms in mathematics, machine learning, computer science and physics implemented in C++ for educational purposes.| | 158|hiroi-sora/Umi-OCR !2025-03-283138129|OCR图片转文字识别软件,完全离线。截屏/批量导入图片,支持多国语言、合并段落、竖排文字。可排除水印区域,提取干净的文本。基于 PaddleOCR 。| | 159|mudler/LocalAI !2025-03-283127815|🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.| | 160|facebookresearch/fairseq !2025-03-28312124 |Facebook AI Research Sequence-to-Sequence Toolkit written in Python.| | 161|alibaba/nacos !2025-03-28310559|an easy-to-use dynamic service discovery, configuration and service management platform for building cloud native applications.| | 162|yunjey/pytorch-tutorial !2025-03-28310326|PyTorch Tutorial for Deep Learning Researchers| | 163|v2fly/v2ray-core !2025-03-28307448|A platform for building proxies to bypass network restrictions.| | 164|mckaywrigley/chatbot-ui !2025-03-283067714|The open-source AI chat interface for everyone.| | 165|TabbyML/tabby !2025-03-28305949 |Self-hosted AI coding assistant| | 166|deepseek-ai/awesome-deepseek-integration !2025-03-283053193|| | 167|danielmiessler/fabric !2025-03-283028914|fabric is an open-source framework for augmenting humans using AI.| | 168|xinntao/Real-ESRGAN !2025-03-283026623 |Real-ESRGAN aims at developing Practical Algorithms for General Image/Video Restoration.| | 169|paul-gauthier/aider !2025-03-283014642|aider is GPT powered coding in your terminal| | 170|tatsu-lab/stanfordalpaca !2025-03-28299022 |Code and documentation to train Stanford's Alpaca models, and generate the data.| | 171|DataTalksClub/data-engineering-zoomcamp !2025-03-282971817|Free Data Engineering course!| | 172|HeyPuter/puter !2025-03-282967014|🌐 The Internet OS! Free, Open-Source, and Self-Hostable.| | 173|mli/paper-reading !2025-03-282962314|Classic Deep Learning and In-Depth Reading of New Papers Paragraph by Paragraph| | 174|linexjlin/GPTs !2025-03-28295568|leaked prompts of GPTs| | 175|s0md3v/roop !2025-03-28295286 |one-click deepfake (face swap)| | 176|JushBJJ/Mr.-Ranedeer-AI-Tutor !2025-03-2829465-1 |A GPT-4 AI Tutor Prompt for customizable personalized learning experiences.| | 177|opendatalab/MinerU !2025-03-282927074|A one-stop, open-source, high-quality data extraction tool, supports PDF/webpage/e-book extraction.一站式开源高质量数据提取工具,支持PDF/网页/多格式电子书提取。| | 178|mouredev/Hello-Python !2025-03-282920720|Curso para aprender el lenguaje de programación Python desde cero y para principiantes. 75 clases, 37 horas en vídeo, código, proyectos y grupo de chat. Fundamentos, frontend, backend, testing, IA...| | 179|Lightning-AI/pytorch-lightning !2025-03-28292039|Pretrain, finetune and deploy AI models on multiple GPUs, TPUs with zero code changes.| | 180|crewAIInc/crewAI !2025-03-282919344|Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.| | 181|facebook/folly !2025-03-282916612|An open-source C++ library developed and used at Facebook.| | 182|google-ai-edge/mediapipe !2025-03-28291519|Cross-platform, customizable ML solutions for live and streaming media.| | 183| getcursor/cursor !2025-03-282892025 | An editor made for programming with AI| | 184|chatanywhere/GPTAPIfree !2025-03-282856424|Free ChatGPT API Key, Free ChatGPT API, supports GPT-4 API (free), ChatGPT offers a free domestic forwarding API that allows direct connections without the need for a proxy. It can be used in conjunction with software/plugins like ChatBox, significantly reducing interface usage costs. Enjoy unlimited and unrestricted chatting within China| | 185|meta-llama/llama3 !2025-03-28285552|The official Meta Llama 3 GitHub site| | 186|tinygrad/tinygrad !2025-03-282845811|You like pytorch? You like micrograd? You love tinygrad! ❤️| | 187|google-research/tuningplaybook !2025-03-282841514|A playbook for systematically maximizing the performance of deep learning models.| | 188|huggingface/diffusers !2025-03-282830222|🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.| | 189|tokio-rs/tokio !2025-03-28282408|A runtime for writing reliable asynchronous applications with Rust. Provides I/O, networking, scheduling, timers, ...| | 190|RVC-Project/Retrieval-based-Voice-Conversion-WebUI !2025-03-282823817|Voice data !2025-03-282822612|Jan is an open source alternative to ChatGPT that runs 100% offline on your computer| | 192|openai/CLIP !2025-03-282814720|CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image| | 193|🔥khoj-ai/khoj !2025-03-2828112313|Your AI second brain. A copilot to get answers to your questions, whether they be from your own notes or from the internet. Use powerful, online (e.g gpt4) or private, local (e.g mistral) LLMs. Self-host locally or use our web app. Access from Obsidian, Emacs, Desktop app, Web or Whatsapp.| | 194| acheong08/ChatGPT !2025-03-2828054-2 | Reverse engineered ChatGPT API | | 195|iperov/DeepFaceLive !2025-03-28279345 |Real-time face swap for PC streaming or video calls| | 196|eugeneyan/applied-ml !2025-03-28278471|📚 Papers & tech blogs by companies sharing their work on data science & machine learning in production.| | 197|XTLS/Xray-core !2025-03-282778213|Xray, Penetrates Everything. Also the best v2ray-core, with XTLS support. Fully compatible configuration.| | 198|feder-cr/JobsApplierAIAgent !2025-03-282776410|AutoJobsApplierAI_Agent aims to easy job hunt process by automating the job application process. Utilizing artificial intelligence, it enables users to apply for multiple jobs in an automated and personalized way.| | 199|mindsdb/mindsdb !2025-03-282750631|The platform for customizing AI from enterprise data| | 200|DataExpert-io/data-engineer-handbook !2025-03-282721611|This is a repo with links to everything you'd ever want to learn about data engineering| | 201|exo-explore/exo !2025-03-282721633|Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚| | 202|taichi-dev/taichi !2025-03-2826926-1|Productive, portable, and performant GPU programming in Python.| | 203|mem0ai/mem0 !2025-03-282689134|The memory layer for Personalized AI| | 204|svc-develop-team/so-vits-svc !2025-03-28268096 |SoftVC VITS Singing Voice Conversion| | 205|OpenBMB/ChatDev !2025-03-28265624|Create Customized Software using Natural Language Idea (through Multi-Agent Collaboration)| | 206|roboflow/supervision !2025-03-282632010|We write your reusable computer vision tools. 💜| | 207|drawdb-io/drawdb !2025-03-282626913|Free, simple, and intuitive online database design tool and SQL generator.| | 208|karpathy/llm.c !2025-03-28261633|LLM training in simple, raw C/CUDA| | 209|airbnb/lottie-ios !2025-03-28261431|An iOS library to natively render After Effects vector animations| | 210|openai/openai-python !2025-03-282607713|The OpenAI Python library provides convenient access to the OpenAI API from applications written in the Python language.| | 211|academic/awesome-datascience !2025-03-28259876|📝 An awesome Data Science repository to learn and apply for real world problems.| | 212|harry0703/MoneyPrinterTurbo !2025-03-282576618|Generate short videos with one click using a large model| | 213|gabime/spdlog !2025-03-282571511|Fast C++ logging library.| | 214|ocrmypdf/OCRmyPDF !2025-03-2825674217|OCRmyPDF adds an OCR text layer to scanned PDF files, allowing them to be searched| | 215|Vision-CAIR/MiniGPT-4 !2025-03-28256170 |Enhancing Vision-language Understanding with Advanced Large Language Models| | 216|Stability-AI/generative-models !2025-03-28255936|Generative Models by Stability AI| | 217|DS4SD/docling !2025-03-282555662|Get your docs ready for gen AI| | 218|PostHog/posthog !2025-03-282533227|🦔 PostHog provides open-source product analytics, session recording, feature flagging and A/B testing that you can self-host.| | 219|nrwl/nx !2025-03-282509612|Smart Monorepos · Fast CI| | 220|continuedev/continue !2025-03-282500737|⏩ the open-source copilot chat for software development—bring the power of ChatGPT to VS Code| | 221|opentofu/opentofu !2025-03-28247968|OpenTofu lets you declaratively manage your cloud infrastructure.| | 222|invoke-ai/InvokeAI !2025-03-28247293|InvokeAI is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The solution offers an industry leading WebUI, supports terminal use through a CLI, and serves as the foundation for multiple commercial products.| | 223|deepinsight/insightface !2025-03-282471615 |State-of-the-art 2D and 3D Face Analysis Project| | 224|apache/flink !2025-03-28246865|Apache Flink| | 225|ComposioHQ/composio !2025-03-28246436|Composio equips agents with well-crafted tools empowering them to tackle complex tasks| | 226|Genesis-Embodied-AI/Genesis !2025-03-282458314|A generative world for general-purpose robotics & embodied AI learning.| | 227|stretchr/testify !2025-03-28243184|A toolkit with common assertions and mocks that plays nicely with the standard library| | 228| yetone/openai-translator !2025-03-28242921 | Browser extension and cross-platform desktop application for translation based on ChatGPT API | | 229|frappe/erpnext !2025-03-282425211|Free and Open Source Enterprise Resource Planning (ERP)| | 230|songquanpeng/one-api !2025-03-282410034|OpenAI 接口管理 & 分发系统,支持 Azure、Anthropic Claude、Google PaLM 2 & Gemini、智谱 ChatGLM、百度文心一言、讯飞星火认知、阿里通义千问、360 智脑以及腾讯混元,可用于二次分发管理 key,仅单可执行文件,已打包好 Docker 镜像,一键部署,开箱即用. OpenAI key management & redistribution system, using a single API for all LLMs, and features an English UI.| | 231| microsoft/JARVIS !2025-03-28240604 | a system to connect LLMs with ML community | | 232|google/flatbuffers !2025-03-28239965|FlatBuffers: Memory Efficient Serialization Library| | 233|microsoft/graphrag !2025-03-282398928|A modular graph-based Retrieval-Augmented Generation (RAG) system| | 234|rancher/rancher !2025-03-28239675|Complete container management platform| | 235|bazelbuild/bazel !2025-03-282384618|a fast, scalable, multi-language and extensible build system| | 236|modularml/mojo !2025-03-28238236 |The Mojo Programming Language| | 237|danny-avila/LibreChat !2025-03-282378753|Enhanced ChatGPT Clone: Features OpenAI, GPT-4 Vision, Bing, Anthropic, OpenRouter, Google Gemini, AI model switching, message search, langchain, DALL-E-3, ChatGPT Plugins, OpenAI Functions, Secure Multi-User System, Presets, completely open-source for self-hosting. More features in development| |!green-up-arrow.svg 238|🔥🔥🔥Shubhamsaboo/awesome-llm-apps !2025-03-28237391211|Collection of awesome LLM apps with RAG using OpenAI, Anthropic, Gemini and opensource models.| |!red-down-arrow 239|microsoft/semantic-kernel !2025-03-282373611|Integrate cutting-edge LLM technology quickly and easily into your apps| |!red-down-arrow 240|TheAlgorithms/Rust !2025-03-28236995|All Algorithms implemented in Rust| | 241|stanford-oval/storm !2025-03-28236326|An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.| | 242|openai/gpt-2 !2025-03-28232483|Code for the paper "Language Models are Unsupervised Multitask Learners"| | 243|labring/FastGPT !2025-03-282319445|A platform that uses the OpenAI API to quickly build an AI knowledge base, supporting many-to-many relationships.| | 244|pathwaycom/llm-app !2025-03-2822928-10|Ready-to-run cloud templates for RAG, AI pipelines, and enterprise search with live data. 🐳Docker-friendly.⚡Always in sync with Sharepoint, Google Drive, S3, Kafka, PostgreSQL, real-time data APIs, and more.| | 245|warpdotdev/Warp !2025-03-282286825|Warp is a modern, Rust-based terminal with AI built in so you and your team can build great software, faster.| | 246|🔥agno-agi/agno !2025-03-2822833298|Agno is a lightweight library for building Multimodal Agents. It exposes LLMs as a unified API and gives them superpowers like memory, knowledge, tools and reasoning.| | 247|qdrant/qdrant !2025-03-282275214 |Qdrant - Vector Database for the next generation of AI applications. Also available in the cloud https://cloud.qdrant.io/| | 248|ashishpatel26/500-AI-Machine-learning-Deep-learning-Computer-vision-NLP-Projects-with-code !2025-03-282271815|500 AI Machine learning Deep learning Computer vision NLP Projects with code| | 249|stanfordnlp/dspy !2025-03-282268321|Stanford DSPy: The framework for programming—not prompting—foundation models| | 250|PaddlePaddle/Paddle !2025-03-28226246|PArallel Distributed Deep LEarning: Machine Learning Framework from Industrial Practice (『飞桨』核心框架,深度学习&机器学习高性能单机、分布式训练和跨平台部署)| | 251|zulip/zulip !2025-03-28225464|Zulip server and web application. Open-source team chat that helps teams stay productive and focused.| | 252|Hannibal046/Awesome-LLM !2025-03-282240721|Awesome-LLM: a curated list of Large Language Model| | 253|facefusion/facefusion !2025-03-282218812|Next generation face swapper and enhancer| | 254|Mozilla-Ocho/llamafile !2025-03-28220624|Distribute and run LLMs with a single file.| | 255|yuliskov/SmartTube !2025-03-282201614|SmartTube - an advanced player for set-top boxes and tvs running Android OS| | 256|haotian-liu/LLaVA !2025-03-282201316 |Large Language-and-Vision Assistant built towards multimodal GPT-4 level capabilities.| | 257|ashishps1/awesome-system-design-resources !2025-03-282189367|This repository contains System Design resources which are useful while preparing for interviews and learning Distributed Systems| | 258|Cinnamon/kotaemon !2025-03-28218248|An open-source RAG-based tool for chatting with your documents.| | 259|CodePhiliaX/Chat2DB !2025-03-282179757|🔥🔥🔥AI-driven database tool and SQL client, The hottest GUI client, supporting MySQL, Oracle, PostgreSQL, DB2, SQL Server, DB2, SQLite, H2, ClickHouse, and more.| | 260|blakeblackshear/frigate !2025-03-282177113|NVR with realtime local object detection for IP cameras| | 261|facebookresearch/audiocraft !2025-03-28217111|Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable music generation LM with textual and melodic conditioning.| | 262|karpathy/minGPT !2025-03-28216567|A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training| | 263|grpc/grpc-go !2025-03-282159510|The Go language implementation of gRPC. HTTP/2 based RPC| | 264|HumanSignal/label-studio !2025-03-282137618|Label Studio is a multi-type data labeling and annotation tool with standardized output format| | 265|yoheinakajima/babyagi !2025-03-28212764 |uses OpenAI and Pinecone APIs to create, prioritize, and execute tasks, This is a pared-down version of the original Task-Driven Autonomous Agent| | 266|deepseek-ai/DeepSeek-Coder !2025-03-282118210|DeepSeek Coder: Let the Code Write Itself| | 267|BuilderIO/gpt-crawler !2025-03-282118010|Crawl a site to generate knowledge files to create your own custom GPT from a URL| | 268| openai/chatgpt-retrieval-plugin !2025-03-2821152-1 | Plugins are chat extensions designed specifically for language models like ChatGPT, enabling them to access up-to-date information, run computations, or interact with third-party services in response to a user's request.| | 269|microsoft/OmniParser !2025-03-282113123|A simple screen parsing tool towards pure vision based GUI agent| | 270|black-forest-labs/flux !2025-03-282107219|Official inference repo for FLUX.1 models| | 271|ItzCrazyKns/Perplexica !2025-03-282099154|Perplexica is an AI-powered search engine. It is an Open source alternative to Perplexity AI| | 272|microsoft/unilm !2025-03-28209876|Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities| | 273|Sanster/lama-cleaner !2025-03-282077614|Image inpainting tool powered by SOTA AI Model. Remove any unwanted object, defect, people from your pictures or erase and replace(powered by stable diffusion) any thing on your pictures.| | 274|assafelovic/gpt-researcher !2025-03-282057222|GPT based autonomous agent that does online comprehensive research on any given topic| | 275|PromtEngineer/localGPT !2025-03-28204230 |Chat with your documents on your local device using GPT models. No data leaves your device and 100% private.| | 276|elastic/kibana !2025-03-28203482|Your window into the Elastic Stack| | 277|fishaudio/fish-speech !2025-03-282033222|Brand new TTS solution| | 278|mlc-ai/mlc-llm !2025-03-282028110 |Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.| | 279|deepset-ai/haystack !2025-03-282005320|🔍 Haystack is an open source NLP framework to interact with your data using Transformer models and LLMs (GPT-4, ChatGPT and alike). Haystack offers production-ready tools to quickly build complex question answering, semantic search, text generation applications, and more.| | 280|tree-sitter/tree-sitter !2025-03-28200487|An incremental parsing system for programming tools| | 281|Anjok07/ultimatevocalremovergui !2025-03-281999811|GUI for a Vocal Remover that uses Deep Neural Networks.| | 282|guidance-ai/guidance !2025-03-28199622|A guidance language for controlling large language models.| | 283|ml-explore/mlx !2025-03-28199619|MLX: An array framework for Apple silicon| | 284|mlflow/mlflow !2025-03-281995314|Open source platform for the machine learning lifecycle| | 285|ml-tooling/best-of-ml-python !2025-03-28198631|🏆 A ranked list of awesome machine learning Python libraries. Updated weekly.| | 286|BerriAI/litellm !2025-03-281981862|Call all LLM APIs using the OpenAI format. Use Bedrock, Azure, OpenAI, Cohere, Anthropic, Ollama, Sagemaker, HuggingFace, Replicate (100+ LLMs)| | 287|LazyVim/LazyVim !2025-03-281981320|Neovim config for the lazy| | 288|wez/wezterm !2025-03-281976018|A GPU-accelerated cross-platform terminal emulator and multiplexer written by @wez and implemented in Rust| | 289|valkey-io/valkey !2025-03-281970416|A flexible distributed key-value datastore that supports both caching and beyond caching workloads.| | 290|LiLittleCat/awesome-free-chatgpt !2025-03-28196185|🆓免费的 ChatGPT 镜像网站列表,持续更新。List of free ChatGPT mirror sites, continuously updated.| | 291|Byaidu/PDFMathTranslate !2025-03-281947645|PDF scientific paper translation with preserved formats - 基于 AI 完整保留排版的 PDF 文档全文双语翻译,支持 Google/DeepL/Ollama/OpenAI 等服务,提供 CLI/GUI/Docker| | 292|openai/swarm !2025-03-281947111|Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.| | 293|HqWu-HITCS/Awesome-Chinese-LLM !2025-03-281921423|Organizing smaller, cost-effective, privately deployable open-source Chinese language models, including related datasets and tutorials| | 294|stitionai/devika !2025-03-28190903|Devika is an Agentic AI Software Engineer that can understand high-level human instructions, break them down into steps, research relevant information, and write code to achieve the given objective. Devika aims to be a competitive open-source alternative to Devin by Cognition AI.| | 295|OpenBMB/MiniCPM-o !2025-03-28190887|MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone| | 296|samber/lo !2025-03-281904815|💥 A Lodash-style Go library based on Go 1.18+ Generics (map, filter, contains, find...)| | 297|chroma-core/chroma !2025-03-281895221 |the AI-native open-source embedding database| | 298|DarkFlippers/unleashed-firmware !2025-03-28189278|Flipper Zero Unleashed Firmware| | 299|brave/brave-browser !2025-03-281892710|Brave browser for Android, iOS, Linux, macOS, Windows.| | 300| tloen/alpaca-lora !2025-03-28188641 | Instruct-tune LLaMA on consumer hardware| | 301|VinciGit00/Scrapegraph-ai !2025-03-281884618|Python scraper based on AI| | 302|gitroomhq/postiz-app !2025-03-281879110|📨 Schedule social posts, measure them, exchange with other members and get a lot of help from AI 🚀| | 303|PrefectHQ/prefect !2025-03-281878715|Prefect is a workflow orchestration tool empowering developers to build, observe, and react to data pipelines| | 304|ymcui/Chinese-LLaMA-Alpaca !2025-03-28187723 |Chinese LLaMA & Alpaca LLMs| | 305|kenjihiranabe/The-Art-of-Linear-Algebra !2025-03-28187335|Graphic notes on Gilbert Strang's "Linear Algebra for Everyone"| | 306|joonspk-research/generativeagents !2025-03-28187288|Generative Agents: Interactive Simulacra of Human Behavior| | 307|renovatebot/renovate !2025-03-28186820|Universal dependency update tool that fits into your workflows.| | 308|gventuri/pandas-ai !2025-03-28186109 |Pandas AI is a Python library that integrates generative artificial intelligence capabilities into Pandas, making dataframes conversational| | 309|thingsboard/thingsboard !2025-03-28185184|Open-source IoT Platform - Device management, data collection, processing and visualization.| | 310|ente-io/ente !2025-03-28184722|Fully open source, End to End Encrypted alternative to Google Photos and Apple Photos| | 311|serengil/deepface !2025-03-281840113|A Lightweight Face Recognition and Facial Attribute Analysis (Age, Gender, Emotion and Race) Library for Python| | 312|Raphire/Win11Debloat !2025-03-281840132|A simple, easy to use PowerShell script to remove pre-installed apps from windows, disable telemetry, remove Bing from windows search as well as perform various other changes to declutter and improve your windows experience. This script works for both windows 10 and windows 11.| | 313|Avaiga/taipy !2025-03-28179235|Turns Data and AI algorithms into production-ready web applications in no time.| | 314|microsoft/qlib !2025-03-281784231|Qlib is an AI-oriented quantitative investment platform that aims to realize the potential, empower research, and create value using AI technologies in quantitative investment, from exploring ideas to implementing productions. Qlib supports diverse machine learning modeling paradigms. including supervised learning, market dynamics modeling, and RL.| | 315|CopilotKit/CopilotKit !2025-03-281778571|Build in-app AI chatbots 🤖, and AI-powered Textareas ✨, into react web apps.| | 316|QwenLM/Qwen-7B !2025-03-281766017|The official repo of Qwen-7B (通义千问-7B) chat & pretrained large language model proposed by Alibaba Cloud.| | 317|w-okada/voice-changer !2025-03-28176078 |リアルタイムボイスチェンジャー Realtime Voice Changer| | 318|rlabbe/Kalman-and-Bayesian-Filters-in-Python !2025-03-281756011|Kalman Filter book using Jupyter Notebook. Focuses on building intuition and experience, not formal proofs. Includes Kalman filters,extended Kalman filters, unscented Kalman filters, particle filters, and more. All exercises include solutions.| | 319|Mikubill/sd-webui-controlnet !2025-03-28174794 |WebUI extension for ControlNet| | 320|jingyaogong/minimind !2025-03-2817380116|「大模型」3小时完全从0训练26M的小参数GPT,个人显卡即可推理训练!| | 321|apify/crawlee !2025-03-28172696|Crawlee—A web scraping and browser automation library for Node.js to build reliable crawlers. In JavaScript and TypeScript. Extract data for AI, LLMs, RAG, or GPTs. Download HTML, PDF, JPG, PNG, and other files from websites. Works with Puppeteer, Playwright, Cheerio, JSDOM, and raw HTTP. Both headful and headless mode. With proxy rotation.| | 322|apple/ml-stable-diffusion !2025-03-28172395|Stable Diffusion with Core ML on Apple Silicon| | 323| transitive-bullshit/chatgpt-api !2025-03-28172095 | Node.js client for the official ChatGPT API. | | 324|teableio/teable !2025-03-281719222|✨ The Next Gen Airtable Alternative: No-Code Postgres| | 325| xx025/carrot !2025-03-28170900 | Free ChatGPT Site List | | 326|microsoft/LightGBM !2025-03-28170723|A fast, distributed, high-performance gradient boosting (GBT, GBDT, GBRT, GBM or MART) framework based on decision tree algorithms, used for ranking, classification and many other machine learning tasks.| | 327|VikParuchuri/surya !2025-03-28169827|Accurate line-level text detection and recognition (OCR) in any language| | 328|deepseek-ai/Janus !2025-03-281692825|Janus-Series: Unified Multimodal Understanding and Generation Models| | 329|ardalis/CleanArchitecture !2025-03-28168823|Clean Architecture Solution Template: A starting point for Clean Architecture with ASP.NET Core| | 330|neondatabase/neon !2025-03-28166466|Neon: Serverless Postgres. We separated storage and compute to offer autoscaling, code-like database branching, and scale to zero.| | 331|kestra-io/kestra !2025-03-281661313|⚡ Workflow Automation Platform. Orchestrate & Schedule code in any language, run anywhere, 500+ plugins. Alternative to Zapier, Rundeck, Camunda, Airflow...| | 332|Dao-AILab/flash-attention !2025-03-281659720|Fast and memory-efficient exact attention| | 333|RPCS3/rpcs3 !2025-03-281655712|PS3 emulator/debugger| | 334|meta-llama/llama-recipes !2025-03-28165486|Scripts for fine-tuning Llama2 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization & question answering. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment.Demo apps to showcase Llama2 for WhatsApp & Messenger| | 335|emilwallner/Screenshot-to-code !2025-03-28165180|A neural network that transforms a design mock-up into a static website.| | 336|datawhalechina/llm-cookbook !2025-03-281650922|面向开发者的 LLM 入门教程,吴恩达大模型系列课程中文版| | 337|e2b-dev/awesome-ai-agents !2025-03-281643923|A list of AI autonomous agents| | 338|QwenLM/Qwen2.5 !2025-03-281641114|Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.| | 339|dair-ai/ML-YouTube-Courses !2025-03-28164114|📺 Discover the latest machine learning / AI courses on YouTube.| | 340|pybind/pybind11 !2025-03-28163620|Seamless operability between C++11 and Python| | 341|graphdeco-inria/gaussian-splatting !2025-03-281627116|Original reference implementation of "3D Gaussian Splatting for Real-Time Radiance Field Rendering"| | 342|meta-llama/codellama !2025-03-28162531|Inference code for CodeLlama models| | 343|TransformerOptimus/SuperAGI !2025-03-28161292 | SuperAGI - A dev-first open source autonomous AI agent framework. Enabling developers to build, manage & run useful autonomous agents quickly and reliably.| | 344|microsoft/onnxruntime !2025-03-28161169|ONNX Runtime: cross-platform, high-performance ML inferencing and training accelerator| | 345|IDEA-Research/Grounded-Segment-Anything !2025-03-281601411 |Marrying Grounding DINO with Segment Anything & Stable Diffusion & BLIP - Automatically Detect, Segment and Generate Anything with Image and Text Inputs| | 346|ddbourgin/numpy-ml !2025-03-28160054|Machine learning, in numpy| | 347|eosphoros-ai/DB-GPT !2025-03-281585225|Revolutionizing Database Interactions with Private LLM Technology| | 348|Stability-AI/StableLM !2025-03-28158310 |Stability AI Language Models| | 349|openai/evals !2025-03-28157935 |Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.| | 350|THUDM/ChatGLM2-6B !2025-03-28157500|ChatGLM2-6B: An Open Bilingual Chat LLM | | 351|sunner/ChatALL !2025-03-28156761 |Concurrently chat with ChatGPT, Bing Chat, Bard, Alpaca, Vincuna, Claude, ChatGLM, MOSS, iFlytek Spark, ERNIE and more, discover the best answers| | 352|abseil/abseil-cpp !2025-03-28156656|Abseil Common Libraries (C++)| | 353|NVIDIA/open-gpu-kernel-modules !2025-03-28156531|NVIDIA Linux open GPU kernel module source| | 354|letta-ai/letta !2025-03-281563718|Letta (formerly MemGPT) is a framework for creating LLM services with memory.| | 355|typescript-eslint/typescript-eslint !2025-03-28156211|✨ Monorepo for all the tooling which enables ESLint to support TypeScript| | 356|umijs/umi !2025-03-28156211|A framework in react community ✨| | 357|AI4Finance-Foundation/FinGPT !2025-03-281561215|Data-Centric FinGPT. Open-source for open finance! Revolutionize 🔥 We'll soon release the trained model.| | 358|amplication/amplication !2025-03-28156022|🔥🔥🔥 The Only Production-Ready AI-Powered Backend Code Generation| | 359|KindXiaoming/pykan !2025-03-28155477|Kolmogorov Arnold Networks| | 360|arc53/DocsGPT !2025-03-28154900|GPT-powered chat for documentation, chat with your documents| | 361|influxdata/telegraf !2025-03-28154502|Agent for collecting, processing, aggregating, and writing metrics, logs, and other arbitrary data.| | 362|microsoft/Bringing-Old-Photos-Back-to-Life !2025-03-28154084|Bringing Old Photo Back to Life (CVPR 2020 oral)| | 363|GaiZhenbiao/ChuanhuChatGPT !2025-03-2815394-2|GUI for ChatGPT API and many LLMs. Supports agents, file-based QA, GPT finetuning and query with web search. All with a neat UI.| | 364|Zeyi-Lin/HivisionIDPhotos !2025-03-281529710|⚡️HivisionIDPhotos: a lightweight and efficient AI ID photos tools. 一个轻量级的AI证件照制作算法。| | 365| mayooear/gpt4-pdf-chatbot-langchain !2025-03-281529518 | GPT4 & LangChain Chatbot for large PDF docs | | 366|1Panel-dev/MaxKB !2025-03-2815277148|? Based on LLM large language model knowledge base Q&A system. Ready to use out of the box, supports quick integration into third-party business systems. Officially produced by 1Panel| | 367|ai16z/eliza !2025-03-281526811|Conversational Agent for Twitter and Discord| | 368|apache/arrow !2025-03-28151684|Apache Arrow is a multi-language toolbox for accelerated data interchange and in-memory processing| | 369|princeton-nlp/SWE-agent !2025-03-281516119|SWE-agent: Agent Computer Interfaces Enable Software Engineering Language Models| | 370|mlc-ai/web-llm !2025-03-281509311 |Bringing large-language models and chat to web browsers. Everything runs inside the browser with no server support.| | 371|guillaumekln/faster-whisper !2025-03-281507117 |Faster Whisper transcription with CTranslate2| | 372|overleaf/overleaf !2025-03-28150316|A web-based collaborative LaTeX editor| | 373|triton-lang/triton !2025-03-28150169|Development repository for the Triton language and compiler| | 374|soxoj/maigret !2025-03-281500410|🕵️‍♂️ Collect a dossier on a person by username from thousands of sites| | 375|alibaba/lowcode-engine !2025-03-28149841|An enterprise-class low-code technology stack with scale-out design / 一套面向扩展设计的企业级低代码技术体系| | 376|espressif/esp-idf !2025-03-28148545|Espressif IoT Development Framework. Official development framework for Espressif SoCs.| | 377|pgvector/pgvector !2025-03-281484913|Open-source vector similarity search for Postgres| | 378|datawhalechina/leedl-tutorial !2025-03-28148246|《李宏毅深度学习教程》(李宏毅老师推荐👍),PDF下载地址:https://github.com/datawhalechina/leedl-tutorial/releases| | 379|xcanwin/KeepChatGPT !2025-03-28147972 |Using ChatGPT is more efficient and smoother, perfectly solving ChatGPT network errors. No longer do you need to frequently refresh the webpage, saving over 10 unnecessary steps| | 380|m-bain/whisperX !2025-03-281471313|WhisperX: Automatic Speech Recognition with Word-level Timestamps (& Diarization)| | 381|HumanAIGC/AnimateAnyone !2025-03-2814706-1|Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation| |!green-up-arrow.svg 382|naklecha/llama3-from-scratch !2025-03-281469024|llama3 implementation one matrix multiplication at a time| |!red-down-arrow 383| fauxpilot/fauxpilot !2025-03-28146871 | An open-source GitHub Copilot server | | 384|LlamaFamily/Llama-Chinese !2025-03-28145111|Llama Chinese Community, the best Chinese Llama large model, fully open source and commercially available| | 385|BradyFU/Awesome-Multimodal-Large-Language-Models !2025-03-281450121|Latest Papers and Datasets on Multimodal Large Language Models| | 386|vanna-ai/vanna !2025-03-281449819|🤖 Chat with your SQL database 📊. Accurate Text-to-SQL Generation via LLMs using RAG 🔄.| | 387|bleedline/aimoneyhunter !2025-03-28144845|AI Side Hustle Money Mega Collection: Teaching You How to Utilize AI for Various Side Projects to Earn Extra Income.| | 388|stefan-jansen/machine-learning-for-trading !2025-03-28144629|Code for Machine Learning for Algorithmic Trading, 2nd edition.| | 389|state-spaces/mamba !2025-03-28144139|Mamba: Linear-Time Sequence Modeling with Selective State Spaces| | 390|vercel/ai-chatbot !2025-03-281434614|A full-featured, hackable Next.js AI chatbot built by Vercel| | 391|steven-tey/novel !2025-03-281428410|Notion-style WYSIWYG editor with AI-powered autocompletions| | 392|unifyai/ivy !2025-03-281409348|Unified AI| | 393|chidiwilliams/buzz !2025-03-281402411 |Buzz transcribes and translates audio offline on your personal computer. Powered by OpenAI's Whisper.| | 394|lukas-blecher/LaTeX-OCR !2025-03-28139769|pix2tex: Using a ViT to convert images of equations into LaTeX code.| | 395|openai/tiktoken !2025-03-28139599|tiktoken is a fast BPE tokeniser for use with OpenAI's models.| | 396|nocobase/nocobase !2025-03-281391522|NocoBase is a scalability-first, open-source no-code/low-code platform for building business applications and enterprise solutions.| | 397|neonbjb/tortoise-tts !2025-03-28139010 |A multi-voice TTS system trained with an emphasis on quality| | 398|yamadashy/repomix !2025-03-281382036|📦 Repomix (formerly Repopack) is a powerful tool that packs your entire repository into a single, AI-friendly file. Perfect for when you need to feed your codebase to Large Language Models (LLMs) or other AI tools like Claude, ChatGPT, and Gemini.| | 399|adobe/react-spectrum !2025-03-28136766|A collection of libraries and tools that help you build adaptive, accessible, and robust user experiences.| | 400|THUDM/ChatGLM3 !2025-03-28136684|ChatGLM3 series: Open Bilingual Chat LLMs | | 401|NVIDIA/NeMo !2025-03-28134837|A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)| | 402|BlinkDL/RWKV-LM !2025-03-28134346 |RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it combines the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.| | 403| fuergaosi233/wechat-chatgpt !2025-03-28133330 | Use ChatGPT On Wechat via wechaty | | 404|udecode/plate !2025-03-28133325|A rich-text editor powered by AI| | 405|xenova/transformers.js !2025-03-281331219|State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!| | 406|stas00/ml-engineering !2025-03-281325615|Machine Learning Engineering Guides and Tools| | 407| wong2/chatgpt-google-extension !2025-03-2813241-1 | A browser extension that enhances search engines with ChatGPT, this repos will not be updated from 2023-02-20| | 408|mrdbourke/pytorch-deep-learning !2025-03-281317520|Materials for the Learn PyTorch for Deep Learning: Zero to Mastery course.| | 409|Koenkk/zigbee2mqtt !2025-03-28131544|Zigbee 🐝 to MQTT bridge 🌉, get rid of your proprietary Zigbee bridges 🔨| | 410|vercel-labs/ai !2025-03-281298528|Build AI-powered applications with React, Svelte, and Vue| | 411|netease-youdao/QAnything !2025-03-28129318|Question and Answer based on Anything.| | 412|huggingface/trl !2025-03-281289622|Train transformer language models with reinforcement learning.| | 413|microsoft/BitNet !2025-03-28128503|Official inference framework for 1-bit LLMs| | 414|mediar-ai/screenpipe !2025-03-281283915|24/7 local AI screen & mic recording. Build AI apps that have the full context. Works with Ollama. Alternative to Rewind.ai. Open. Secure. You own your data. Rust.| | 415|Skyvern-AI/skyvern !2025-03-281277612|Automate browser-based workflows with LLMs and Computer Vision| | 416|pytube/pytube !2025-03-28126591|A lightweight, dependency-free Python library (and command-line utility) for downloading YouTube Videos.| | 417|official-stockfish/Stockfish !2025-03-28126574|UCI chess engine| | 418|sgl-project/sglang !2025-03-281260143|SGLang is a structured generation language designed for large language models (LLMs). It makes your interaction with LLMs faster and more controllable.| | 419|plasma-umass/scalene !2025-03-28125535|Scalene: a high-performance, high-precision CPU, GPU, and memory profiler for Python with AI-powered optimization proposals| | 420|danswer-ai/danswer !2025-03-28125503|Ask Questions in natural language and get Answers backed by private sources. Connects to tools like Slack, GitHub, Confluence, etc.| | 421|OpenTalker/SadTalker !2025-03-28125226|[CVPR 2023] SadTalker:Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation| | 422|facebookresearch/AnimatedDrawings !2025-03-28123693 |Code to accompany "A Method for Animating Children's Drawings of the Human Figure"| | 423|activepieces/activepieces !2025-03-28123609|Your friendliest open source all-in-one automation tool ✨ Workflow automation tool 100+ integration / Enterprise automation tool / Zapier Alternative| | 424|ggerganov/ggml !2025-03-28121992 |Tensor library for machine learning| | 425|bytebase/bytebase !2025-03-28121694|World's most advanced database DevOps and CI/CD for Developer, DBA and Platform Engineering teams. The GitLab/GitHub for database DevOps.| | 426| willwulfken/MidJourney-Styles-and-Keywords-Reference !2025-03-28120971 | A reference containing Styles and Keywords that you can use with MidJourney AI| | 427|Huanshere/VideoLingo !2025-03-281207013|Netflix-level subtitle cutting, translation, alignment, and even dubbing - one-click fully automated AI video subtitle team | | 428|OpenLMLab/MOSS !2025-03-28120330 |An open-source tool-augmented conversational language model from Fudan University| | 429|llmware-ai/llmware !2025-03-281200727|Providing enterprise-grade LLM-based development framework, tools, and fine-tuned models.| | 430|PKU-YuanGroup/Open-Sora-Plan !2025-03-28119362|This project aim to reproduce Sora (Open AI T2V model), but we only have limited resource. We deeply wish the all open source community can contribute to this project.| | 431|ShishirPatil/gorilla !2025-03-28119332 |Gorilla: An API store for LLMs| | 432|NVIDIA/Megatron-LM !2025-03-281192716|Ongoing research training transformer models at scale| | 433|illacloud/illa-builder !2025-03-28119192|Create AI-Driven Apps like Assembling Blocks| | 434|marimo-team/marimo !2025-03-281191521|A reactive notebook for Python — run reproducible experiments, execute as a script, deploy as an app, and version with git.| | 435|smol-ai/developer !2025-03-28119111 | With 100k context windows on the way, it's now feasible for every dev to have their own smol developer| | 436|Lightning-AI/litgpt !2025-03-28118878|Pretrain, finetune, deploy 20+ LLMs on your own data. Uses state-of-the-art techniques: flash attention, FSDP, 4-bit, LoRA, and more.| | 437|openai/shap-e !2025-03-28118474 |Generate 3D objects conditioned on text or images| | 438|eugeneyan/open-llms !2025-03-28118451 |A list of open LLMs available for commercial use.| | 439|andrewyng/aisuite !2025-03-28118124|Simple, unified interface to multiple Generative AI providers| | 440|hajimehoshi/ebiten !2025-03-28117816|Ebitengine - A dead simple 2D game engine for Go| | 441|kgrzybek/modular-monolith-with-ddd !2025-03-28117493|Full Modular Monolith application with Domain-Driven Design approach.| | 442|h2oai/h2ogpt !2025-03-2811736-1 |Come join the movement to make the world's best open source GPT led by H2O.ai - 100% private chat and document search, no data leaks, Apache 2.0| | 443|owainlewis/awesome-artificial-intelligence !2025-03-28117332|A curated list of Artificial Intelligence (AI) courses, books, video lectures and papers.| | 444|DataTalksClub/mlops-zoomcamp !2025-03-28116643|Free MLOps course from DataTalks.Club| | 445|Rudrabha/Wav2Lip !2025-03-281163410|This repository contains the codes of "A Lip Sync Expert Is All You Need for Speech to Lip Generation In the Wild", published at ACM Multimedia 2020.| | 446|aishwaryanr/awesome-generative-ai-guide !2025-03-281152810|A one stop repository for generative AI research updates, interview resources, notebooks and much more!| | 447|karpathy/micrograd !2025-03-28115146|A tiny scalar-valued autograd engine and a neural net library on top of it with PyTorch-like API| | 448|InstantID/InstantID !2025-03-28115111|InstantID : Zero-shot Identity-Preserving Generation in Seconds 🔥| | 449|facebookresearch/seamlesscommunication !2025-03-28114434|Foundational Models for State-of-the-Art Speech and Text Translation| | 450|anthropics/anthropic-cookbook !2025-03-281140112|A collection of notebooks/recipes showcasing some fun and effective ways of using Claude.| | 451|mastra-ai/mastra !2025-03-281139240|the TypeScript AI agent framework| | 452|NVIDIA/TensorRT !2025-03-28113864|NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.| | 453|plandex-ai/plandex !2025-03-28113645|An AI coding engine for complex tasks| | 454|RUCAIBox/LLMSurvey !2025-03-28112735 |A collection of papers and resources related to Large Language Models.| | 455|kubeshark/kubeshark !2025-03-28112711|The API traffic analyzer for Kubernetes providing real-time K8s protocol-level visibility, capturing and monitoring all traffic and payloads going in, out and across containers, pods, nodes and clusters. Inspired by Wireshark, purposely built for Kubernetes| | 456|electric-sql/pglite !2025-03-28112617|Lightweight Postgres packaged as WASM into a TypeScript library for the browser, Node.js, Bun and Deno from https://electric-sql.com| | 457|lightaime/camel !2025-03-281124441 |🐫 CAMEL: Communicative Agents for “Mind” Exploration of Large Scale Language Model Society| | 458|huggingface/lerobot !2025-03-281120184|🤗 LeRobot: State-of-the-art Machine Learning for Real-World Robotics in Pytorch| | 459|normal-computing/outlines !2025-03-28111657|Generative Model Programming| | 460|libretro/RetroArch !2025-03-28110701|Cross-platform, sophisticated frontend for the libretro API. Licensed GPLv3.| | 461|THUDM/CogVideo !2025-03-28110599|Text-to-video generation: CogVideoX (2024) and CogVideo (ICLR 2023)| | 462|bentoml/OpenLLM !2025-03-28110495|An open platform for operating large language models (LLMs) in production. Fine-tune, serve, deploy, and monitor any LLMs with ease.| | 463|vosen/ZLUDA !2025-03-28110429|CUDA on AMD GPUs| | 464|dair-ai/ML-Papers-of-the-Week !2025-03-28110304 |🔥Highlighting the top ML papers every week.| | 465|WordPress/gutenberg !2025-03-28110212|The Block Editor project for WordPress and beyond. Plugin is available from the official repository.| | 466|microsoft/data-formulator !2025-03-281099827|🪄 Create rich visualizations with AI| | 467|LibreTranslate/LibreTranslate !2025-03-28109887|Free and Open Source Machine Translation API. Self-hosted, offline capable and easy to setup.| | 468|block/goose !2025-03-281097737|an open-source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM| | 469|getumbrel/llama-gpt !2025-03-28109553|A self-hosted, offline, ChatGPT-like chatbot. Powered by Llama 2. 100% private, with no data leaving your device.| | 470|HigherOrderCO/HVM !2025-03-28109182|A massively parallel, optimal functional runtime in Rust| | 471|databrickslabs/dolly !2025-03-2810812-3 | A large language model trained on the Databricks Machine Learning Platform| | 472|srush/GPU-Puzzles !2025-03-28108014|Solve puzzles. Learn CUDA.| | 473|Z3Prover/z3 !2025-03-28107952|The Z3 Theorem Prover| | 474|UFund-Me/Qbot !2025-03-281079313 |Qbot is an AI-oriented quantitative investment platform, which aims to realize the potential, empower AI technologies in quantitative investment| | 475|langchain-ai/langgraph !2025-03-281077336|| | 476|lz4/lz4 !2025-03-28107647|Extremely Fast Compression algorithm| | 477|magic-research/magic-animate !2025-03-28107160|MagicAnimate: Temporally Consistent Human Image Animation using Diffusion Model| | 478|PaperMC/Paper !2025-03-281071410|The most widely used, high performance Minecraft server that aims to fix gameplay and mechanics inconsistencies| | 479|getomni-ai/zerox !2025-03-281071015|Zero shot pdf OCR with gpt-4o-mini| |!green-up-arrow.svg 480|🔥NirDiamant/GenAIAgents !2025-03-2810693318|This repository provides tutorials and implementations for various Generative AI Agent techniques, from basic to advanced. It serves as a comprehensive guide for building intelligent, interactive AI systems.| |!red-down-arrow 481|Unstructured-IO/unstructured !2025-03-28106889|Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.| | 482|apache/thrift !2025-03-28106610|Apache Thrift| | 483| TheR1D/shellgpt !2025-03-28106097 | A command-line productivity tool powered by ChatGPT, will help you accomplish your tasks faster and more efficiently | | 484|TheRamU/Fay !2025-03-281060312 |Fay is a complete open source project that includes Fay controller and numeral models, which can be used in different applications such as virtual hosts, live promotion, numeral human interaction and so on| | 485|zyronon/douyin !2025-03-28105566|Vue3 + Pinia + Vite5 仿抖音,Vue 在移动端的最佳实践 . Imitate TikTok ,Vue Best practices on Mobile| | 486|THU-MIG/yolov10 !2025-03-28105485|YOLOv10: Real-Time End-to-End Object Detection| | 487|idootop/mi-gpt !2025-03-281052522|? Transform XiaoAi speaker into a personal voice assistant with ChatGPT and DouBao integration.| | 488|SakanaAI/AI-Scientist !2025-03-281051310|The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery 🧑‍🔬| | 489|szimek/sharedrop !2025-03-28105101|Easy P2P file transfer powered by WebRTC - inspired by Apple AirDrop| | 490|salesforce/LAVIS !2025-03-28103942 |LAVIS - A One-stop Library for Language-Vision Intelligence| | 491|aws/amazon-sagemaker-examples !2025-03-28103654|Example 📓 Jupyter notebooks that demonstrate how to build, train, and deploy machine learning models using 🧠 Amazon SageMaker.| | 492|artidoro/qlora !2025-03-28103402 |QLoRA: Efficient Finetuning of Quantized LLMs| | 493|lllyasviel/stable-diffusion-webui-forge !2025-03-281029314| a platform on top of Stable Diffusion WebUI (based on Gradio) to make development easier, optimize resource management, and speed up inference| | 494|NielsRogge/Transformers-Tutorials !2025-03-28102487|This repository contains demos I made with the Transformers library by HuggingFace.| | 495|kedro-org/kedro !2025-03-28102371|Kedro is a toolbox for production-ready data science. It uses software engineering best practices to help you create data engineering and data science pipelines that are reproducible, maintainable, and modular.| | 496| chathub-dev/chathub !2025-03-28102301 | All-in-one chatbot client | | 497|microsoft/promptflow !2025-03-28101612|Build high-quality LLM apps - from prototyping, testing to production deployment and monitoring.| | 498|mistralai/mistral-src !2025-03-28101372|Reference implementation of Mistral AI 7B v0.1 model.| | 499|burn-rs/burn !2025-03-28101183|Burn - A Flexible and Comprehensive Deep Learning Framework in Rust| | 500|AIGC-Audio/AudioGPT !2025-03-28101150 |AudioGPT: Understanding and Generating Speech, Music, Sound, and Talking Head| | 501|facebookresearch/dinov2 !2025-03-281011210 |PyTorch code and models for the DINOv2 self-supervised learning method.| | 502|RockChinQ/LangBot !2025-03-281008455|😎丰富生态、🧩支持扩展、🦄多模态 - 大模型原生即时通信机器人平台 🤖 | | 503|78/xiaozhi-esp32 !2025-03-281008180|Build your own AI friend| | 504|cumulo-autumn/StreamDiffusion !2025-03-28100761|StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation| | 505|DataTalksClub/machine-learning-zoomcamp !2025-03-28100664|The code from the Machine Learning Bookcamp book and a free course based on the book| | 506|nerfstudio-project/nerfstudio !2025-03-28100343|A collaboration friendly studio for NeRFs| | 507|cupy/cupy !2025-03-28100344|NumPy & SciPy for GPU| | 508|NVIDIA/TensorRT-LLM !2025-03-281000823|TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.| | 509|wasp-lang/open-saas !2025-03-2899665|A free, open-source SaaS app starter for React & Node.js with superpowers. Production-ready. Community-driven.| | 510|huggingface/text-generation-inference !2025-03-2899383|Large Language Model Text Generation Inference| | 511|jxnl/instructor !2025-03-2899224|structured outputs for llms| | 512|GoogleCloudPlatform/generative-ai !2025-03-2899086|Sample code and notebooks for Generative AI on Google Cloud| | 513|manticoresoftware/manticoresearch !2025-03-2898799|Easy to use open source fast database for search | | 514|langfuse/langfuse !2025-03-28985134|🪢 Open source LLM engineering platform. Observability, metrics, evals, prompt management, testing, prompt playground, datasets, LLM evaluations -- 🍊YC W23 🤖 integrate via Typescript, Python / Decorators, OpenAI, Langchain, LlamaIndex, Litellm, Instructor, Mistral, Perplexity, Claude, Gemini, Vertex| | 515|keephq/keep !2025-03-2897949|The open-source alert management and AIOps platform| | 516|sashabaranov/go-openai !2025-03-2897843|OpenAI ChatGPT, GPT-3, GPT-4, DALL·E, Whisper API wrapper for Go| | 517|autowarefoundation/autoware !2025-03-2897766|Autoware - the world's leading open-source software project for autonomous driving| | 518|anthropics/courses !2025-03-2897269|Anthropic's educational courses| | 519|popcorn-official/popcorn-desktop !2025-03-2896853|Popcorn Time is a multi-platform, free software BitTorrent client that includes an integrated media player ( Windows / Mac / Linux ) A Butter-Project Fork| | 520|getmaxun/maxun !2025-03-28968515|🔥 Open-source no-code web data extraction platform. Turn websites to APIs and spreadsheets with no-code robots in minutes! [In Beta]| | 521|wandb/wandb !2025-03-2896763|🔥 A tool for visualizing and tracking your machine learning experiments. This repo contains the CLI and Python API.| | 522|karpathy/minbpe !2025-03-2895353|Minimal, clean, code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.| | 523|bigscience-workshop/petals !2025-03-2895142|🌸 Run large language models at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading| | 524|OthersideAI/self-operating-computer !2025-03-2894931|A framework to enable multimodal models to operate a computer.| | 525|mshumer/gpt-prompt-engineer !2025-03-2894911|| | 526| BloopAI/bloop !2025-03-2894710 | A fast code search engine written in Rust| | 527|BlinkDL/ChatRWKV !2025-03-289467-1 |ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.| | 528|timlrx/tailwind-nextjs-starter-blog !2025-03-2894677|This is a Next.js, Tailwind CSS blogging starter template. Comes out of the box configured with the latest technologies to make technical writing a breeze. Easily configurable and customizable. Perfect as a replacement to existing Jekyll and Hugo individual blogs.| | 529|google/benchmark !2025-03-2893634|A microbenchmark support library| | 530|facebookresearch/nougat !2025-03-2893603|Implementation of Nougat Neural Optical Understanding for Academic Documents| | 531|modelscope/facechain !2025-03-2893536|FaceChain is a deep-learning toolchain for generating your Digital-Twin.| | 532|DrewThomasson/ebook2audiobook !2025-03-2893388|Convert ebooks to audiobooks with chapters and metadata using dynamic AI models and voice cloning. Supports 1,107+ languages!| | 533|RayTracing/raytracing.github.io !2025-03-2893035|Main Web Site (Online Books)| | 534|QwenLM/Qwen2.5-VL !2025-03-28930249|Qwen2.5-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.| | 535|WongKinYiu/yolov9 !2025-03-2892201|Implementation of paper - YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information| | 536|alibaba-damo-academy/FunASR !2025-03-28920222|A Fundamental End-to-End Speech Recognition Toolkit and Open Source SOTA Pretrained Models.| | 537|Visualize-ML/Book4Power-of-Matrix !2025-03-2891931|Book4 'Power of Matrix' | | 538|dice2o/BingGPT !2025-03-289185-1 |Desktop application of new Bing's AI-powered chat (Windows, macOS and Linux)| | 539|browserbase/stagehand !2025-03-28917621|An AI web browsing framework focused on simplicity and extensibility.| | 540|FlagOpen/FlagEmbedding !2025-03-28914111|Dense Retrieval and Retrieval-augmented LLMs| | 541|Const-me/Whisper !2025-03-2890979|High-performance GPGPU inference of OpenAI's Whisper automatic speech recognition (ASR) model| | 542|lucidrains/denoising-diffusion-pytorch !2025-03-2890942|Implementation of Denoising Diffusion Probabilistic Model in Pytorch| | 543|Chainlit/chainlit !2025-03-28904422|Build Conversational AI in minutes ⚡️| | 544|togethercomputer/OpenChatKit !2025-03-2890160 |OpenChatKit provides a powerful, open-source base to create both specialized and general purpose chatbots for various applications| | 545|Stability-AI/StableStudio !2025-03-2889631 |Community interface for generative AI| | 546|voicepaw/so-vits-svc-fork !2025-03-2889482 |so-vits-svc fork with realtime support, improved interface and more features.| | 547|pymc-devs/pymc !2025-03-2889413|Bayesian Modeling and Probabilistic Programming in Python| | 548|espnet/espnet !2025-03-2889302|End-to-End Speech Processing Toolkit| | 549|kedacore/keda !2025-03-2888991|KEDA is a Kubernetes-based Event Driven Autoscaling component. It provides event driven scale for any container running in Kubernetes| | 550|open-mmlab/Amphion !2025-03-28886911|Amphion (/æmˈfaɪən/) is a toolkit for Audio, Music, and Speech Generation. Its purpose is to support reproducible research and help junior researchers and engineers get started in the field of audio, music, and speech generation research and development.| | 551|gorse-io/gorse !2025-03-2888451|Gorse open source recommender system engine| | 552|adams549659584/go-proxy-bingai !2025-03-288768-1 |A Microsoft New Bing demo site built with Vue3 and Go, providing a consistent UI experience, supporting ChatGPT prompts, and accessible within China| | 553|open-mmlab/mmsegmentation !2025-03-2887513|OpenMMLab Semantic Segmentation Toolbox and Benchmark.| | 554|bytedance/monolith !2025-03-2887223|ByteDance's Recommendation System| | 555|LouisShark/chatgptsystemprompt !2025-03-2887216|store all agent's system prompt| | 556|brexhq/prompt-engineering !2025-03-2887080 |Tips and tricks for working with Large Language Models like OpenAI's GPT-4.| | 557|erincatto/box2d !2025-03-2886841|Box2D is a 2D physics engine for games| | 558|🔥microsoft/ai-agents-for-beginners !2025-03-288669323|10 Lessons to Get Started Building AI Agents| | 559|nashsu/FreeAskInternet !2025-03-2886102|FreeAskInternet is a completely free, private and locally running search aggregator & answer generate using LLM, without GPU needed. The user can ask a question and the system will make a multi engine search and combine the search result to the ChatGPT3.5 LLM and generate the answer based on search results.| | 560|goldmansachs/gs-quant !2025-03-2885981|Python toolkit for quantitative finance| | 561|srbhr/Resume-Matcher !2025-03-2885800|Open Source Free ATS Tool to compare Resumes with Job Descriptions and create a score to rank them.| | 562|facebookresearch/ImageBind !2025-03-2885681 |ImageBind One Embedding Space to Bind Them All| | 563|ashawkey/stable-dreamfusion !2025-03-2885481 |A pytorch implementation of text-to-3D dreamfusion, powered by stable diffusion.| | 564|meetecho/janus-gateway !2025-03-2885232|Janus WebRTC Server| | 565|google/magika !2025-03-2885003|Detect file content types with deep learning| | 566|huggingface/chat-ui !2025-03-2884871 |Open source codebase powering the HuggingChat app| | 567|EleutherAI/lm-evaluation-harness !2025-03-28843012|A framework for few-shot evaluation of autoregressive language models.| | 568|jina-ai/reader !2025-03-2884089|Convert any URL to an LLM-friendly input with a simple prefix https://r.jina.ai/| | 569|microsoft/TypeChat !2025-03-288406-1|TypeChat is a library that makes it easy to build natural language interfaces using types.| | 570|thuml/Time-Series-Library !2025-03-28839715|A Library for Advanced Deep Time Series Models.| | 571|OptimalScale/LMFlow !2025-03-2883882|An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Model for All.| | 572|baptisteArno/typebot.io !2025-03-2883845|💬 Typebot is a powerful chatbot builder that you can self-host.| | 573|jzhang38/TinyLlama !2025-03-2883504|The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.| | 574|fishaudio/Bert-VITS2 !2025-03-2883472|vits2 backbone with multilingual-bert| | 575|OpenBMB/XAgent !2025-03-2882683|An Autonomous LLM Agent for Complex Task Solving| | 576|Acly/krita-ai-diffusion !2025-03-2882387|Streamlined interface for generating images with AI in Krita. Inpaint and outpaint with optional text prompt, no tweaking required.| | 577|jasonppy/VoiceCraft !2025-03-2882151|Zero-Shot Speech Editing and Text-to-Speech in the Wild| | 578|SJTU-IPADS/PowerInfer !2025-03-2881693|High-speed Large Language Model Serving on PCs with Consumer-grade GPUs| | 579|modelscope/DiffSynth-Studio !2025-03-28814713|Enjoy the magic of Diffusion models!| | 580|o3de/o3de !2025-03-2881443|Open 3D Engine (O3DE) is an Apache 2.0-licensed multi-platform 3D engine that enables developers and content creators to build AAA games, cinema-quality 3D worlds, and high-fidelity simulations without any fees or commercial obligations.| | 581|zmh-program/chatnio !2025-03-2881325|🚀 Next Generation AI One-Stop Internationalization Solution. 🚀 下一代 AI 一站式 B/C 端解决方案,支持 OpenAI,Midjourney,Claude,讯飞星火,Stable Diffusion,DALL·E,ChatGLM,通义千问,腾讯混元,360 智脑,百川 AI,火山方舟,新必应,Gemini,Moonshot 等模型,支持对话分享,自定义预设,云端同步,模型市场,支持弹性计费和订阅计划模式,支持图片解析,支持联网搜索,支持模型缓存,丰富美观的后台管理与仪表盘数据统计。| | 582|leptonai/searchwithlepton !2025-03-2880632|Building a quick conversation-based search demo with Lepton AI.| | 583|sebastianstarke/AI4Animation !2025-03-2880620|Bringing Characters to Life with Computer Brains in Unity| | 584|wangrongding/wechat-bot !2025-03-2880528|🤖一个基于 WeChaty 结合 DeepSeek / ChatGPT / Kimi / 讯飞等Ai服务实现的微信机器人 ,可以用来帮助你自动回复微信消息,或者管理微信群/好友,检测僵尸粉等...| | 585|openvinotoolkit/openvino !2025-03-2880528|OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference| | 586|steven2358/awesome-generative-ai !2025-03-28802610|A curated list of modern Generative Artificial Intelligence projects and services| | 587|adam-maj/tiny-gpu !2025-03-2880234|A minimal GPU design in Verilog to learn how GPUs work from the ground up| | 588| anse-app/chatgpt-demo !2025-03-2880180 | A demo repo based on OpenAI API (gpt-3.5-turbo) | | 589| acheong08/EdgeGPT !2025-03-288015-1 |Reverse engineered API of Microsoft's Bing Chat | | 590|ai-collection/ai-collection !2025-03-2879994 |The Generative AI Landscape - A Collection of Awesome Generative AI Applications| | 591|GreyDGL/PentestGPT !2025-03-2879953 |A GPT-empowered penetration testing tool| | 592|delta-io/delta !2025-03-2879112|An open-source storage framework that enables building a Lakehouse architecture with compute engines including Spark, PrestoDB, Flink, Trino, and Hive and APIs| | 593|dataelement/bisheng !2025-03-2879085|Bisheng is an open LLM devops platform for next generation AI applications.| | 594|e2b-dev/e2b !2025-03-2878447 |Vercel for AI agents. We help developers to build, deploy, and monitor AI agents. Focusing on specialized AI agents that build software for you - your personal software developers.| | 595|01-ai/Yi !2025-03-2878311|A series of large language models trained from scratch by developers @01-ai| | 596|Plachtaa/VALL-E-X !2025-03-287830-1|An open source implementation of Microsoft's VALL-E X zero-shot TTS model. The demo is available at https://plachtaa.github.io| | 597|abhishekkrthakur/approachingalmost !2025-03-2878204|Approaching (Almost) Any Machine Learning Problem| | 598|pydantic/pydantic-ai !2025-03-28781041|Agent Framework / shim to use Pydantic with LLMs| | 599|rany2/edge-tts !2025-03-2877901|Use Microsoft Edge's online text-to-speech service from Python WITHOUT needing Microsoft Edge or Windows or an API key| | 600|CASIA-IVA-Lab/FastSAM !2025-03-2877881|Fast Segment Anything| | 601|netease-youdao/EmotiVoice !2025-03-2877817|EmotiVoice 😊: a Multi-Voice and Prompt-Controlled TTS Engine| | 602|lllyasviel/IC-Light !2025-03-2877804|More relighting!| | 603|kroma-network/tachyon !2025-03-287774-1|Modular ZK(Zero Knowledge) backend accelerated by GPU| | 604|deep-floyd/IF !2025-03-2877731 |A novel state-of-the-art open-source text-to-image model with a high degree of photorealism and language understanding| | 605|oumi-ai/oumi !2025-03-2877705|Everything you need to build state-of-the-art foundation models, end-to-end.| | 606|reorproject/reor !2025-03-2877681|AI note-taking app that runs models locally.| | 607|lightpanda-io/browser !2025-03-28775813|Lightpanda: the headless browser designed for AI and automation| | 608|xiangsx/gpt4free-ts !2025-03-287755-1|Providing a free OpenAI GPT-4 API ! This is a replication project for the typescript version of xtekky/gpt4free| | 609|IDEA-Research/GroundingDINO !2025-03-28773311|Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"| | 610|bunkerity/bunkerweb !2025-03-2877326|🛡️ Make your web services secure by default !| | 611|vikhyat/moondream !2025-03-2877057|tiny vision language model| | 612|firmai/financial-machine-learning !2025-03-287703-1|A curated list of practical financial machine learning tools and applications.| | 613|n8n-io/self-hosted-ai-starter-kit !2025-03-28765121|The Self-hosted AI Starter Kit is an open-source template that quickly sets up a local AI environment. Curated by n8n, it provides essential tools for creating secure, self-hosted AI workflows.| | 614|intel-analytics/ipex-llm !2025-03-2876507|Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with llama.cpp, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, ModelScope, etc.| | 615|jrouwe/JoltPhysics !2025-03-28764510|A multi core friendly rigid body physics and collision detection library. Written in C++. Suitable for games and VR applications. Used by Horizon Forbidden West.| | 616|THUDM/CodeGeeX2 !2025-03-2876270|CodeGeeX2: A More Powerful Multilingual Code Generation Model| | 617|meta-llama/llama-stack !2025-03-2875866|Composable building blocks to build Llama Apps| | 618|sweepai/sweep !2025-03-287530-1|Sweep is an AI junior developer| | 619|lllyasviel/Omost !2025-03-2875301|Your image is almost there!| | 620|ahmedbahaaeldin/From-0-to-Research-Scientist-resources-guide !2025-03-2875050|Detailed and tailored guide for undergraduate students or anybody want to dig deep into the field of AI with solid foundation.| | 621|dair-ai/ML-Papers-Explained !2025-03-2875050|Explanation to key concepts in ML| | 622|zaidmukaddam/scira !2025-03-28750110|Scira (Formerly MiniPerplx) is a minimalistic AI-powered search engine that helps you find information on the internet. Powered by Vercel AI SDK! Search with models like Grok 2.0.| | 623|Portkey-AI/gateway !2025-03-28749416|A Blazing Fast AI Gateway. Route to 100+ LLMs with 1 fast & friendly API.| | 624|web-infra-dev/midscene !2025-03-28748729|An AI-powered automation SDK can control the page, perform assertions, and extract data in JSON format using natural language.| | 625|zilliztech/GPTCache !2025-03-2874801 |GPTCache is a library for creating semantic cache to store responses from LLM queries.| | 626|niedev/RTranslator !2025-03-2874742|RTranslator is the world's first open source real-time translation app.| |!green-up-arrow.svg 627|roboflow/notebooks !2025-03-2874666|Examples and tutorials on using SOTA computer vision models and techniques. Learn everything from old-school ResNet, through YOLO and object-detection transformers like DETR, to the latest models like Grounding DINO and SAM.| |!red-down-arrow 628|openlm-research/openllama !2025-03-2874652|OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset| | 629|LiheYoung/Depth-Anything !2025-03-2874155|Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data| | 630|enso-org/enso !2025-03-2874040|Hybrid visual and textual functional programming.| | 631|bigcode-project/starcoder !2025-03-287401-1 |Home of StarCoder: fine-tuning & inference!| | 632|git-ecosystem/git-credential-manager !2025-03-2873975|Secure, cross-platform Git credential storage with authentication to GitHub, Azure Repos, and other popular Git hosting services.| | 633|OpenGVLab/InternVL !2025-03-2873634|[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4V. 接近GPT-4V表现的可商用开源模型| | 634|WooooDyy/LLM-Agent-Paper-List !2025-03-2873551|The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.| | 635|lencx/Noi !2025-03-2873157|🦄 AI + Tools + Plugins + Community| | 636|udlbook/udlbook !2025-03-2873075|Understanding Deep Learning - Simon J.D. Prince| | 637|OpenBMB/MiniCPM !2025-03-2872841|MiniCPM-2B: An end-side LLM outperforms Llama2-13B.| | 638|jaywalnut310/vits !2025-03-2872815 |VITS: Conditional Variational Autoencoder with Adversarial Learning for End-to-End Text-to-Speech| | 639|xorbitsai/inference !2025-03-28727528|Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.| | 640|PWhiddy/PokemonRedExperiments !2025-03-2872492|Playing Pokemon Red with Reinforcement Learning| | 641|Canner/WrenAI !2025-03-28723213|🤖 Open-source AI Agent that empowers data-driven teams to chat with their data to generate Text-to-SQL, charts, spreadsheets, reports, and BI. 📈📊📋🧑‍💻| | 642|miurla/morphic !2025-03-2872258|An AI-powered answer engine with a generative UI| | 643|ml-explore/mlx-examples !2025-03-2872168|Examples in the MLX framework| | 644|PKU-YuanGroup/ChatLaw !2025-03-2872010|Chinese Legal Large Model| | 645|NVIDIA/cutlass !2025-03-2871883|CUDA Templates for Linear Algebra Subroutines| | 646|FoundationVision/VAR !2025-03-28717444|[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction"| | 647|ymcui/Chinese-LLaMA-Alpaca-2 !2025-03-2871561|Chinese LLaMA-2 & Alpaca-2 LLMs| | 648|nadermx/backgroundremover !2025-03-2871514 |Background Remover lets you Remove Background from images and video using AI with a simple command line interface that is free and open source.| | 649|onuratakan/gpt-computer-assistant !2025-03-28714514|gpt-4o for windows, macos and ubuntu| | 650|graviraja/MLOps-Basics !2025-03-2871326|| | 651|Future-House/paper-qa !2025-03-287118-1|High accuracy RAG for answering questions from scientific documents with citations| | 652|open-mmlab/mmagic !2025-03-2871102 |OpenMMLab Multimodal Advanced, Generative, and Intelligent Creation Toolbox| | 653|bhaskatripathi/pdfGPT !2025-03-2870941 |PDF GPT allows you to chat with the contents of your PDF file by using GPT capabilities. The only open source solution to turn your pdf files in a chatbot!| | 654|ollama/ollama-python !2025-03-28709117|Ollama Python library| | 655|facebookresearch/DiT !2025-03-2870376|Official PyTorch Implementation of "Scalable Diffusion Models with Transformers"| | 656|geekyutao/Inpaint-Anything !2025-03-2870262 |Inpaint anything using Segment Anything and inpainting models.| | 657|AbdullahAlfaraj/Auto-Photoshop-StableDiffusion-Plugin !2025-03-2870160 |A user-friendly plug-in that makes it easy to generate stable diffusion images inside Photoshop using Automatic1111-sd-webui as a backend.| | 658|apple/corenet !2025-03-2869990|CoreNet: A library for training deep neural networks| | 659|openstatusHQ/openstatus !2025-03-2869926|🏓 The open-source synthetic monitoring platform 🏓| | 660|weaviate/Verba !2025-03-2869772|Retrieval Augmented Generation (RAG) chatbot powered by Weaviate| | 661|meshery/meshery !2025-03-2869630|Meshery, the cloud native manager| | 662|OpenTalker/video-retalking !2025-03-2869530|[SIGGRAPH Asia 2022] VideoReTalking: Audio-based Lip Synchronization for Talking Head Video Editing In the Wild| | 663|digitalinnovationone/dio-lab-open-source !2025-03-28689013|Repositório do lab "Contribuindo em um Projeto Open Source no GitHub" da Digital Innovation One.| | 664|jianchang512/ChatTTS-ui !2025-03-2868842|一个简单的本地网页界面,直接使用ChatTTS将文字合成为语音,同时支持对外提供API接口。| | 665|patchy631/ai-engineering-hub !2025-03-28686434|In-depth tutorials on LLMs, RAGs and real-world AI agent applications.| | 666|gunnarmorling/1brc !2025-03-2868512|1️⃣🐝🏎️ The One Billion Row Challenge -- A fun exploration of how quickly 1B rows from a text file can be aggregated with Java| | 667|Azure-Samples/azure-search-openai-demo !2025-03-2868482 |A sample app for the Retrieval-Augmented Generation pattern running in Azure, using Azure Cognitive Search for retrieval and Azure OpenAI large language models to power ChatGPT-style and Q&A experiences.| | 668|mit-han-lab/streaming-llm !2025-03-2868382|Efficient Streaming Language Models with Attention Sinks| | 669|InternLM/InternLM !2025-03-2868352|InternLM has open-sourced a 7 billion parameter base model, a chat model tailored for practical scenarios and the training system.| | 670|dependency-check/DependencyCheck !2025-03-2868191|OWASP dependency-check is a software composition analysis utility that detects publicly disclosed vulnerabilities in application dependencies.| | 671|Soulter/AstrBot !2025-03-28678643|✨易上手的多平台 LLM 聊天机器人及开发框架✨。支持 QQ、QQ频道、Telegram、微信平台(Gewechat, 企业微信)、内置 Web Chat,OpenAI GPT、DeepSeek、Ollama、Llama、GLM、Gemini、OneAPI、LLMTuner,支持 LLM Agent 插件开发,可视化面板。一键部署。支持 Dify 工作流、代码执行器、Whisper 语音转文字。| | 672|react-native-webview/react-native-webview !2025-03-2867792|React Native Cross-Platform WebView| | 673|modelscope/agentscope !2025-03-28676916|Start building LLM-empowered multi-agent applications in an easier way.| | 674|mylxsw/aidea !2025-03-2867381|AIdea is a versatile app that supports GPT and domestic large language models,also supports "Stable Diffusion" text-to-image generation, image-to-image generation, SDXL 1.0, super-resolution, and image colorization| | 675|langchain-ai/ollama-deep-researcher !2025-03-28668635|Fully local web research and report writing assistant| | 676|threestudio-project/threestudio !2025-03-2866653|A unified framework for 3D content generation.| | 677|gaomingqi/Track-Anything !2025-03-2866631 |A flexible and interactive tool for video object tracking and segmentation, based on Segment Anything, XMem, and E2FGVI.| | 678|spdustin/ChatGPT-AutoExpert !2025-03-2866570|🚀🧠💬 Supercharged Custom Instructions for ChatGPT (non-coding) and ChatGPT Advanced Data Analysis (coding).| | 679|HariSekhon/DevOps-Bash-tools !2025-03-2866463|1000+ DevOps Bash Scripts - AWS, GCP, Kubernetes, Docker, CI/CD, APIs, SQL, PostgreSQL, MySQL, Hive, Impala, Kafka, Hadoop, Jenkins, GitHub, GitLab, BitBucket, Azure DevOps, TeamCity, Spotify, MP3, LDAP, Code/Build Linting, pkg mgmt for Linux, Mac, Python, Perl, Ruby, NodeJS, Golang, Advanced dotfiles: .bashrc, .vimrc, .gitconfig, .screenrc, tmux..| | 680|modelscope/swift !2025-03-28661530|ms-swift: Use PEFT or Full-parameter to finetune 200+ LLMs or 15+ MLLMs| | 681|langchain-ai/opengpts !2025-03-2866080|This is an open source effort to create a similar experience to OpenAI's GPTs and Assistants API| | 682| yihong0618/xiaogpt !2025-03-2865131 | Play ChatGPT with xiaomi ai speaker | | 683| civitai/civitai !2025-03-2865111 | Build a platform where people can share their stable diffusion models | | 684|KoljaB/RealtimeSTT !2025-03-28649513|A robust, efficient, low-latency speech-to-text library with advanced voice activity detection, wake word activation and instant transcription.| | 685|qunash/chatgpt-advanced !2025-03-2864910 | A browser extension that augments your ChatGPT prompts with web results.| | 686|Licoy/ChatGPT-Midjourney !2025-03-2864850|🎨 Own your own ChatGPT+Midjourney web service with one click| | 687|friuns2/BlackFriday-GPTs-Prompts !2025-03-2864744|List of free GPTs that doesn't require plus subscription| | 688|PixarAnimationStudios/OpenUSD !2025-03-2864700|Universal Scene Description| | 689|linyiLYi/street-fighter-ai !2025-03-2864630 |This is an AI agent for Street Fighter II Champion Edition.| | 690|run-llama/rags !2025-03-2864380|Build ChatGPT over your data, all with natural language| | 691|frdel/agent-zero !2025-03-2864154|Agent Zero AI framework| | 692|microsoft/DeepSpeedExamples !2025-03-2863911 |Example models using DeepSpeed| | 693|k8sgpt-ai/k8sgpt !2025-03-2863882|Giving Kubernetes Superpowers to everyone| | 694|open-metadata/OpenMetadata !2025-03-2863514|OpenMetadata is a unified platform for discovery, observability, and governance powered by a central metadata repository, in-depth lineage, and seamless team collaboration.| | 695|google/gemma.cpp !2025-03-2863163|lightweight, standalone C++ inference engine for Google's Gemma models.| | 696|RayVentura/ShortGPT !2025-03-286314-1|🚀🎬 ShortGPT - An experimental AI framework for automated short/video content creation. Enables creators to rapidly produce, manage, and deliver content using AI and automation.| | 697|openai/consistencymodels !2025-03-2862940 |Official repo for consistency models.| | 698|yangjianxin1/Firefly !2025-03-2862924|Firefly: Chinese conversational large language model (full-scale fine-tuning + QLoRA), supporting fine-tuning of Llma2, Llama, Baichuan, InternLM, Ziya, Bloom, and other large models| | 699|enricoros/big-AGI !2025-03-2862665|Generative AI suite powered by state-of-the-art models and providing advanced AI/AGI functions. It features AI personas, AGI functions, multi-model chats, text-to-image, voice, response streaming, code highlighting and execution, PDF import, presets for developers, much more. Deploy on-prem or in the cloud.| | 700|aptos-labs/aptos-core !2025-03-2862633|Aptos is a layer 1 blockchain built to support the widespread use of blockchain through better technology and user experience.| | 701|wenda-LLM/wenda !2025-03-286262-1 |Wenda: An LLM invocation platform. Its objective is to achieve efficient content generation tailored to specific environments while considering the limited computing resources of individuals and small businesses, as well as knowledge security and privacy concerns| | 702|Project-MONAI/MONAI !2025-03-2862603|AI Toolkit for Healthcare Imaging| | 703|HVision-NKU/StoryDiffusion !2025-03-2862470|Create Magic Story!| | 704|deepseek-ai/DeepSeek-LLM !2025-03-2862463|DeepSeek LLM: Let there be answers| | 705|Tohrusky/Final2x !2025-03-2862393|2^x Image Super-Resolution| | 706|OpenSPG/KAG !2025-03-28619611|KAG is a logical form-guided reasoning and retrieval framework based on OpenSPG engine and LLMs. It is used to build logical reasoning and factual Q&A solutions for professional domain knowledge bases. It can effectively overcome the shortcomings of the traditional RAG vector similarity calculation model.| | 707|Moonvy/OpenPromptStudio !2025-03-2861861 |AIGC Hint Word Visualization Editor| | 708|levihsu/OOTDiffusion !2025-03-2861761|Official implementation of OOTDiffusion| | 709|tmc/langchaingo !2025-03-2861729|LangChain for Go, the easiest way to write LLM-based programs in Go| | 710|vladmandic/automatic !2025-03-2861374|SD.Next: Advanced Implementation of Stable Diffusion and other Diffusion-based generative image models| | 711|clovaai/donut !2025-03-2861231 |Official Implementation of OCR-free Document Understanding Transformer (Donut) and Synthetic Document Generator (SynthDoG), ECCV 2022| | 712|Shaunwei/RealChar !2025-03-286121-1|🎙️🤖Create, Customize and Talk to your AI Character/Companion in Realtime(All in One Codebase!). Have a natural seamless conversation with AI everywhere(mobile, web and terminal) using LLM OpenAI GPT3.5/4, Anthropic Claude2, Chroma Vector DB, Whisper Speech2Text, ElevenLabs Text2Speech🎙️🤖| | 713|microsoft/TinyTroupe !2025-03-2861142|LLM-powered multiagent persona simulation for imagination enhancement and business insights.| | 714| rustformers/llm !2025-03-2861010 | Run inference for Large Language Models on CPU, with Rust| | 715|firebase/firebase-ios-sdk !2025-03-2860950|Firebase SDK for Apple App Development| | 716|vespa-engine/vespa !2025-03-2860824|The open big data serving engine. https://vespa.ai| | 717|n4ze3m/page-assist !2025-03-28607610|Use your locally running AI models to assist you in your web browsing| | 718|Dooy/chatgpt-web-midjourney-proxy !2025-03-2860646|chatgpt web, midjourney, gpts,tts, whisper 一套ui全搞定| | 719|ethereum-optimism/optimism !2025-03-2860213|Optimism is Ethereum, scaled.| | 720|sczhou/ProPainter !2025-03-2859971|[ICCV 2023] ProPainter: Improving Propagation and Transformer for Video Inpainting| | 721|MineDojo/Voyager !2025-03-2859951 |An Open-Ended Embodied Agent with Large Language Models| | 722|lavague-ai/LaVague !2025-03-2859800|Automate automation with Large Action Model framework| | 723|SevaSk/ecoute !2025-03-2859770 |Ecoute is a live transcription tool that provides real-time transcripts for both the user's microphone input (You) and the user's speakers output (Speaker) in a textbox. It also generates a suggested response using OpenAI's GPT-3.5 for the user to say based on the live transcription of the conversation.| | 724|google/mesop !2025-03-2859661|| | 725|pengxiao-song/LaWGPT !2025-03-2859542 |Repo for LaWGPT, Chinese-Llama tuned with Chinese Legal knowledge| | 726|fr0gger/Awesome-GPT-Agents !2025-03-2859434|A curated list of GPT agents for cybersecurity| | 727|google-deepmind/graphcast !2025-03-2859412|| | 728|comet-ml/opik !2025-03-28594126|Open-source end-to-end LLM Development Platform| | 729|SciPhi-AI/R2R !2025-03-28594033|A framework for rapid development and deployment of production-ready RAG systems| | 730|SkalskiP/courses !2025-03-2859272 |This repository is a curated collection of links to various courses and resources about Artificial Intelligence (AI)| | 731|QuivrHQ/MegaParse !2025-03-2859122|File Parser optimised for LLM Ingestion with no loss 🧠 Parse PDFs, Docx, PPTx in a format that is ideal for LLMs.| | 732|pytorch-labs/gpt-fast !2025-03-2858971|Simple and efficient pytorch-native transformer text generation in !2025-03-2858886|Curated list of chatgpt prompts from the top-rated GPTs in the GPTs Store. Prompt Engineering, prompt attack & prompt protect. Advanced Prompt Engineering papers.| | 734|nilsherzig/LLocalSearch !2025-03-2858852|LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a chain of LLMs to find the answer. The user can see the progress of the agents and the final answer. No OpenAI or Google API keys are needed.| | 735|kuafuai/DevOpsGPT !2025-03-285874-2|Multi agent system for AI-driven software development. Convert natural language requirements into working software. Supports any development language and extends the existing base code.| | 736|myshell-ai/MeloTTS !2025-03-2858486|High-quality multi-lingual text-to-speech library by MyShell.ai. Support English, Spanish, French, Chinese, Japanese and Korean.| | 737|OpenGVLab/LLaMA-Adapter !2025-03-2858421 |Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters| | 738|volcengine/verl !2025-03-28582563|veRL: Volcano Engine Reinforcement Learning for LLM| | 739|a16z-infra/companion-app !2025-03-2858171|AI companions with memory: a lightweight stack to create and host your own AI companions| | 740|HumanAIGC/OutfitAnyone !2025-03-285816-1|Outfit Anyone: Ultra-high quality virtual try-on for Any Clothing and Any Person| | 741|josStorer/RWKV-Runner !2025-03-2857472|A RWKV management and startup tool, full automation, only 8MB. And provides an interface compatible with the OpenAI API. RWKV is a large language model that is fully open source and available for commercial use.| | 742|648540858/wvp-GB28181-pro !2025-03-2857414|WEB VIDEO PLATFORM是一个基于GB28181-2016标准实现的网络视频平台,支持NAT穿透,支持海康、大华、宇视等品牌的IPC、NVR、DVR接入。支持国标级联,支持rtsp/rtmp等视频流转发到国标平台,支持rtsp/rtmp等推流转发到国标平台。| | 743|ToonCrafter/ToonCrafter !2025-03-2857345|a research paper for generative cartoon interpolation| | 744|PawanOsman/ChatGPT !2025-03-2857191|OpenAI API Free Reverse Proxy| | 745|apache/hudi !2025-03-2857091|Upserts, Deletes And Incremental Processing on Big Data.| | 746| nsarrazin/serge !2025-03-2857081 | A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API| | 747|homanp/superagent !2025-03-2857021|🥷 Superagent - Build, deploy, and manage LLM-powered agents| | 748|ramonvc/freegpt-webui !2025-03-2856910|GPT 3.5/4 with a Chat Web UI. No API key is required.| | 749|baichuan-inc/baichuan-7B !2025-03-2856901|A large-scale 7B pretraining language model developed by BaiChuan-Inc.| | 750|Azure/azure-sdk-for-net !2025-03-2856792|This repository is for active development of the Azure SDK for .NET. For consumers of the SDK we recommend visiting our public developer docs at https://learn.microsoft.com/dotnet/azure/ or our versioned developer docs at https://azure.github.io/azure-sdk-for-net.| | 751|mnotgod96/AppAgent !2025-03-2856643|AppAgent: Multimodal Agents as Smartphone Users, an LLM-based multimodal agent framework designed to operate smartphone apps.| | 752|microsoft/TaskWeaver !2025-03-2856243|A code-first agent framework for seamlessly planning and executing data analytics tasks.| | 753| yetone/bob-plugin-openai-translator !2025-03-285600-1 | A Bob Plugin base ChatGPT API | | 754|PrefectHQ/marvin !2025-03-2855840 |A batteries-included library for building AI-powered software| | 755|microsoft/promptbase !2025-03-2855832|All things prompt engineering| | 756|fullstackhero/dotnet-starter-kit !2025-03-2855560|Production Grade Cloud-Ready .NET 8 Starter Kit (Web API + Blazor Client) with Multitenancy Support, and Clean/Modular Architecture that saves roughly 200+ Development Hours! All Batteries Included.| | 757|deepseek-ai/DeepSeek-Coder-V2 !2025-03-2855435|DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence| | 758|aiwaves-cn/agents !2025-03-2855391|An Open-source Framework for Autonomous Language Agents| | 759|microsoft/Mastering-GitHub-Copilot-for-Paired-Programming !2025-03-2855158|A 6 Lesson course teaching everything you need to know about harnessing GitHub Copilot and an AI Paired Programing resource.| | 760|allenai/OLMo !2025-03-2854506|Modeling, training, eval, and inference code for OLMo| | 761|apify/crawlee-python !2025-03-2854493|Crawlee—A web scraping and browser automation library for Python to build reliable crawlers. Extract data for AI, LLMs, RAG, or GPTs. Download HTML, PDF, JPG, PNG, and other files from websites. Works with BeautifulSoup, Playwright, and raw HTTP. Both headful and headless mode. With proxy rotation.| | 762|k2-fsa/sherpa-onnx !2025-03-28541520|Speech-to-text, text-to-speech, and speaker recongition using next-gen Kaldi with onnxruntime without Internet connection. Support embedded systems, Android, iOS, Raspberry Pi, RISC-V, x86_64 servers, websocket server/client, C/C++, Python, Kotlin, C#, Go, NodeJS, Java, Swift| | 763|TEN-framework/TEN-Agent !2025-03-28541411|TEN Agent is a realtime conversational AI agent powered by TEN. It seamlessly integrates the OpenAI Realtime API, RTC capabilities, and advanced features like weather updates, web search, computer vision, and Retrieval-Augmented Generation (RAG).| | 764|google/gemmapytorch !2025-03-2854010|The official PyTorch implementation of Google's Gemma models| | 765|snakers4/silero-vad !2025-03-2853858|Silero VAD: pre-trained enterprise-grade Voice Activity Detector| | 766|livekit/agents !2025-03-2853836|Build real-time multimodal AI applications 🤖🎙️📹| | 767|pipecat-ai/pipecat !2025-03-28537811|Open Source framework for voice and multimodal conversational AI| | 768|EricLBuehler/mistral.rs !2025-03-28536324|Blazingly fast LLM inference.| | 769|asg017/sqlite-vec !2025-03-28535810|Work-in-progress vector search SQLite extension that runs anywhere.| | 770|albertan017/LLM4Decompile !2025-03-2853563|Reverse Engineering: Decompiling Binary Code with Large Language Models| | 771|Permify/permify !2025-03-2853235|An open-source authorization as a service inspired by Google Zanzibar, designed to build and manage fine-grained and scalable authorization systems for any application.| | 772|imoneoi/openchat !2025-03-2853171|OpenChat: Advancing Open-source Language Models with Imperfect Data| | 773|mosaicml/composer !2025-03-2853140|Train neural networks up to 7x faster| | 774|dsdanielpark/Bard-API !2025-03-285277-1 |The python package that returns a response of Google Bard through API.| | 775|lxfater/inpaint-web !2025-03-2852552|A free and open-source inpainting & image-upscaling tool powered by webgpu and wasm on the browser。| | 776|leanprover/lean4 !2025-03-2852441|Lean 4 programming language and theorem prover| | 777|AILab-CVC/YOLO-World !2025-03-2852415|Real-Time Open-Vocabulary Object Detection| | 778|openchatai/OpenChat !2025-03-2852260 |Run and create custom ChatGPT-like bots with OpenChat, embed and share these bots anywhere, the open-source chatbot console.| | 779|mufeedvh/code2prompt !2025-03-28519414|A CLI tool to convert your codebase into a single LLM prompt with source tree, prompt templating, and token counting.| | 780|biobootloader/wolverine !2025-03-2851700 |Automatically repair python scripts through GPT-4 to give them regenerative abilities.| | 781|huggingface/parler-tts !2025-03-2851671|Inference and training library for high-quality TTS models.| | 782|Akegarasu/lora-scripts !2025-03-2851308 |LoRA training scripts use kohya-ss's trainer, for diffusion model.| | 783|openchatai/OpenCopilot !2025-03-285128-3|🤖 🔥 Let your users chat with your product features and execute things by text - open source Shopify sidekick| | 784|e2b-dev/fragments !2025-03-2851228|Open-source Next.js template for building apps that are fully generated by AI. By E2B.| | 785|microsoft/SynapseML !2025-03-2851132|Simple and Distributed Machine Learning| | 786|aigc-apps/sd-webui-EasyPhoto !2025-03-285108-1|📷 EasyPhoto | | 787|ChaoningZhang/MobileSAM !2025-03-2850944|This is the official code for Faster Segment Anything (MobileSAM) project that makes SAM lightweight| | 788|huggingface/alignment-handbook !2025-03-2850932|Robust recipes for to align language models with human and AI preferences| | 789|alpkeskin/mosint !2025-03-2850920|An automated e-mail OSINT tool| | 790|TaskingAI/TaskingAI !2025-03-2850891|The open source platform for AI-native application development.| | 791|lipku/metahuman-stream !2025-03-28507615|Real time interactive streaming digital human| | 792|OpenInterpreter/01 !2025-03-2850530|The open-source language model computer| | 793|open-compass/opencompass !2025-03-28505111|OpenCompass is an LLM evaluation platform, supporting a wide range of models (InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.| | 794|xxlong0/Wonder3D !2025-03-2850491|A cross-domain diffusion model for 3D reconstruction from a single image| | 795|pytorch/torchtune !2025-03-2850342|A Native-PyTorch Library for LLM Fine-tuning| | 796|SuperDuperDB/superduperdb !2025-03-2850192|🔮 SuperDuperDB: Bring AI to your database: Integrate, train and manage any AI models and APIs directly with your database and your data.| | 797|WhiskeySockets/Baileys !2025-03-2850057|Lightweight full-featured typescript/javascript WhatsApp Web API| | 798| mpociot/chatgpt-vscode !2025-03-2849890 | A VSCode extension that allows you to use ChatGPT | | 799|OpenGVLab/DragGAN !2025-03-2849880|Unofficial Implementation of DragGAN - "Drag Your GAN: Interactive Point-based Manipulation on the Generative Image Manifold" (DragGAN 全功能实现,在线Demo,本地部署试用,代码、模型已全部开源,支持Windows, macOS, Linux)| | 800|microsoft/LLMLingua !2025-03-2849824|To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.| | 801|Zipstack/unstract !2025-03-2849745|No-code LLM Platform to launch APIs and ETL Pipelines to structure unstructured documents| | 802|OpenBMB/ToolBench !2025-03-2849621|An open platform for training, serving, and evaluating large language model for tool learning.| | 803|Fanghua-Yu/SUPIR !2025-03-2849593|SUPIR aims at developing Practical Algorithms for Photo-Realistic Image Restoration In the Wild| | 804|GaiaNet-AI/gaianet-node !2025-03-2849360|Install and run your own AI agent service| | 805|qodo-ai/qodo-cover !2025-03-284922-1|Qodo-Cover: An AI-Powered Tool for Automated Test Generation and Code Coverage Enhancement! 💻🤖🧪🐞| | 806|Zejun-Yang/AniPortrait !2025-03-2849042|AniPortrait: Audio-Driven Synthesis of Photorealistic Portrait Animation| | 807|lvwzhen/law-cn-ai !2025-03-2848901 |⚖️ AI Legal Assistant| | 808|developersdigest/llm-answer-engine !2025-03-2848740|Build a Perplexity-Inspired Answer Engine Using Next.js, Groq, Mixtral, Langchain, OpenAI, Brave & Serper| | 809|Plachtaa/VITS-fast-fine-tuning !2025-03-2848640|This repo is a pipeline of VITS finetuning for fast speaker adaptation TTS, and many-to-many voice conversion| | 810|espeak-ng/espeak-ng !2025-03-2848601|eSpeak NG is an open source speech synthesizer that supports more than hundred languages and accents.| | 811|ant-research/CoDeF !2025-03-2848581|[CVPR'24 Highlight] Official PyTorch implementation of CoDeF: Content Deformation Fields for Temporally Consistent Video Processing| | 812|deepseek-ai/DeepSeek-V2 !2025-03-2848512|| | 813|XRPLF/rippled !2025-03-2848210|Decentralized cryptocurrency blockchain daemon implementing the XRP Ledger protocol in C++| | 814|AutoMQ/automq !2025-03-28478721|AutoMQ is a cloud-first alternative to Kafka by decoupling durability to S3 and EBS. 10x cost-effective. Autoscale in seconds. Single-digit ms latency.| | 815|AILab-CVC/VideoCrafter !2025-03-2847800|VideoCrafter1: Open Diffusion Models for High-Quality Video Generation| | 816|nautechsystems/nautilustrader !2025-03-2847702|A high-performance algorithmic trading platform and event-driven backtester| | 817|kyegomez/swarms !2025-03-2847563|The Enterprise-Grade Production-Ready Multi-Agent Orchestration Framework Join our Community: https://discord.com/servers/agora-999382051935506503| | 818|Deci-AI/super-gradients !2025-03-2847310 |Easily train or fine-tune SOTA computer vision models with one open source training library. The home of Yolo-NAS.| | 819|QwenLM/Qwen2.5-Coder !2025-03-2847236|Qwen2.5-Coder is the code version of Qwen2.5, the large language model series developed by Qwen team, Alibaba Cloud.| | 820|SCIR-HI/Huatuo-Llama-Med-Chinese !2025-03-2847191 |Repo for HuaTuo (华驼), Llama-7B tuned with Chinese medical knowledge| | 821|togethercomputer/RedPajama-Data !2025-03-2846841 |code for preparing large datasets for training large language models| | 822|mishushakov/llm-scraper !2025-03-2846704|Turn any webpage into structured data using LLMs| | 823|1rgs/jsonformer !2025-03-2846663 |A Bulletproof Way to Generate Structured JSON from Language Models| | 824|anti-work/shortest !2025-03-2846565|QA via natural language AI tests| | 825|dnhkng/GlaDOS !2025-03-2846510|This is the Personality Core for GLaDOS, the first steps towards a real-life implementation of the AI from the Portal series by Valve.| | 826|Nukem9/dlssg-to-fsr3 !2025-03-2846380|Adds AMD FSR3 Frame Generation to games by replacing Nvidia DLSS-G Frame Generation (nvngx_dlssg).| | 827|BuilderIO/ai-shell !2025-03-2846373 |A CLI that converts natural language to shell commands.| | 828|facebookincubator/AITemplate !2025-03-2846220 |AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.| | 829|terraform-aws-modules/terraform-aws-eks !2025-03-2846030|Terraform module to create AWS Elastic Kubernetes (EKS) resources 🇺🇦| | 830|timescale/pgai !2025-03-2845915|A suite of tools to develop RAG, semantic search, and other AI applications more easily with PostgreSQL| | 831|awslabs/multi-agent-orchestrator !2025-03-2845788|Flexible and powerful framework for managing multiple AI agents and handling complex conversations| | 832|sanchit-gandhi/whisper-jax !2025-03-2845771 |Optimised JAX code for OpenAI's Whisper Model, largely built on the Hugging Face Transformers Whisper implementation| | 833|NVIDIA/NeMo-Guardrails !2025-03-2845755|NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.| | 834|PathOfBuildingCommunity/PathOfBuilding !2025-03-2845480|Offline build planner for Path of Exile.| | 835|UX-Decoder/Segment-Everything-Everywhere-All-At-Once !2025-03-2845412 |Official implementation of the paper "Segment Everything Everywhere All at Once"| | 836|build-trust/ockam !2025-03-2845171|Orchestrate end-to-end encryption, cryptographic identities, mutual authentication, and authorization policies between distributed applications – at massive scale.| | 837|google-research/timesfm !2025-03-2845135|TimesFM (Time Series Foundation Model) is a pretrained time-series foundation model developed by Google Research for time-series forecasting.| | 838|luosiallen/latent-consistency-model !2025-03-2844842|Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference| | 839|NVlabs/neuralangelo !2025-03-2844740|Official implementation of "Neuralangelo: High-Fidelity Neural Surface Reconstruction" (CVPR 2023)| | 840|kyegomez/tree-of-thoughts !2025-03-2844720 |Plug in and Play Implementation of Tree of Thoughts: Deliberate Problem Solving with Large Language Models that Elevates Model Reasoning by atleast 70%| | 841|sjvasquez/handwriting-synthesis !2025-03-2844720 |Handwriting Synthesis with RNNs ✏️| | 842| madawei2699/myGPTReader !2025-03-2844420 | A slack bot that can read any webpage, ebook or document and summarize it with chatGPT | | 843|OpenBMB/AgentVerse !2025-03-2844413|🤖 AgentVerse 🪐 provides a flexible framework that simplifies the process of building custom multi-agent environments for large language models (LLMs).| | 844|argmaxinc/WhisperKit !2025-03-2844395|Swift native speech recognition on-device for iOS and macOS applications.| | 845|landing-ai/vision-agent !2025-03-2844346|Vision agent| | 846|InternLM/xtuner !2025-03-2844273|An efficient, flexible and full-featured toolkit for fine-tuning large models (InternLM, Llama, Baichuan, Qwen, ChatGLM)| | 847|google-deepmind/alphageometry !2025-03-284421-1|Solving Olympiad Geometry without Human Demonstrations| | 848|ostris/ai-toolkit !2025-03-2844093|Various AI scripts. Mostly Stable Diffusion stuff.| | 849|LLM-Red-Team/kimi-free-api !2025-03-2844004|🚀 KIMI AI 长文本大模型白嫖服务,支持高速流式输出、联网搜索、长文档解读、图像解析、多轮对话,零配置部署,多路token支持,自动清理会话痕迹。| | 850|argilla-io/argilla !2025-03-2843991|Argilla is a collaboration platform for AI engineers and domain experts that require high-quality outputs, full data ownership, and overall efficiency.| | 851|spring-projects/spring-ai !2025-03-28438419|An Application Framework for AI Engineering| | 852|alibaba-damo-academy/FunClip !2025-03-2843555|Open-source, accurate and easy-to-use video clipping tool, LLM based AI clipping intergrated | | 853|yisol/IDM-VTON !2025-03-2843541|IDM-VTON : Improving Diffusion Models for Authentic Virtual Try-on in the Wild| | 854|fchollet/ARC-AGI !2025-03-2843368|The Abstraction and Reasoning Corpus| | 855|MahmoudAshraf97/whisper-diarization !2025-03-2843064|Automatic Speech Recognition with Speaker Diarization based on OpenAI Whisper| | 856|Speykious/cve-rs !2025-03-2843047|Blazingly 🔥 fast 🚀 memory vulnerabilities, written in 100% safe Rust. 🦀| | 857|Blealtan/efficient-kan !2025-03-2842770|An efficient pure-PyTorch implementation of Kolmogorov-Arnold Network (KAN).| | 858|smol-ai/GodMode !2025-03-284249-1|AI Chat Browser: Fast, Full webapp access to ChatGPT / Claude / Bard / Bing / Llama2! I use this 20 times a day.| | 859|openai/plugins-quickstart !2025-03-284235-4 |Get a ChatGPT plugin up and running in under 5 minutes!| | 860|Doriandarko/maestro !2025-03-2842260|A framework for Claude Opus to intelligently orchestrate subagents.| | 861|philz1337x/clarity-upscaler !2025-03-2842204|Clarity-Upscaler: Reimagined image upscaling for everyone| | 862|facebookresearch/co-tracker !2025-03-2842142|CoTracker is a model for tracking any point (pixel) on a video.| | 863|xlang-ai/OpenAgents !2025-03-2842031|OpenAgents: An Open Platform for Language Agents in the Wild| | 864|alibaba/higress !2025-03-28419514|🤖 AI Gateway | | 865|ray-project/llm-numbers !2025-03-2841920 |Numbers every LLM developer should know| | 866|fudan-generative-vision/champ !2025-03-2841820|Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance| | 867|NVIDIA/garak !2025-03-2841795|the LLM vulnerability scanner| | 868|leetcode-mafia/cheetah !2025-03-2841740 |Whisper & GPT-based app for passing remote SWE interviews| | 869|ragapp/ragapp !2025-03-2841710|The easiest way to use Agentic RAG in any enterprise| | 870|collabora/WhisperSpeech !2025-03-2841692|An Open Source text-to-speech system built by inverting Whisper.| | 871|Facico/Chinese-Vicuna !2025-03-2841520 |Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model| | 872|openai/grok !2025-03-2841381|| | 873|CrazyBoyM/llama3-Chinese-chat !2025-03-2841361|Llama3 Chinese Repository with modified versions, and training and deployment resources| | 874|luban-agi/Awesome-AIGC-Tutorials !2025-03-2841301|Curated tutorials and resources for Large Language Models, AI Painting, and more.| | 875|damo-vilab/AnyDoor !2025-03-2841192|Official implementations for paper: Anydoor: zero-shot object-level image customization| | 876|raspberrypi/pico-sdk !2025-03-2841072|| | 877|mshumer/gpt-llm-trainer !2025-03-284097-1|| | 878|metavoiceio/metavoice-src !2025-03-284076-1|AI for human-level speech intelligence| | 879|intelowlproject/IntelOwl !2025-03-2840763|IntelOwl: manage your Threat Intelligence at scale| | 880|a16z-infra/ai-getting-started !2025-03-2840682|A Javascript AI getting started stack for weekend projects, including image/text models, vector stores, auth, and deployment configs| | 881|MarkFzp/mobile-aloha !2025-03-2840641|Mobile ALOHA: Learning Bimanual Mobile Manipulation with Low-Cost Whole-Body Teleoperation| | 882| keijiro/AICommand !2025-03-2840380 | ChatGPT integration with Unity Editor | | 883|Tencent/HunyuanDiT !2025-03-2840214|Hunyuan-DiT : A Powerful Multi-Resolution Diffusion Transformer with Fine-Grained Chinese Understanding| | 884|hengyoush/kyanos !2025-03-2840061|Visualize the time packets spend in the kernel, watch & analyze in command line.| | 885|agiresearch/AIOS !2025-03-2840045|AIOS: LLM Agent Operating System| | 886|truefoundry/cognita !2025-03-2839773|RAG (Retrieval Augmented Generation) Framework for building modular, open source applications for production by TrueFoundry| | 887|X-PLUG/MobileAgent !2025-03-2839557|Mobile-Agent: Autonomous Multi-Modal Mobile Device Agent with Visual Perception| | 888|jackMort/ChatGPT.nvim !2025-03-2839231|ChatGPT Neovim Plugin: Effortless Natural Language Generation with OpenAI's ChatGPT API| | 889|microsoft/RD-Agent !2025-03-28388422|Research and development (R&D) is crucial for the enhancement of industrial productivity, especially in the AI era, where the core aspects of R&D are mainly focused on data and models. We are committed to automate these high-value generic R&D processes through our open source R&D automation tool RD-Agent, which let AI drive data-driven AI.| | 890|Significant-Gravitas/Auto-GPT-Plugins !2025-03-283882-1 |Plugins for Auto-GPT| | 891|apple/ml-mgie !2025-03-2838770|| | 892|OpenDriveLab/UniAD !2025-03-2838727|[CVPR 2023 Best Paper] Planning-oriented Autonomous Driving| | 893|llSourcell/DoctorGPT !2025-03-2838640|DoctorGPT is an LLM that can pass the US Medical Licensing Exam. It works offline, it's cross-platform, & your health data stays private.| | 894|FlagAI-Open/FlagAI !2025-03-2838601|FlagAI (Fast LArge-scale General AI models) is a fast, easy-to-use and extensible toolkit for large-scale model.| | 895|krishnaik06/Roadmap-To-Learn-Generative-AI-In-2024 !2025-03-2838513|Roadmap To Learn Generative AI In 2024| | 896|SysCV/sam-hq !2025-03-2838491|Segment Anything in High Quality| | 897|google/security-research !2025-03-2838420|This project hosts security advisories and their accompanying proof-of-concepts related to research conducted at Google which impact non-Google owned code.| | 898|shroominic/codeinterpreter-api !2025-03-2838330|Open source implementation of the ChatGPT Code Interpreter 👾| | 899|Yonom/assistant-ui !2025-03-2838308|React Components for AI Chat 💬 🚀| | 900|nucleuscloud/neosync !2025-03-2838262|Open source data anonymization and synthetic data orchestration for developers. Create high fidelity synthetic data and sync it across your environments.| | 901|ravenscroftj/turbopilot !2025-03-2838230 |Turbopilot is an open source large-language-model based code completion engine that runs locally on CPU| | 902|NVlabs/Sana !2025-03-28380810|SANA: Efficient High-Resolution Image Synthesis with Linear Diffusion Transformer| | 903|huggingface/distil-whisper !2025-03-2838061|Distilled variant of Whisper for speech recognition. 6x faster, 50% smaller, within 1% word error rate.| | 904|Codium-ai/AlphaCodium !2025-03-2837971|code generation tool that surpasses most human competitors in CodeContests| | 905|fixie-ai/ultravox !2025-03-2837710|A fast multimodal LLM for real-time voice| | 906|unit-mesh/auto-dev !2025-03-28375715|🧙‍AutoDev: The AI-powered coding wizard with multilingual support 🌐, auto code generation 🏗️, and a helpful bug-slaying assistant 🐞! Customizable prompts 🎨 and a magic Auto Dev/Testing/Document/Agent feature 🧪 included! 🚀| | 907|Marker-Inc-Korea/AutoRAG !2025-03-2837432|AutoML tool for RAG| | 908|deepseek-ai/DeepSeek-VL !2025-03-283734-1|DeepSeek-VL: Towards Real-World Vision-Language Understanding| | 909|hiyouga/ChatGLM-Efficient-Tuning !2025-03-283692-1|Fine-tuning ChatGLM-6B with PEFT | | 910| Yue-Yang/ChatGPT-Siri !2025-03-2836921 | Shortcuts for Siri using ChatGPT API gpt-3.5-turbo model | | 911|0hq/WebGPT !2025-03-2836901 |Run GPT model on the browser with WebGPU. An implementation of GPT inference in less than ~2000 lines of vanilla Javascript.| | 912|cvg/LightGlue !2025-03-2836903|LightGlue: Local Feature Matching at Light Speed (ICCV 2023)| | 913|deanxv/coze-discord-proxy !2025-03-2836791|代理Discord-Bot对话Coze-Bot,实现API形式请求GPT4对话模型/微调模型| | 914|MervinPraison/PraisonAI !2025-03-2836764|PraisonAI application combines AutoGen and CrewAI or similar frameworks into a low-code solution for building and managing multi-agent LLM systems, focusing on simplicity, customisation, and efficient human-agent collaboration.| | 915|Ironclad/rivet !2025-03-2836345 |The open-source visual AI programming environment and TypeScript library| | 916|BasedHardware/OpenGlass !2025-03-2835851|Turn any glasses into AI-powered smart glasses| | 917|ricklamers/gpt-code-ui !2025-03-2835840 |An open source implementation of OpenAI's ChatGPT Code interpreter| | 918|whoiskatrin/chart-gpt !2025-03-2835830 |AI tool to build charts based on text input| | 919|github/CopilotForXcode !2025-03-2835788|Xcode extension for GitHub Copilot| | 920|hemansnation/God-Level-Data-Science-ML-Full-Stack !2025-03-2835570 |A collection of scientific methods, processes, algorithms, and systems to build stories & models. This roadmap contains 16 Chapters, whether you are a fresher in the field or an experienced professional who wants to transition into Data Science & AI| | 921|pytorch/torchchat !2025-03-2835461|Run PyTorch LLMs locally on servers, desktop and mobile| | 922| Kent0n-Li/ChatDoctor !2025-03-2835451 | A Medical Chat Model Fine-tuned on LLaMA Model using Medical Domain Knowledge | | 923|xtekky/chatgpt-clone !2025-03-283519-1 |ChatGPT interface with better UI| | 924|jupyterlab/jupyter-ai !2025-03-2835120|A generative AI extension for JupyterLab| | 925|pytorch/torchtitan !2025-03-2835064|A native PyTorch Library for large model training| | 926|minimaxir/simpleaichat !2025-03-2835031|Python package for easily interfacing with chat apps, with robust features and minimal code complexity.| | 927|srush/Tensor-Puzzles !2025-03-2834930|Solve puzzles. Improve your pytorch.| | 928|Helicone/helicone !2025-03-2834918|🧊 Open source LLM-Observability Platform for Developers. One-line integration for monitoring, metrics, evals, agent tracing, prompt management, playground, etc. Supports OpenAI SDK, Vercel AI SDK, Anthropic SDK, LiteLLM, LLamaIndex, LangChain, and more. 🍓 YC W23| | 929|run-llama/llama-hub !2025-03-2834740|A library of data loaders for LLMs made by the community -- to be used with LlamaIndex and/or LangChain| | 930|NExT-GPT/NExT-GPT !2025-03-2834700|Code and models for NExT-GPT: Any-to-Any Multimodal Large Language Model| | 931|souzatharsis/podcastfy !2025-03-2834661|An Open Source Python alternative to NotebookLM's podcast feature: Transforming Multimodal Content into Captivating Multilingual Audio Conversations with GenAI| | 932|Dataherald/dataherald !2025-03-2834450|Interact with your SQL database, Natural Language to SQL using LLMs| | 933|iryna-kondr/scikit-llm !2025-03-2834350 |Seamlessly integrate powerful language models like ChatGPT into scikit-learn for enhanced text analysis tasks.| | 934|Netflix/maestro !2025-03-2834230|Maestro: Netflix’s Workflow Orchestrator| | 935|CanadaHonk/porffor !2025-03-2833560|A from-scratch experimental AOT JS engine, written in JS| | 936|hustvl/Vim !2025-03-2833323|Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model| | 937|pashpashpash/vault-ai !2025-03-2833250 |OP Vault ChatGPT: Give ChatGPT long-term memory using the OP Stack (OpenAI + Pinecone Vector Database). Upload your own custom knowledge base files (PDF, txt, etc) using a simple React frontend.| | 938|tencentmusic/supersonic !2025-03-28330611|SuperSonic is the next-generation BI platform that integrates Chat BI (powered by LLM) and Headless BI (powered by semantic layer) paradigms.| | 939|billmei/every-chatgpt-gui !2025-03-2832981|Every front-end GUI client for ChatGPT| | 940|microsoft/torchgeo !2025-03-2832772|TorchGeo: datasets, samplers, transforms, and pre-trained models for geospatial data| | 941|LLMBook-zh/LLMBook-zh.github.io !2025-03-28326110|《大语言模型》作者:赵鑫,李军毅,周昆,唐天一,文继荣| | 942|dvlab-research/MiniGemini !2025-03-2832601|Official implementation for Mini-Gemini| | 943|rashadphz/farfalle !2025-03-2832460|🔍 AI search engine - self-host with local or cloud LLMs| | 944|Luodian/Otter !2025-03-2832450|🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.| | 945|AprilNEA/ChatGPT-Admin-Web !2025-03-2832370 | ChatGPT WebUI with user management and admin dashboard system| | 946|MarkFzp/act-plus-plus !2025-03-2832365|Imitation Learning algorithms with Co-traing for Mobile ALOHA: ACT, Diffusion Policy, VINN| | 947|ethen8181/machine-learning !2025-03-2832310|🌎 machine learning tutorials (mainly in Python3)| | 948|opengeos/segment-geospatial !2025-03-2832312 |A Python package for segmenting geospatial data with the Segment Anything Model (SAM)| | 949|iusztinpaul/hands-on-llms !2025-03-283225-2|🦖 𝗟𝗲𝗮𝗿𝗻 about 𝗟𝗟𝗠𝘀, 𝗟𝗟𝗠𝗢𝗽𝘀, and 𝘃𝗲𝗰𝘁𝗼𝗿 𝗗𝗕𝘀 for free by designing, training, and deploying a real-time financial advisor LLM system ~ 𝘴𝘰𝘶𝘳𝘤𝘦 𝘤𝘰𝘥𝘦 + 𝘷𝘪𝘥𝘦𝘰 & 𝘳𝘦𝘢𝘥𝘪𝘯𝘨 𝘮𝘢𝘵𝘦𝘳𝘪𝘢𝘭𝘴| | 950|ToTheBeginning/PuLID !2025-03-2832221|Official code for PuLID: Pure and Lightning ID Customization via Contrastive Alignment| | 951|neo4j-labs/llm-graph-builder !2025-03-2832164|Neo4j graph construction from unstructured data using LLMs| | 952|OpenGVLab/InternGPT !2025-03-2832150 |InternGPT (iGPT) is an open source demo platform where you can easily showcase your AI models. Now it supports DragGAN, ChatGPT, ImageBind, multimodal chat like GPT-4, SAM, interactive image editing, etc. Try it at igpt.opengvlab.com (支持DragGAN、ChatGPT、ImageBind、SAM的在线Demo系统)| | 953|PKU-YuanGroup/Video-LLaVA !2025-03-2832060 |Video-LLaVA: Learning United Visual Representation by Alignment Before Projection| | 954|DataTalksClub/llm-zoomcamp !2025-03-2832030|LLM Zoomcamp - a free online course about building an AI bot that can answer questions about your knowledge base| | 955|gptscript-ai/gptscript !2025-03-2832010|Natural Language Programming| |!green-up-arrow.svg 956|isaac-sim/IsaacLab !2025-03-28320113|Unified framework for robot learning built on NVIDIA Isaac Sim| |!red-down-arrow 957|ai-boost/Awesome-GPTs !2025-03-2832003|Curated list of awesome GPTs 👍.| | 958|huggingface/safetensors !2025-03-2831901|Simple, safe way to store and distribute tensors| | 959|linyiLYi/bilibot !2025-03-2831771|A local chatbot fine-tuned by bilibili user comments.| | 960| project-baize/baize-chatbot !2025-03-283168-1 | Let ChatGPT teach your own chatbot in hours with a single GPU! | | 961|Azure-Samples/cognitive-services-speech-sdk !2025-03-2831280|Sample code for the Microsoft Cognitive Services Speech SDK| | 962|microsoft/Phi-3CookBook !2025-03-2831231|This is a Phi-3 book for getting started with Phi-3. Phi-3, a family of open AI models developed by Microsoft. Phi-3 models are the most capable and cost-effective small language models (SLMs) available, outperforming models of the same size and next size up across a variety of language, reasoning, coding, and math benchmarks.| | 963|neuralmagic/deepsparse !2025-03-2831180|Sparsity-aware deep learning inference runtime for CPUs| | 964|sugarforever/chat-ollama !2025-03-2831000|ChatOllama is an open source chatbot based on LLMs. It supports a wide range of language models, and knowledge base management.| | 965|amazon-science/chronos-forecasting !2025-03-2830974|Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting| | 966|damo-vilab/i2vgen-xl !2025-03-2830902|Official repo for VGen: a holistic video generation ecosystem for video generation building on diffusion models| | 967|google-deepmind/gemma !2025-03-2830733|Open weights LLM from Google DeepMind.| | 968|iree-org/iree !2025-03-2830733|A retargetable MLIR-based machine learning compiler and runtime toolkit.| | 969|NVlabs/VILA !2025-03-2830724|VILA - a multi-image visual language model with training, inference and evaluation recipe, deployable from cloud to edge (Jetson Orin and laptops)| | 970|microsoft/torchscale !2025-03-2830661|Foundation Architecture for (M)LLMs| | 971|openai/openai-realtime-console !2025-03-2830656|React app for inspecting, building and debugging with the Realtime API| | 972|daveshap/OpenAIAgentSwarm !2025-03-2830610|HAAS = Hierarchical Autonomous Agent Swarm - "Resistance is futile!"| | 973|microsoft/PromptWizard !2025-03-2830555|Task-Aware Agent-driven Prompt Optimization Framework| | 974|CVI-SZU/Linly !2025-03-2830490 |Chinese-LLaMA basic model; ChatFlow Chinese conversation model; NLP pre-training/command fine-tuning dataset| | 975|cohere-ai/cohere-toolkit !2025-03-2830130|Toolkit is a collection of prebuilt components enabling users to quickly build and deploy RAG applications.| | 976|adamcohenhillel/ADeus !2025-03-2830131|An open source AI wearable device that captures what you say and hear in the real world and then transcribes and stores it on your own server. You can then chat with Adeus using the app, and it will have all the right context about what you want to talk about - a truly personalized, personal AI.| | 977|Lightning-AI/LitServe !2025-03-2830132|Lightning-fast serving engine for AI models. Flexible. Easy. Enterprise-scale.| | 978|potpie-ai/potpie !2025-03-2829973|Prompt-To-Agent : Create custom engineering agents for your codebase| | 979|ant-design/x !2025-03-28299529|Craft AI-driven interfaces effortlessly 🤖| | 980|meta-llama/PurpleLlama !2025-03-2829832|Set of tools to assess and improve LLM security.| | 981|williamyang1991/RerenderAVideo !2025-03-2829800|[SIGGRAPH Asia 2023] Rerender A Video: Zero-Shot Text-Guided Video-to-Video Translation| | 982|baichuan-inc/Baichuan-13B !2025-03-2829790|A 13B large language model developed by Baichuan Intelligent Technology| | 983|Stability-AI/stable-audio-tools !2025-03-2829761|Generative models for conditional audio generation| | 984|li-plus/chatglm.cpp !2025-03-2829720|C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & more LLMs| | 985|NVIDIA/GenerativeAIExamples !2025-03-2829546|Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.| | 986|Josh-XT/AGiXT !2025-03-2829521 |AGiXT is a dynamic AI Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.| | 987|MrForExample/ComfyUI-3D-Pack !2025-03-2829515|An extensive node suite that enables ComfyUI to process 3D inputs (Mesh & UV Texture, etc) using cutting edge algorithms (3DGS, NeRF, etc.)| | 988|olimorris/codecompanion.nvim !2025-03-28295111|✨ AI-powered coding, seamlessly in Neovim. Supports Anthropic, Copilot, Gemini, Ollama, OpenAI and xAI LLMs| | 989|salesforce/CodeT5 !2025-03-282940-1 |Home of CodeT5: Open Code LLMs for Code Understanding and Generation| | 990|facebookresearch/ijepa !2025-03-2829391|Official codebase for I-JEPA, the Image-based Joint-Embedding Predictive Architecture. First outlined in the CVPR paper, "Self-supervised learning from images with a joint-embedding predictive architecture."| | 991|eureka-research/Eureka !2025-03-2829351|Official Repository for "Eureka: Human-Level Reward Design via Coding Large Language Models"| | 992|NVIDIA/trt-llm-rag-windows !2025-03-282934-1|A developer reference project for creating Retrieval Augmented Generation (RAG) chatbots on Windows using TensorRT-LLM| | 993|gmpetrov/databerry !2025-03-282930-1|The no-code platform for building custom LLM Agents| | 994|AI4Finance-Foundation/FinRobot !2025-03-28291946|FinRobot: An Open-Source AI Agent Platform for Financial Applications using LLMs 🚀 🚀 🚀| | 995|nus-apr/auto-code-rover !2025-03-2829013|A project structure aware autonomous software engineer aiming for autonomous program improvement| | 996|deepseek-ai/DreamCraft3D !2025-03-2828921|[ICLR 2024] Official implementation of DreamCraft3D: Hierarchical 3D Generation with Bootstrapped Diffusion Prior| | 997|mlabonne/llm-datasets !2025-03-2828848|High-quality datasets, tools, and concepts for LLM fine-tuning.| | 998|facebookresearch/jepa !2025-03-2828712|PyTorch code and models for V-JEPA self-supervised learning from video.| | 999|facebookresearch/habitat-sim !2025-03-2828604|A flexible, high-performance 3D simulator for Embodied AI research.| | 1000|xenova/whisper-web !2025-03-2828581|ML-powered speech recognition directly in your browser| | 1001|cvlab-columbia/zero123 !2025-03-2828530|Zero-1-to-3: Zero-shot One Image to 3D Object: https://zero123.cs.columbia.edu/| | 1002|yuruotong1/autoMate !2025-03-28285121|Like Manus, Computer Use Agent(CUA) and Omniparser, we are computer-using agents.AI-driven local automation assistant that uses natural language to make computers work by themselves| | 1003|muellerberndt/mini-agi !2025-03-282845-1 |A minimal generic autonomous agent based on GPT3.5/4. Can analyze stock prices, perform network security tests, create art, and order pizza.| | 1004|allenai/open-instruct !2025-03-2828432|| | 1005|CodingChallengesFYI/SharedSolutions !2025-03-2828360|Publicly shared solutions to Coding Challenges| | 1006|hegelai/prompttools !2025-03-2828220|Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate).| | 1007|mazzzystar/Queryable !2025-03-2828222|Run CLIP on iPhone to Search Photos.| | 1008|Doubiiu/DynamiCrafter !2025-03-2828173|DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors| | 1009|SamurAIGPT/privateGPT !2025-03-282805-1 |An app to interact privately with your documents using the power of GPT, 100% privately, no data leaks| | 1010|facebookresearch/Pearl !2025-03-2827951|A Production-ready Reinforcement Learning AI Agent Library brought by the Applied Reinforcement Learning team at Meta.| | 1011|intuitem/ciso-assistant-community !2025-03-2827954|CISO Assistant is a one-stop-shop for GRC, covering Risk, AppSec and Audit Management and supporting +70 frameworks worldwide with auto-mapping: NIST CSF, ISO 27001, SOC2, CIS, PCI DSS, NIS2, CMMC, PSPF, GDPR, HIPAA, Essential Eight, NYDFS-500, DORA, NIST AI RMF, 800-53, 800-171, CyFun, CJIS, AirCyber, NCSC, ECC, SCF and so much more| | 1012|facebookresearch/audio2photoreal !2025-03-2827840|Code and dataset for photorealistic Codec Avatars driven from audio| | 1013|Azure/azure-rest-api-specs !2025-03-2827770|The source for REST API specifications for Microsoft Azure.| | 1014|SCUTlihaoyu/open-chat-video-editor !2025-03-2827690 |Open source short video automatic generation tool| | 1015|Alpha-VLLM/LLaMA2-Accessory !2025-03-2827642|An Open-source Toolkit for LLM Development| | 1016|johnma2006/mamba-minimal !2025-03-2827601|Simple, minimal implementation of the Mamba SSM in one file of PyTorch.| | 1017|nerfstudio-project/gsplat !2025-03-2827576|CUDA accelerated rasterization of gaussian splatting| | 1018|Physical-Intelligence/openpi !2025-03-28274617|| | 1019|leptonai/leptonai !2025-03-2827246|A Pythonic framework to simplify AI service building| |!green-up-arrow.svg 1020|joanrod/star-vector !2025-03-28271149|StarVector is a foundation model for SVG generation that transforms vectorization into a code generation task. Using a vision-language modeling architecture, StarVector processes both visual and textual inputs to produce high-quality SVG code with remarkable precision.| |!red-down-arrow 1021|jqnatividad/qsv !2025-03-2827092|CSVs sliced, diced & analyzed.| | 1022|FranxYao/chain-of-thought-hub !2025-03-2826991|Benchmarking large language models' complex reasoning ability with chain-of-thought prompting| | 1023|princeton-nlp/SWE-bench !2025-03-2826965|[ICLR 2024] SWE-Bench: Can Language Models Resolve Real-world Github Issues?| | 1024|elastic/otel-profiling-agent !2025-03-2826930|The production-scale datacenter profiler| | 1025|src-d/hercules !2025-03-2826900|Gaining advanced insights from Git repository history.| | 1026|lanqian528/chat2api !2025-03-2826695|A service that can convert ChatGPT on the web to OpenAI API format.| | 1027|ishan0102/vimGPT !2025-03-2826681|Browse the web with GPT-4V and Vimium| | 1028|TMElyralab/MuseV !2025-03-2826650|MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising| | 1029|georgia-tech-db/eva !2025-03-2826600 |AI-Relational Database System | | 1030|kubernetes-sigs/controller-runtime !2025-03-2826590|Repo for the controller-runtime subproject of kubebuilder (sig-apimachinery)| | 1031|gptlink/gptlink !2025-03-2826550 |Build your own free commercial ChatGPT environment in 10 minutes. The setup is simple and includes features such as user management, orders, tasks, and payments| | 1032|pytorch/executorch !2025-03-2826534|On-device AI across mobile, embedded and edge for PyTorch| | 1033|NVIDIA/nv-ingest !2025-03-2826290|NVIDIA Ingest is an early access set of microservices for parsing hundreds of thousands of complex, messy unstructured PDFs and other enterprise documents into metadata and text to embed into retrieval systems.| | 1034|SuperTux/supertux !2025-03-2826081|SuperTux source code| | 1035|abi/secret-llama !2025-03-2826050|Fully private LLM chatbot that runs entirely with a browser with no server needed. Supports Mistral and LLama 3.| | 1036|liou666/polyglot !2025-03-2825841 |Desktop AI Language Practice Application| | 1037|janhq/nitro !2025-03-2825821|A fast, lightweight, embeddable inference engine to supercharge your apps with local AI. OpenAI-compatible API| | 1038|deepseek-ai/DeepSeek-Math !2025-03-2825825|DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models| | 1039|anthropics/prompt-eng-interactive-tutorial !2025-03-2825781|Anthropic's Interactive Prompt Engineering Tutorial| | 1040|microsoft/promptbench !2025-03-2825741|A unified evaluation framework for large language models| | 1041|baaivision/Painter !2025-03-2825580 |Painter & SegGPT Series: Vision Foundation Models from BAAI| | 1042|OpenPipe/OpenPipe !2025-03-2825581|Turn expensive prompts into cheap fine-tuned models| | 1043|TracecatHQ/tracecat !2025-03-2825531|😼 The AI-native, open source alternative to Tines / Splunk SOAR.| | 1044|JoshuaC215/agent-service-toolkit !2025-03-2825528|Full toolkit for running an AI agent service built with LangGraph, FastAPI and Streamlit| | 1045|databricks/dbrx !2025-03-2825460|Code examples and resources for DBRX, a large language model developed by Databricks| | 1046|lamini-ai/lamini !2025-03-2825271 |Official repo for Lamini's data generator for generating instructions to train instruction-following LLMs| | 1047|mshumer/gpt-author !2025-03-282510-1|| | 1048|TMElyralab/MusePose !2025-03-2824971|MusePose: a Pose-Driven Image-to-Video Framework for Virtual Human Generation| | 1049|Kludex/fastapi-tips !2025-03-2824974|FastAPI Tips by The FastAPI Expert!| | 1050|openai/simple-evals !2025-03-2824813|| | 1051|iterative/datachain !2025-03-2824732|AI-data warehouse to enrich, transform and analyze data from cloud storages| | 1052|girafe-ai/ml-course !2025-03-2824703|Open Machine Learning course| | 1053|kevmo314/magic-copy !2025-03-2824620 |Magic Copy is a Chrome extension that uses Meta's Segment Anything Model to extract a foreground object from an image and copy it to the clipboard.| | 1054|Eladlev/AutoPrompt !2025-03-2824432|A framework for prompt tuning using Intent-based Prompt Calibration| | 1055|OpenBMB/CPM-Bee !2025-03-282434-1 |A bilingual large-scale model with trillions of parameters| | 1056|IDEA-Research/T-Rex !2025-03-2824310|T-Rex2: Towards Generic Object Detection via Text-Visual Prompt Synergy| | 1057|microsoft/genaiscript !2025-03-2824202|Automatable GenAI Scripting| | 1058|paulpierre/RasaGPT !2025-03-2824090 |💬 RasaGPT is the first headless LLM chatbot platform built on top of Rasa and Langchain. Built w/ Rasa, FastAPI, Langchain, LlamaIndex, SQLModel, pgvector, ngrok, telegram| | 1059|ashishpatel26/LLM-Finetuning !2025-03-2823911|LLM Finetuning with peft| | 1060|SoraWebui/SoraWebui !2025-03-2823570|SoraWebui is an open-source Sora web client, enabling users to easily create videos from text with OpenAI's Sora model.| | 1061|6drf21e/ChatTTScolab !2025-03-2823491|🚀 一键部署(含离线整合包)!基于 ChatTTS ,支持音色抽卡、长音频生成和分角色朗读。简单易用,无需复杂安装。| | 1062|Azure/PyRIT !2025-03-2823343|The Python Risk Identification Tool for generative AI (PyRIT) is an open access automation framework to empower security professionals and machine learning engineers to proactively find risks in their generative AI systems.| | 1063|tencent-ailab/V-Express !2025-03-2823201|V-Express aims to generate a talking head video under the control of a reference image, an audio, and a sequence of V-Kps images.| | 1064|THUDM/CogVLM2 !2025-03-2823170|GPT4V-level open-source multi-modal model based on Llama3-8B| | 1065|dvmazur/mixtral-offloading !2025-03-2823001|Run Mixtral-8x7B models in Colab or consumer desktops| | 1066|semanser/codel !2025-03-2822950|✨ Fully autonomous AI Agent that can perform complicated tasks and projects using terminal, browser, and editor.| | 1067|mshumer/gpt-investor !2025-03-2822590|| | 1068|aixcoder-plugin/aiXcoder-7B !2025-03-2822550|official repository of aiXcoder-7B Code Large Language Model| | 1069|Azure-Samples/graphrag-accelerator !2025-03-2822503|One-click deploy of a Knowledge Graph powered RAG (GraphRAG) in Azure| | 1070|emcf/engshell !2025-03-2821830 |An English-language shell for any OS, powered by LLMs| | 1071|hncboy/chatgpt-web-java !2025-03-2821771|ChatGPT project developed in Java, based on Spring Boot 3 and JDK 17, supports both AccessToken and ApiKey modes| | 1072|openai/consistencydecoder !2025-03-2821692|Consistency Distilled Diff VAE| | 1073|Alpha-VLLM/Lumina-T2X !2025-03-2821681|Lumina-T2X is a unified framework for Text to Any Modality Generation| | 1074|bghira/SimpleTuner !2025-03-2821612|A general fine-tuning kit geared toward Stable Diffusion 2.1, Stable Diffusion 3, DeepFloyd, and SDXL.| | 1075|JiauZhang/DragGAN !2025-03-2821530 |Implementation of DragGAN: Interactive Point-based Manipulation on the Generative Image Manifold| | 1076|cgpotts/cs224u !2025-03-2821390|Code for Stanford CS224u| | 1077|PKU-YuanGroup/MoE-LLaVA !2025-03-2821300|Mixture-of-Experts for Large Vision-Language Models| | 1078|darrenburns/elia !2025-03-2820831|A snappy, keyboard-centric terminal user interface for interacting with large language models. Chat with ChatGPT, Claude, Llama 3, Phi 3, Mistral, Gemma and more.| | 1079|ageerle/ruoyi-ai !2025-03-28207898|RuoYi AI 是一个全栈式 AI 开发平台,旨在帮助开发者快速构建和部署个性化的 AI 应用。| | 1080|NVIDIA/gpu-operator !2025-03-2820510|NVIDIA GPU Operator creates/configures/manages GPUs atop Kubernetes| | 1081|BAAI-Agents/Cradle !2025-03-2820481|The Cradle framework is a first attempt at General Computer Control (GCC). Cradle supports agents to ace any computer task by enabling strong reasoning abilities, self-improvment, and skill curation, in a standardized general environment with minimal requirements.| | 1082|microsoft/aici !2025-03-2820080|AICI: Prompts as (Wasm) Programs| | 1083|PRIS-CV/DemoFusion !2025-03-2820040|Let us democratise high-resolution generation! (arXiv 2023)| | 1084|apple/axlearn !2025-03-2820012|An Extensible Deep Learning Library| | 1085|naver/mast3r !2025-03-2819685|Grounding Image Matching in 3D with MASt3R| | 1086|liltom-eth/llama2-webui !2025-03-281958-1|Run Llama 2 locally with gradio UI on GPU or CPU from anywhere (Linux/Windows/Mac). Supporting Llama-2-7B/13B/70B with 8-bit, 4-bit. Supporting GPU inference (6 GB VRAM) and CPU inference.| | 1087|GaParmar/img2img-turbo !2025-03-2819582|One-step image-to-image with Stable Diffusion turbo: sketch2image, day2night, and more| | 1088|Niek/chatgpt-web !2025-03-2819560|ChatGPT web interface using the OpenAI API| | 1089|huggingface/cookbook !2025-03-2819421|Open-source AI cookbook| | 1090|pytorch/ao !2025-03-2819241|PyTorch native quantization and sparsity for training and inference| | 1091|emcie-co/parlant !2025-03-2819053|The behavior guidance framework for customer-facing LLM agents| | 1092|ymcui/Chinese-LLaMA-Alpaca-3 !2025-03-2818980|中文羊驼大模型三期项目 (Chinese Llama-3 LLMs) developed from Meta Llama 3| | 1093|Nutlope/notesGPT !2025-03-2818811|Record voice notes & transcribe, summarize, and get tasks| | 1094|InstantStyle/InstantStyle !2025-03-2818791|InstantStyle: Free Lunch towards Style-Preserving in Text-to-Image Generation 🔥| | 1095|idaholab/moose !2025-03-2818771|Multiphysics Object Oriented Simulation Environment| | 1096|The-OpenROAD-Project/OpenROAD !2025-03-2818351|OpenROAD's unified application implementing an RTL-to-GDS Flow. Documentation at https://openroad.readthedocs.io/en/latest/| | 1097|alibaba/spring-ai-alibaba !2025-03-281831121|Agentic AI Framework for Java Developers| | 1098|ytongbai/LVM !2025-03-2817990|Sequential Modeling Enables Scalable Learning for Large Vision Models| | 1099|microsoft/sample-app-aoai-chatGPT !2025-03-2817981|[PREVIEW] Sample code for a simple web chat experience targeting chatGPT through AOAI.| | 1100|AI-Citizen/SolidGPT !2025-03-2817830|Chat everything with your code repository, ask repository level code questions, and discuss your requirements. AI Scan and learning your code repository, provide you code repository level answer🧱 🧱| | 1101|YangLing0818/RPG-DiffusionMaster !2025-03-2817784|Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs (PRG)| | 1102|kyegomez/BitNet !2025-03-2817710|Implementation of "BitNet: Scaling 1-bit Transformers for Large Language Models" in pytorch| | 1103|eloialonso/diamond !2025-03-2817671|DIAMOND (DIffusion As a Model Of eNvironment Dreams) is a reinforcement learning agent trained in a diffusion world model.| | 1104|flowdriveai/flowpilot !2025-03-2817250|flow-pilot is an openpilot based driver assistance system that runs on linux, windows and android powered machines.| | 1105|xlang-ai/OSWorld !2025-03-2817200|OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments| | 1106|linyiLYi/snake-ai !2025-03-2817031|An AI agent that beats the classic game "Snake".| | 1107|baaivision/Emu !2025-03-2816991|Emu Series: Generative Multimodal Models from BAAI| | 1108|kevmo314/scuda !2025-03-2816870|SCUDA is a GPU over IP bridge allowing GPUs on remote machines to be attached to CPU-only machines.| | 1109|SharifiZarchi/IntroductiontoMachineLearning !2025-03-2816701|دوره‌ی مقدمه‌ای بر یادگیری ماشین، برای دانشجویان| | 1110|google/maxtext !2025-03-2816670|A simple, performant and scalable Jax LLM!| | 1111|ml-explore/mlx-swift-examples !2025-03-2816471|Examples using MLX Swift| | 1112|unitreerobotics/unitreerlgym !2025-03-2816256|| | 1113|collabora/WhisperFusion !2025-03-2815901|WhisperFusion builds upon the capabilities of WhisperLive and WhisperSpeech to provide a seamless conversations with an AI.| | 1114|lichao-sun/Mora !2025-03-2815520|Mora: More like Sora for Generalist Video Generation| | 1115|GoogleCloudPlatform/localllm !2025-03-2815370|Run LLMs locally on Cloud Workstations| | 1116|TencentARC/BrushNet !2025-03-2815330|The official implementation of paper "BrushNet: A Plug-and-Play Image Inpainting Model with Decomposed Dual-Branch Diffusion"| | 1117|ai-christianson/RA.Aid !2025-03-2815288|Develop software autonomously.| | 1118|stephansturges/WALDO !2025-03-2815170|Whereabouts Ascertainment for Low-lying Detectable Objects. The SOTA in FOSS AI for drones!| | 1119|skills/copilot-codespaces-vscode !2025-03-2815112|Develop with AI-powered code suggestions using GitHub Copilot and VS Code| | 1120|andrewnguonly/Lumos !2025-03-2814920|A RAG LLM co-pilot for browsing the web, powered by local LLMs| | 1121|TeamNewPipe/NewPipeExtractor !2025-03-2814811|NewPipe's core library for extracting data from streaming sites| | 1122|mhamilton723/FeatUp !2025-03-2814770|Official code for "FeatUp: A Model-Agnostic Frameworkfor Features at Any Resolution" ICLR 2024| | 1123|AnswerDotAI/fsdpqlora !2025-03-2814671|Training LLMs with QLoRA + FSDP| | 1124|jgravelle/AutoGroq !2025-03-2814330|| | 1125|OpenGenerativeAI/llm-colosseum !2025-03-2814130|Benchmark LLMs by fighting in Street Fighter 3! The new way to evaluate the quality of an LLM| | 1126|microsoft/vscode-ai-toolkit !2025-03-2814000|| | 1127|McGill-NLP/webllama !2025-03-2813930|Llama-3 agents that can browse the web by following instructions and talking to you| | 1128|lucidrains/self-rewarding-lm-pytorch !2025-03-2813760|Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI| | 1129|ishaan1013/sandbox !2025-03-2813650|A cloud-based code editing environment with an AI copilot and real-time collaboration.| | 1130|goatcorp/Dalamud !2025-03-2813275|FFXIV plugin framework and API| | 1131|Lightning-AI/lightning-thunder !2025-03-2813151|Make PyTorch models Lightning fast! Thunder is a source to source compiler for PyTorch. It enables using different hardware executors at once.| | 1132|PKU-YuanGroup/MagicTime !2025-03-2813052|MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators| | 1133|SakanaAI/evolutionary-model-merge !2025-03-2813000|Official repository of Evolutionary Optimization of Model Merging Recipes| | 1134|a-real-ai/pywinassistant !2025-03-2812950|The first open source Large Action Model generalist Artificial Narrow Intelligence that controls completely human user interfaces by only using natural language. PyWinAssistant utilizes Visualization-of-Thought Elicits Spatial Reasoning in Large Language Models.| | 1135|TraceMachina/nativelink !2025-03-2812630|NativeLink is an open source high-performance build cache and remote execution server, compatible with Bazel, Buck2, Reclient, and other RBE-compatible build systems. It offers drastically faster builds, reduced test flakiness, and significant infrastructure cost savings.| | 1136|MLSysOps/MLE-agent !2025-03-2812500|🤖 MLE-Agent: Your intelligent companion for seamless AI engineering and research. 🔍 Integrate with arxiv and paper with code to provide better code/research plans 🧰 OpenAI, Ollama, etc supported. 🎆 Code RAG| | 1137|wpilibsuite/allwpilib !2025-03-2811610|Official Repository of WPILibJ and WPILibC| | 1138|elfvingralf/macOSpilot-ai-assistant !2025-03-2811470|Voice + Vision powered AI assistant that answers questions about any application, in context and in audio.| | 1139|langchain-ai/langchain-extract !2025-03-2811210|🦜⛏️ Did you say you like data?| | 1140|FoundationVision/GLEE !2025-03-2811120|【CVPR2024】GLEE: General Object Foundation Model for Images and Videos at Scale| | 1141|Profluent-AI/OpenCRISPR !2025-03-2810990|AI-generated gene editing systems| | 1142|zju3dv/EasyVolcap !2025-03-2810821|[SIGGRAPH Asia 2023 (Technical Communications)] EasyVolcap: Accelerating Neural Volumetric Video Research| | 1143|PaddlePaddle/PaddleHelix !2025-03-2810560|Bio-Computing Platform Featuring Large-Scale Representation Learning and Multi-Task Deep Learning “螺旋桨”生物计算工具集| | 1144|myshell-ai/JetMoE !2025-03-289800|Reaching LLaMA2 Performance with 0.1M Dollars| | 1145|likejazz/llama3.np !2025-03-289770|llama3.np is pure NumPy implementation for Llama 3 model.| | 1146|mustafaaljadery/gemma-2B-10M !2025-03-289500|Gemma 2B with 10M context length using Infini-attention.| | 1147|HITsz-TMG/FilmAgent !2025-03-289382|Resources of our paper "FilmAgent: A Multi-Agent Framework for End-to-End Film Automation in Virtual 3D Spaces". New versions in the making!| | 1148|aws-samples/amazon-bedrock-samples !2025-03-289362|This repository contains examples for customers to get started using the Amazon Bedrock Service. This contains examples for all available foundational models| | 1149|Akkudoktor-EOS/EOS !2025-03-2893154|This repository features an Energy Optimization System (EOS) that optimizes energy distribution, usage for batteries, heat pumps& household devices. It includes predictive models for electricity prices (planned), load forecasting& dynamic optimization to maximize energy efficiency & minimize costs. Founder Dr. Andreas Schmitz (YouTube @akkudoktor)| Tip: | symbol| rule | | :----| :---- | |🔥 | 256 1k| |!green-up-arrow.svg !red-down-arrow | ranking up / down| |⭐ | on trending page today| [Back to Top] Tools | No. | Tool | Description | | ----:|:----------------------------------------------- |:------------------------------------------------------------------------------------------- | | 1 | ChatGPT | A sibling model to InstructGPT, which is trained to follow instructions in a prompt and provide a detailed response | | 2 | DALL·E 2 | Create original, realistic images and art from a text description | | 3 | Murf AI | AI enabled, real people's voices| | 4 | Midjourney | An independent research lab that produces an artificial intelligence program under the same name that creates images from textual descriptions, used in Discord | 5 | Make-A-Video | Make-A-Video is a state-of-the-art AI system that generates videos from text | | 6 | Creative Reality™ Studio by D-ID| Use generative AI to create future-facing videos| | 7 | chat.D-ID| The First App Enabling Face-to-Face Conversations with ChatGPT| | 8 | Notion AI| Access the limitless power of AI, right inside Notion. Work faster. Write better. Think bigger. | | 9 | Runway| Text to Video with Gen-2 | | 10 | Resemble AI| Resemble’s AI voice generator lets you create human–like voice overs in seconds | | 11 | Cursor| Write, edit, and chat about your code with a powerful AI | | 12 | Hugging Face| Build, train and deploy state of the art models powered by the reference open source in machine learning | | 13 | Claude | A next-generation AI assistant for your tasks, no matter the scale | | 14 | Poe| Poe lets you ask questions, get instant answers, and have back-and-forth conversations with AI. Gives access to GPT-4, gpt-3.5-turbo, Claude from Anthropic, and a variety of other bots| [Back to Top] Websites | No. | WebSite |Description | | ----:|:------------------------------------------ |:---------------------------------------------------------------------------------------- | | 1 | OpenAI | An artificial intelligence research lab | | 2 | Bard | Base Google's LaMDA chatbots and pull from internet | | 3 | ERNIE Bot | Baidu’s new generation knowledge-enhanced large language model is a new member of the Wenxin large model family | | 4 | DALL·E 2 | An AI system that can create realistic images and art from a description in natural language | | 5 | Whisper | A general-purpose speech recognition model | | 6| CivitAI| A platform that makes it easy for people to share and discover resources for creating AI art| | 7|D-ID| D-ID’s Generative AI enables users to transform any picture or video into extraordinary experiences| | 8| Nvidia eDiff-I| Text-to-Image Diffusion Models with Ensemble of Expert Denoisers | | 9| Stability AI| The world's leading open source generative AI company which opened source Stable Diffusion | | 10| Meta AI| Whether it be research, product or infrastructure development, we’re driven to innovate responsibly with AI to benefit the world | | 11| ANTHROPIC| AI research and products that put safety at the frontier | [Back to Top] Reports&Papers | No. | Report&Paper | Description | |:---- |:-------------------------------------------------------------------------------------------------------------- |:---------------------------------------------------- | | 1 | GPT-4 Technical Report | GPT-4 Technical Report | | 2 | mli/paper-reading | Deep learning classics and new papers are read carefully paragraph by paragraph. | | 3 | labmlai/annotateddeeplearningpaperimplementations| A collection of simple PyTorch implementations of neural networks and related algorithms, which are documented with explanations | | 4 | Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models | Talking, Drawing and Editing with Visual Foundation Models | | 5 | OpenAI Research | The latest research report and papers from OpenAI | | 6 | Make-A-Video: Text-to-Video Generation without Text-Video Data|Meta's Text-to-Video Generation| | 7 | eDiff-I: Text-to-Image Diffusion Models with Ensemble of Expert Denoisers| Nvidia eDiff-I - New generation of generative AI content creation tool | | 8 | Training an Assistant-style Chatbot with Large Scale Data Distillation from GPT-3.5-Turbo | 2023 GPT4All Technical Report | | 9 | Segment Anything| Meta Segment Anything | | 10 | LLaMA: Open and Efficient Foundation Language Models| LLaMA: a collection of foundation language models ranging from 7B to 65B parameters| | 11 | papers-we-love/papers-we-love |Papers from the computer science community to read and discuss| | 12 | CVPR 2023 papers |The most exciting and influential CVPR 2023 papers| [Back to Top] Tutorials | No. | Tutorial | Description| |:---- |:---------------------------------------------------------------- | --- | | 1 | Coursera - Machine Learning | The Machine Learning Specialization Course taught by Dr. Andrew Ng| | 2 | microsoft/ML-For-Beginners | 12 weeks, 26 lessons, 52 quizzes, classic Machine Learning for all| | 3 | ChatGPT Prompt Engineering for Developers | This short course taught by Isa Fulford (OpenAI) and Andrew Ng (DeepLearning.AI) will teach how to use a large language model (LLM) to quickly build new and powerful applications | | 4 | Dive into Deep Learning |Targeting Chinese readers, functional and open for discussion. The Chinese and English versions are used for teaching in over 400 universities across more than 60 countries | | 5 | AI Expert Roadmap | Roadmap to becoming an Artificial Intelligence Expert in 2022 | | 6 | Computer Science courses |List of Computer Science courses with video lectures| | 7 | Machine Learning with Python | Machine Learning with Python Certification on freeCodeCamp| | 8 | Building Systems with the ChatGPT API | This short course taught by Isa Fulford (OpenAI) and Andrew Ng (DeepLearning.AI), you will learn how to automate complex workflows using chain calls to a large language model| | 9 | LangChain for LLM Application Development | This short course taught by Harrison Chase (Co-Founder and CEO at LangChain) and Andrew Ng. you will gain essential skills in expanding the use cases and capabilities of language models in application development using the LangChain framework| | 10 | How Diffusion Models Work | This short course taught by Sharon Zhou (CEO, Co-founder, Lamini). you will gain a deep familiarity with the diffusion process and the models which carry it out. More than simply pulling in a pre-built model or using an API, this course will teach you to build a diffusion model from scratch| | 11 | Free Programming Books For AI |📚 Freely available programming books for AI | | 12 | microsoft/AI-For-Beginners |12 Weeks, 24 Lessons, AI for All!| | 13 | hemansnation/God-Level-Data-Science-ML-Full-Stack |A collection of scientific methods, processes, algorithms, and systems to build stories & models. This roadmap contains 16 Chapters, whether you are a fresher in the field or an experienced professional who wants to transition into Data Science & AI| | 14 | datawhalechina/prompt-engineering-for-developers |Chinese version of Andrew Ng's Big Model Series Courses, including "Prompt Engineering", "Building System", and "LangChain"| | 15 | ossu/computer-science |🎓 Path to a free self-taught education in Computer Science!| | 16 | microsoft/Data-Science-For-Beginners | 10 Weeks, 20 Lessons, Data Science for All! | |17 |jwasham/coding-interview-university !2023-09-29268215336 |A complete computer science study plan to become a software engineer.| [Back to Top] Thanks If this project has been helpful to you in any way, please give it a ⭐️ by clicking on the star.

h2o-llmstudio
github
LLM Vibe Score0.499
Human Vibe Score0.04822694170894296
h2oaiMar 28, 2025

h2o-llmstudio

Welcome to H2O LLM Studio, a framework and no-code GUI designed for fine-tuning state-of-the-art large language models (LLMs). Jump to With H2O LLM Studio, you can Quickstart What's New Setup Recommended Install Virtual Environments Run H2O LLM Studio GUI Run H2O LLM Studio GUI using Docker Run H2O LLM Studio with command line interface (CLI) Troubleshooting Data format and example data Training your model Example: Run on OASST data via CLI Model checkpoints Documentation Contributing License With H2O LLM Studio, you can easily and effectively fine-tune LLMs without the need for any coding experience. use a graphic user interface (GUI) specially designed for large language models. finetune any LLM using a large variety of hyperparameters. use recent finetuning techniques such as Low-Rank Adaptation (LoRA) and 8-bit model training with a low memory footprint. use Reinforcement Learning (RL) to finetune your model (experimental) use advanced evaluation metrics to judge generated answers by the model. track and compare your model performance visually. In addition, Neptune and W&B integration can be used. chat with your model and get instant feedback on your model performance. easily export your model to the Hugging Face Hub and share it with the community. Quickstart For questions, discussing, or just hanging out, come and join our Discord! Use cloud-based runpod.io instance to run the H2O LLM Studio GUI. Using CLI for fine-tuning LLMs: What's New PR 788 New problem type for Causal Regression Modeling allows to train single target regression data using LLMs. PR 747 Fully removed RLHF in favor of DPO/IPO/KTO optimization. PR 741 Removing separate max length settings for prompt and answer in favor of a single maxlength settings better resembling chattemplate functionality from transformers. PR 592 Added KTOPairLoss for DPO modeling allowing to train models with simple preference data. Data currently needs to be manually prepared by randomly matching positive and negative examples as pairs. PR 592 Starting to deprecate RLHF in favor of DPO/IPO optimization. Training is disabled, but old experiments are still viewable. RLHF will be fully removed in a future release. PR 530 Introduced a new problem type for DPO/IPO optimization. This optimization technique can be used as an alternative to RLHF. PR 288 Introduced Deepspeed for sharded training allowing to train larger models on machines with multiple GPUs. Requires NVLink. This feature replaces FSDP and offers more flexibility. Deepspeed requires a system installation of cudatoolkit and we recommend using version 12.1. See Recommended Install. PR 449 New problem type for Causal Classification Modeling allows to train binary and multiclass models using LLMs. PR 364 User secrets are now handled more securely and flexible. Support for handling secrets using the 'keyring' library was added. User settings are tried to be migrated automatically. Please note that due to current rapid development we cannot guarantee full backwards compatibility of new functionality. We thus recommend to pin the version of the framework to the one you used for your experiments. For resetting, please delete/backup your data and output folders. Setup H2O LLM Studio requires a machine with Ubuntu 16.04+ and at least one recent Nvidia GPU with Nvidia drivers version >= 470.57.02. For larger models, we recommend at least 24GB of GPU memory. For more information about installation prerequisites, see the Set up H2O LLM Studio guide in the documentation. For a performance comparison of different GPUs, see the H2O LLM Studio performance guide in the documentation. Recommended Install The recommended way to install H2O LLM Studio is using pipenv with Python 3.10. To install Python 3.10 on Ubuntu 16.04+, execute the following commands: System installs (Python 3.10) Installing NVIDIA Drivers (if required) If deploying on a 'bare metal' machine running Ubuntu, one may need to install the required Nvidia drivers and CUDA. The following commands show how to retrieve the latest drivers for a machine running Ubuntu 20.04 as an example. One can update the following based on their OS. alternatively, one can install cudatoolkits in a conda environment: Virtual environments We offer various ways of setting up the necessary python environment. Pipenv virtual environment The following command will create a virtual environment using pipenv and will install the dependencies using pipenv: If you are having troubles installing the flash_attn package, consider running instead. This will install the dependencies without the flash_attn package. Note that this will disable the use of Flash Attention 2 and model training will be slower and consume more memory. Nightly Conda virtual environment You can also setup a conda virtual environment that can also deviate from the recommended setup. The contains a command that installs a fresh conda environment with CUDA 12.4 and current nightly PyTorch. Using requirements.txt If you wish to use another virtual environment, you can also install the dependencies using the requirements.txt file: Run H2O LLM Studio GUI You can start H2O LLM Studio using the following command: This command will start the H2O wave server and app. Navigate to (we recommend using Chrome) to access H2O LLM Studio and start fine-tuning your models! If you are running H2O LLM Studio with a custom environment other than Pipenv, you need to start the app as follows: If you are using the nightly conda environment, you can run . Run H2O LLM Studio GUI using Docker Install Docker first by following instructions from NVIDIA Containers. Make sure to have nvidia-container-toolkit installed on your machine as outlined in the instructions. H2O LLM Studio images are stored in the h2oai dockerhub container repository. Navigate to (we recommend using Chrome) to access H2O LLM Studio and start fine-tuning your models! (Note other helpful docker commands are docker ps and docker kill.) If you prefer to build your own Docker image from source, follow the instructions below. Run H2O LLM Studio with command line interface (CLI) You can also use H2O LLM Studio with the command line interface (CLI) and specify the configuration .yaml file that contains all the experiment parameters. To finetune using H2O LLM Studio with CLI, activate the pipenv environment by running make shell, and then use the following command: To run on multiple GPUs in DDP mode, run the following command: By default, the framework will run on the first k GPUs. If you want to specify specific GPUs to run on, use the CUDAVISIBLEDEVICES environment variable before the command. To start an interactive chat with your trained model, use the following command: where experiment_name is the output folder of the experiment you want to chat with (see configuration). The interactive chat will also work with model that were finetuned using the UI. To publish the model to Hugging Face, use the following command: pathtoexperiment is the output folder of the experiment. device is the target device for running the model, either 'cpu' or 'cuda:0'. Default is 'cuda:0'. api_key is the Hugging Face API Key. If user logged in, it can be omitted. user_id is the Hugging Face user ID. If user logged in, it can be omitted. model_name is the name of the model to be published on Hugging Face. It can be omitted. safe_serialization is a flag indicating whether safe serialization should be used. Default is True. Troubleshooting If running on cloud based machines such as runpod, you may need to set the following environment variable to allow the H2O Wave server to accept connections from the proxy: If you are experiencing timeouts when running the H2O Wave server remotely, you can increase the timeout by setting the following environment variables: All default to 5 (seconds). Increase them if you are experiencing timeouts. Use -1 to disable the timeout. Data format and example data For details on the data format required when importing your data or example data that you can use to try out H2O LLM Studio, see Data format in the H2O LLM Studio documentation. Training your model With H2O LLM Studio, training your large language model is easy and intuitive. First, upload your dataset and then start training your model. Start by creating an experiment. You can then monitor and manage your experiment, compare experiments, or push the model to Hugging Face to share it with the community. Example: Run on OASST data via CLI As an example, you can run an experiment on the OASST data via CLI. For instructions, see Run an experiment on the OASST data guide in the H2O LLM Studio documentation. Model checkpoints All open-source datasets and models are posted on H2O.ai's Hugging Face page and our H2OGPT repository. Documentation Detailed documentation and frequently asked questions (FAQs) for H2O LLM Studio can be found at . If you wish to contribute to the docs, navigate to the /documentation folder of this repo and refer to the README.md for more information. Contributing We are happy to accept contributions to the H2O LLM Studio project. Please refer to the CONTRIBUTING.md file for more information. License H2O LLM Studio is licensed under the Apache 2.0 license. Please see the LICENSE file for more information.

Production-Level-Deep-Learning
github
LLM Vibe Score0.619
Human Vibe Score0.8326638433689385
alirezadirMar 28, 2025

Production-Level-Deep-Learning

:bulb: A Guide to Production Level Deep Learning :clapper: :scroll: :ferry: 🇨🇳 Translation in Chinese.md) :label: NEW: Machine Learning Interviews :label: Note: This repo is under continous development, and all feedback and contribution are very welcome :blush: Deploying deep learning models in production can be challenging, as it is far beyond training models with good performance. Several distinct components need to be designed and developed in order to deploy a production level deep learning system (seen below): This repo aims to be an engineering guideline for building production-level deep learning systems which will be deployed in real world applications. The material presented here is borrowed from Full Stack Deep Learning Bootcamp (by Pieter Abbeel at UC Berkeley, Josh Tobin at OpenAI, and Sergey Karayev at Turnitin), TFX workshop by Robert Crowe, and Pipeline.ai's Advanced KubeFlow Meetup by Chris Fregly. Machine Learning Projects Fun :flushed: fact: 85% of AI projects fail. 1 Potential reasons include: Technically infeasible or poorly scoped Never make the leap to production Unclear success criteria (metrics) Poor team management ML Projects lifecycle Importance of understanding state of the art in your domain: Helps to understand what is possible Helps to know what to try next Mental Model for ML project The two important factors to consider when defining and prioritizing ML projects: High Impact: Complex parts of your pipeline Where "cheap prediction" is valuable Where automating complicated manual process is valuable Low Cost: Cost is driven by: Data availability Performance requirements: costs tend to scale super-linearly in the accuracy requirement Problem difficulty: Some of the hard problems include: unsupervised learning, reinforcement learning, and certain categories of supervised learning Full stack pipeline The following figure represents a high level overview of different components in a production level deep learning system: In the following, we will go through each module and recommend toolsets and frameworks as well as best practices from practitioners that fit each component. Data Management 1.1 Data Sources Supervised deep learning requires a lot of labeled data Labeling own data is costly! Here are some resources for data: Open source data (good to start with, but not an advantage) Data augmentation (a MUST for computer vision, an option for NLP) Synthetic data (almost always worth starting with, esp. in NLP) 1.2 Data Labeling Requires: separate software stack (labeling platforms), temporary labor, and QC Sources of labor for labeling: Crowdsourcing (Mechanical Turk): cheap and scalable, less reliable, needs QC Hiring own annotators: less QC needed, expensive, slow to scale Data labeling service companies: FigureEight Labeling platforms: Diffgram: Training Data Software (Computer Vision) Prodigy: An annotation tool powered by active learning (by developers of Spacy), text and image HIVE: AI as a Service platform for computer vision Supervisely: entire computer vision platform Labelbox: computer vision Scale AI data platform (computer vision & NLP) 1.3. Data Storage Data storage options: Object store: Store binary data (images, sound files, compressed texts) Amazon S3 Ceph Object Store Database: Store metadata (file paths, labels, user activity, etc). Postgres is the right choice for most of applications, with the best-in-class SQL and great support for unstructured JSON. Data Lake: to aggregate features which are not obtainable from database (e.g. logs) Amazon Redshift Feature Store: store, access, and share machine learning features (Feature extraction could be computationally expensive and nearly impossible to scale, hence re-using features by different models and teams is a key to high performance ML teams). FEAST (Google cloud, Open Source) Michelangelo Palette (Uber) Suggestion: At training time, copy data into a local or networked filesystem (NFS). 1 1.4. Data Versioning It's a "MUST" for deployed ML models: Deployed ML models are part code, part data. 1 No data versioning means no model versioning. Data versioning platforms: DVC: Open source version control system for ML projects Pachyderm: version control for data Dolt: a SQL database with Git-like version control for data and schema 1.5. Data Processing Training data for production models may come from different sources, including Stored data in db and object stores, log processing, and outputs of other classifiers*. There are dependencies between tasks, each needs to be kicked off after its dependencies are finished. For example, training on new log data, requires a preprocessing step before training. Makefiles are not scalable. "Workflow manager"s become pretty essential in this regard. Workflow orchestration: Luigi by Spotify Airflow by Airbnb: Dynamic, extensible, elegant, and scalable (the most widely used) DAG workflow Robust conditional execution: retry in case of failure Pusher supports docker images with tensorflow serving Whole workflow in a single .py file Development, Training, and Evaluation 2.1. Software engineering Winner language: Python Editors: Vim Emacs VS Code (Recommended by the author): Built-in git staging and diff, Lint code, open projects remotely through ssh Notebooks: Great as starting point of the projects, hard to scale (fun fact: Netflix’s Notebook-Driven Architecture is an exception, which is entirely based on nteract suites). nteract: a next-gen React-based UI for Jupyter notebooks Papermill: is an nteract library built for parameterizing, executing, and analyzing* Jupyter Notebooks. Commuter: another nteract project which provides a read-only display of notebooks (e.g. from S3 buckets). Streamlit: interactive data science tool with applets Compute recommendations 1: For individuals or startups*: Development: a 4x Turing-architecture PC Training/Evaluation: Use the same 4x GPU PC. When running many experiments, either buy shared servers or use cloud instances. For large companies:* Development: Buy a 4x Turing-architecture PC per ML scientist or let them use V100 instances Training/Evaluation: Use cloud instances with proper provisioning and handling of failures Cloud Providers: GCP: option to connect GPUs to any instance + has TPUs AWS: 2.2. Resource Management Allocating free resources to programs Resource management options: Old school cluster job scheduler ( e.g. Slurm workload manager ) Docker + Kubernetes Kubeflow Polyaxon (paid features) 2.3. DL Frameworks Unless having a good reason not to, use Tensorflow/Keras or PyTorch. 1 The following figure shows a comparison between different frameworks on how they stand for "developement" and "production"*. 2.4. Experiment management Development, training, and evaluation strategy: Always start simple Train a small model on a small batch. Only if it works, scale to larger data and models, and hyperparameter tuning! Experiment management tools: Tensorboard provides the visualization and tooling needed for ML experimentation Losswise (Monitoring for ML) Comet: lets you track code, experiments, and results on ML projects Weights & Biases: Record and visualize every detail of your research with easy collaboration MLFlow Tracking: for logging parameters, code versions, metrics, and output files as well as visualization of the results. Automatic experiment tracking with one line of code in python Side by side comparison of experiments Hyper parameter tuning Supports Kubernetes based jobs 2.5. Hyperparameter Tuning Approaches: Grid search Random search Bayesian Optimization HyperBand and Asynchronous Successive Halving Algorithm (ASHA) Population-based Training Platforms: RayTune: Ray Tune is a Python library for hyperparameter tuning at any scale (with a focus on deep learning and deep reinforcement learning). Supports any machine learning framework, including PyTorch, XGBoost, MXNet, and Keras. Katib: Kubernete's Native System for Hyperparameter Tuning and Neural Architecture Search, inspired by Google vizier and supports multiple ML/DL frameworks (e.g. TensorFlow, MXNet, and PyTorch). Hyperas: a simple wrapper around hyperopt for Keras, with a simple template notation to define hyper-parameter ranges to tune. SIGOPT: a scalable, enterprise-grade optimization platform Sweeps from [Weights & Biases] (https://www.wandb.com/): Parameters are not explicitly specified by a developer. Instead they are approximated and learned by a machine learning model. Keras Tuner: A hyperparameter tuner for Keras, specifically for tf.keras with TensorFlow 2.0. 2.6. Distributed Training Data parallelism: Use it when iteration time is too long (both tensorflow and PyTorch support) Ray Distributed Training Model parallelism: when model does not fit on a single GPU Other solutions: Horovod Troubleshooting [TBD] Testing and Deployment 4.1. Testing and CI/CD Machine Learning production software requires a more diverse set of test suites than traditional software: Unit and Integration Testing: Types of tests: Training system tests: testing training pipeline Validation tests: testing prediction system on validation set Functionality tests: testing prediction system on few important examples Continuous Integration: Running tests after each new code change pushed to the repo SaaS for continuous integration: Argo: Open source Kubernetes native workflow engine for orchestrating parallel jobs (incudes workflows, events, CI and CD). CircleCI: Language-Inclusive Support, Custom Environments, Flexible Resource Allocation, used by instacart, Lyft, and StackShare. Travis CI Buildkite: Fast and stable builds, Open source agent runs on almost any machine and architecture, Freedom to use your own tools and services Jenkins: Old school build system 4.2. Web Deployment Consists of a Prediction System and a Serving System Prediction System: Process input data, make predictions Serving System (Web server): Serve prediction with scale in mind Use REST API to serve prediction HTTP requests Calls the prediction system to respond Serving options: Deploy to VMs, scale by adding instances Deploy as containers, scale via orchestration Containers Docker Container Orchestration: Kubernetes (the most popular now) MESOS Marathon Deploy code as a "serverless function" Deploy via a model serving solution Model serving: Specialized web deployment for ML models Batches request for GPU inference Frameworks: Tensorflow serving MXNet Model server Clipper (Berkeley) SaaS solutions Seldon: serve and scale models built in any framework on Kubernetes Algorithmia Decision making: CPU or GPU? CPU inference: CPU inference is preferable if it meets the requirements. Scale by adding more servers, or going serverless. GPU inference: TF serving or Clipper Adaptive batching is useful (Bonus) Deploying Jupyter Notebooks: Kubeflow Fairing is a hybrid deployment package that let's you deploy your Jupyter notebook* codes! 4.5 Service Mesh and Traffic Routing Transition from monolithic applications towards a distributed microservice architecture could be challenging. A Service mesh (consisting of a network of microservices) reduces the complexity of such deployments, and eases the strain on development teams. Istio: a service mesh to ease creation of a network of deployed services with load balancing, service-to-service authentication, monitoring, with few or no code changes in service code. 4.4. Monitoring: Purpose of monitoring: Alerts for downtime, errors, and distribution shifts Catching service and data regressions Cloud providers solutions are decent Kiali:an observability console for Istio with service mesh configuration capabilities. It answers these questions: How are the microservices connected? How are they performing? Are we done? 4.5. Deploying on Embedded and Mobile Devices Main challenge: memory footprint and compute constraints Solutions: Quantization Reduced model size MobileNets Knowledge Distillation DistillBERT (for NLP) Embedded and Mobile Frameworks: Tensorflow Lite PyTorch Mobile Core ML ML Kit FRITZ OpenVINO Model Conversion: Open Neural Network Exchange (ONNX): open-source format for deep learning models 4.6. All-in-one solutions Tensorflow Extended (TFX) Michelangelo (Uber) Google Cloud AI Platform Amazon SageMaker Neptune FLOYD Paperspace Determined AI Domino data lab Tensorflow Extended (TFX) [TBD] Airflow and KubeFlow ML Pipelines [TBD] Other useful links: Lessons learned from building practical deep learning systems Machine Learning: The High Interest Credit Card of Technical Debt Contributing References: [1]: Full Stack Deep Learning Bootcamp, Nov 2019. [2]: Advanced KubeFlow Workshop by Pipeline.ai, 2019. [3]: TFX: Real World Machine Learning in Production

RD-Agent
github
LLM Vibe Score0.548
Human Vibe Score0.27921589729164453
microsoftMar 28, 2025

RD-Agent

🖥️ Live Demo | 🎥 Demo Video ▶️YouTube | 📖 Documentation | 📃 Papers Data Science Agent Preview Check out our demo video showcasing the current progress of our Data Science Agent under development: https://github.com/user-attachments/assets/3eccbecb-34a4-4c81-bce4-d3f8862f7305 📰 News | 🗞️ News | 📝 Description | | -- | ------ | | Support LiteLLM Backend | We now fully support LiteLLM as a backend for integration with multiple LLM providers. | | More General Data Science Agent | 🚀Coming soon! | | Kaggle Scenario release | We release Kaggle Agent, try the new features! | | Official WeChat group release | We created a WeChat group, welcome to join! (🗪QR Code) | | Official Discord release | We launch our first chatting channel in Discord (🗪) | | First release | RDAgent is released on GitHub | 🌟 Introduction RDAgent aims to automate the most critical and valuable aspects of the industrial R&D process, and we begin with focusing on the data-driven scenarios to streamline the development of models and data. Methodologically, we have identified a framework with two key components: 'R' for proposing new ideas and 'D' for implementing them. We believe that the automatic evolution of R&D will lead to solutions of significant industrial value. R&D is a very general scenario. The advent of RDAgent can be your 💰 Automatic Quant Factory (🎥Demo Video|▶️YouTube) 🤖 Data Mining Agent: Iteratively proposing data & models (🎥Demo Video 1|▶️YouTube) (🎥Demo Video 2|▶️YouTube) and implementing them by gaining knowledge from data. 🦾 Research Copilot: Auto read research papers (🎥Demo Video|▶️YouTube) / financial reports (🎥Demo Video|▶️YouTube) and implement model structures or building datasets. 🤖 Kaggle Agent: Auto Model Tuning and Feature Engineering([🎥Demo Video Coming Soon...]()) and implementing them to achieve more in competitions. ... You can click the links above to view the demo. We're continuously adding more methods and scenarios to the project to enhance your R&D processes and boost productivity. Additionally, you can take a closer look at the examples in our 🖥️ Live Demo. ⚡ Quick start You can try above demos by running the following command: 🐳 Docker installation. Users must ensure Docker is installed before attempting most scenarios. Please refer to the official 🐳Docker page for installation instructions. Ensure the current user can run Docker commands without using sudo. You can verify this by executing docker run hello-world. 🐍 Create a Conda Environment Create a new conda environment with Python (3.10 and 3.11 are well-tested in our CI): Activate the environment: 🛠️ Install the RDAgent You can directly install the RDAgent package from PyPI: 💊 Health check rdagent provides a health check that currently checks two things. whether the docker installation was successful. whether the default port used by the rdagent ui is occupied. ⚙️ Configuration The demos requires following ability: ChatCompletion json_mode embedding query For example: If you are using the OpenAI API, you have to configure your GPT model in the .env file like this. However, not every API services support these features by default. For example: AZURE OpenAI, you have to configure your GPT model in the .env file like this. We now support LiteLLM as a backend for integration with multiple LLM providers. If you use LiteLLM Backend to use models, you can configure as follows: For more configuration information, please refer to the documentation. 🚀 Run the Application The 🖥️ Live Demo is implemented by the following commands(each item represents one demo, you can select the one you prefer): Run the Automated Quantitative Trading & Iterative Factors Evolution: Qlib self-loop factor proposal and implementation application Run the Automated Quantitative Trading & Iterative Model Evolution: Qlib self-loop model proposal and implementation application Run the Automated Medical Prediction Model Evolution: Medical self-loop model proposal and implementation application (1) Apply for an account at PhysioNet. (2) Request access to FIDDLE preprocessed data: FIDDLE Dataset. (3) Place your username and password in .env. Run the Automated Quantitative Trading & Factors Extraction from Financial Reports: Run the Qlib factor extraction and implementation application based on financial reports Run the Automated Model Research & Development Copilot: model extraction and implementation application Run the Automated Kaggle Model Tuning & Feature Engineering: self-loop model proposal and feature engineering implementation application Using sf-crime (San Francisco Crime Classification) as an example. Register and login on the Kaggle website. Configuring the Kaggle API. (1) Click on the avatar (usually in the top right corner of the page) -> Settings -> Create New Token, A file called kaggle.json will be downloaded. (2) Move kaggle.json to ~/.config/kaggle/ (3) Modify the permissions of the kaggle.json file. Reference command: chmod 600 ~/.config/kaggle/kaggle.json Join the competition: Click Join the competition -> I Understand and Accept at the bottom of the competition details page. Description of the above example: Kaggle competition data, contains two parts: competition description file (json file) and competition dataset (zip file). We prepare the competition description file for you, the competition dataset will be downloaded automatically when you run the program, as in the example. If you want to download the competition description file automatically, you need to install chromedriver, The instructions for installing chromedriver can be found in the documentation. The Competition List Available can be found here. 🖥️ Monitor the Application Results You can run the following command for our demo program to see the run logs. Note: Although port 19899 is not commonly used, but before you run this demo, you need to check if port 19899 is occupied. If it is, please change it to another port that is not occupied. You can check if a port is occupied by running the following command. 🏭 Scenarios We have applied RD-Agent to multiple valuable data-driven industrial scenarios. 🎯 Goal: Agent for Data-driven R&D In this project, we are aiming to build an Agent to automate Data-Driven R\&D that can 📄 Read real-world material (reports, papers, etc.) and extract key formulas, descriptions of interested features and models, which are the key components of data-driven R&D . 🛠️ Implement the extracted formulas (e.g., features, factors, and models) in runnable codes. Due to the limited ability of LLM in implementing at once, build an evolving process for the agent to improve performance by learning from feedback and knowledge. 💡 Propose new ideas based on current knowledge and observations. 📈 Scenarios/Demos In the two key areas of data-driven scenarios, model implementation and data building, our system aims to serve two main roles: 🦾Copilot and 🤖Agent. The 🦾Copilot follows human instructions to automate repetitive tasks. The 🤖Agent, being more autonomous, actively proposes ideas for better results in the future. The supported scenarios are listed below: | Scenario/Target | Model Implementation | Data Building | | -- | -- | -- | | 💹 Finance | 🤖 Iteratively Proposing Ideas & Evolving▶️YouTube | 🤖 Iteratively Proposing Ideas & Evolving ▶️YouTube 🦾 Auto reports reading & implementation▶️YouTube | | 🩺 Medical | 🤖 Iteratively Proposing Ideas & Evolving▶️YouTube | - | | 🏭 General | 🦾 Auto paper reading & implementation▶️YouTube 🤖 Auto Kaggle Model Tuning | 🤖Auto Kaggle feature Engineering | RoadMap: Currently, we are working hard to add new features to the Kaggle scenario. Different scenarios vary in entrance and configuration. Please check the detailed setup tutorial in the scenarios documents. Here is a gallery of successful explorations (5 traces showed in 🖥️ Live Demo). You can download and view the execution trace using this command from the documentation. Please refer to 📖readthedocs_scen for more details of the scenarios. ⚙️ Framework Automating the R&D process in data science is a highly valuable yet underexplored area in industry. We propose a framework to push the boundaries of this important research field. The research questions within this framework can be divided into three main categories: | Research Area | Paper/Work List | |--------------------|-----------------| | Benchmark the R&D abilities | Benchmark | | Idea proposal: Explore new ideas or refine existing ones | Research | | Ability to realize ideas: Implement and execute ideas | Development | We believe that the key to delivering high-quality solutions lies in the ability to evolve R&D capabilities. Agents should learn like human experts, continuously improving their R&D skills. More documents can be found in the 📖 readthedocs. 📃 Paper/Work list 📊 Benchmark Towards Data-Centric Automatic R&D !image 🔍 Research In a data mining expert's daily research and development process, they propose a hypothesis (e.g., a model structure like RNN can capture patterns in time-series data), design experiments (e.g., finance data contains time-series and we can verify the hypothesis in this scenario), implement the experiment as code (e.g., Pytorch model structure), and then execute the code to get feedback (e.g., metrics, loss curve, etc.). The experts learn from the feedback and improve in the next iteration. Based on the principles above, we have established a basic method framework that continuously proposes hypotheses, verifies them, and gets feedback from the real-world practice. This is the first scientific research automation framework that supports linking with real-world verification. For more detail, please refer to our 🖥️ Live Demo page. 🛠️ Development Collaborative Evolving Strategy for Automatic Data-Centric Development !image 🤝 Contributing We welcome contributions and suggestions to improve RD-Agent. Please refer to the Contributing Guide for more details on how to contribute. Before submitting a pull request, ensure that your code passes the automatic CI checks. 📝 Guidelines This project welcomes contributions and suggestions. Contributing to this project is straightforward and rewarding. Whether it's solving an issue, addressing a bug, enhancing documentation, or even correcting a typo, every contribution is valuable and helps improve RDAgent. To get started, you can explore the issues list, or search for TODO: comments in the codebase by running the command grep -r "TODO:". Before we released RD-Agent as an open-source project on GitHub, it was an internal project within our group. Unfortunately, the internal commit history was not preserved when we removed some confidential code. As a result, some contributions from our group members, including Haotian Chen, Wenjun Feng, Haoxue Wang, Zeqi Ye, Xinjie Shen, and Jinhui Li, were not included in the public commits. ⚖️ Legal disclaimer The RD-agent is provided “as is”, without warranty of any kind, express or implied, including but not limited to the warranties of merchantability, fitness for a particular purpose and noninfringement. The RD-agent is aimed to facilitate research and development process in the financial industry and not ready-to-use for any financial investment or advice. Users shall independently assess and test the risks of the RD-agent in a specific use scenario, ensure the responsible use of AI technology, including but not limited to developing and integrating risk mitigation measures, and comply with all applicable laws and regulations in all applicable jurisdictions. The RD-agent does not provide financial opinions or reflect the opinions of Microsoft, nor is it designed to replace the role of qualified financial professionals in formulating, assessing, and approving finance products. The inputs and outputs of the RD-agent belong to the users and users shall assume all liability under any theory of liability, whether in contract, torts, regulatory, negligence, products liability, or otherwise, associated with use of the RD-agent and any inputs and outputs thereof.

LLMStack
github
LLM Vibe Score0.535
Human Vibe Score0.022778788676674117
trypromptlyMar 28, 2025

LLMStack

LLMStack is a no-code platform for building generative AI agents, workflows and chatbots, connecting them to your data and business processes. Quickstart | Documentation | Promptly Overview Build tailor-made generative AI agents, applications and chatbots that cater to your unique needs by chaining multiple LLMs. Seamlessly integrate your own data, internal tools and GPT-powered models without any coding experience using LLMStack's no-code builder. Trigger your AI chains from Slack or Discord. Deploy to the cloud or on-premise. !llmstack-quickstart See full demo video here Getting Started Check out our Cloud offering at Promptly or follow the instructions below to deploy LLMStack on your own infrastructure. LLMStack deployment comes with a default admin account whose credentials are admin and promptly. Be sure to change the password from admin panel after logging in. Installation Prerequisites LLMStack depends on a background docker container to run jobs. Make sure you have Docker installed on your machine if want to use jobs. You can follow the instructions here to install Docker. Install LLMStack using pip If you are on windows, please use WSL2 (Windows Subsystem for Linux) to install LLMStack. You can follow the instructions here to install WSL2. Once you are in a WSL2 terminal, you can install LLMStack using the above command. Start LLMStack using the following command: Above commands will install and start LLMStack. It will create .llmstack in your home directory and places the database and config files in it when run for the first time. Once LLMStack is up and running, it should automatically open your browser and point it to localhost:3000. You can add your own keys to providers like OpenAI, Cohere, Stability etc., from Settings page. If you want to provide default keys for all the users of your LLMStack instance, you can add them to the ~/.llmstack/config file. LLMStack: Quickstart video Features 🤖 Agents: Build generative AI agents like AI SDRs, Research Analysts, RPA Automations etc., without writing any code. Connect agents to your internal or external tools, search the web or browse the internet with agents. 🔗 Chain multiple models: LLMStack allows you to chain multiple LLMs together to build complex generative AI applications. 📊 Use generative AI on your Data: Import your data into your accounts and use it in AI chains. LLMStack allows importing various types (CSV, TXT, PDF, DOCX, PPTX etc.,) of data from a variety of sources (gdrive, notion, websites, direct uploads etc.,). Platform will take care of preprocessing and vectorization of your data and store it in the vector database that is provided out of the box. 🛠️ No-code builder: LLMStack comes with a no-code builder that allows you to build AI chains without any coding experience. You can chain multiple LLMs together and connect them to your data and business processes. ☁️ Deploy to the cloud or on-premise: LLMStack can be deployed to the cloud or on-premise. You can deploy it to your own infrastructure or use our cloud offering at Promptly. 🚀 API access: Apps or chatbots built with LLMStack can be accessed via HTTP API. You can also trigger your AI chains from Slack or Discord. 🏢 Multi-tenant: LLMStack is multi-tenant. You can create multiple organizations and add users to them. Users can only access the data and AI chains that belong to their organization. What can you build with LLMStack? Using LLMStack you can build a variety of generative AI applications, chatbots and agents. Here are some examples: 👩🏻‍💼 AI SDRs: You can build AI SDRs (Sales Development Representatives) that can generate personalized emails, LinkedIn messages, cold calls, etc., for your sales team 👩🏻‍💻 Research Analysts: You can build AI Research Analysts that can generate research reports, investment thesis, etc., for your investment team 🤖 RPA Automations: You can build RPA automations that can automate your business processes by generating emails, filling forms, etc., 📝 Text generation: You can build apps that generate product descriptions, blog posts, news articles, tweets, emails, chat messages, etc., by using text generation models and optionally connecting your data. Check out this marketing content generator for example 🤖 Chatbots: You can build chatbots trained on your data powered by ChatGPT like Promptly Help that is embedded on Promptly website 🎨 Multimedia generation: Build complex applications that can generate text, images, videos, audio, etc. from a prompt. This story generator is an example 🗣️ Conversational AI: Build conversational AI systems that can have a conversation with a user. Check out this Harry Potter character chatbot 🔍 Search augmentation: Build search augmentation systems that can augment search results with additional information using APIs. Sharebird uses LLMStack to augment search results with AI generated answer from their content similar to Bing's chatbot 💬 Discord and Slack bots: Apps built on LLMStack can be triggered from Slack or Discord. You can easily connect your AI chains to Slack or Discord from LLMStack's no-code app editor. Check out our Discord server to interact with one such bot. Administration Login to http://localhost:3000/admin using the admin account. You can add users and assign them to organizations in the admin panel. Cloud Offering Check out our cloud offering at Promptly. You can sign up for a free account and start building your own generative AI applications. Documentation Check out our documentation at docs.trypromptly.com/llmstack to learn more about LLMStack. Development Check out our development guide at docs.trypromptly.com/llmstack/development to learn more about how to run and develop LLMStack. Contributing We welcome contributions to LLMStack. Please check out our contributing guide to learn more about how you can contribute to LLMStack.

airspace-hugo
github
LLM Vibe Score0.551
Human Vibe Score0.45061592683949336
themefisherMar 28, 2025

airspace-hugo

Airspace Hugo This theme is suitable for a wide variety of businesses, including marketing, photography, and development enterprises. 👀Demo | Page Speed (95%)🚀 🔑Key Features 📄 9+ Pre-Designed Pages 🌐 Multiple language support (Fr, En) 📊 Google Analytics support 🎨 CSS and JS bundle with Hugo Pipe 🎨 Bootstrap Based ⚙️ Netlify settings predefine 👥 Multiple authors available ✉️ Contact form support 🔄 GDPR consent support 🗺️ Google Maps support 🎉 Fun factors counter 🚀 Google Page Speed optimized 🌐 Open Graph meta tag 🐦 Twitter Card meta tag 📄 9+ Pre-Designed Pages 🏠 Home Page 📚 Blog Page 📝 Blog Single Page 📄 Project Page 🛠️ Services 💰 Pricing ❓ FAQ ℹ️ About Page 📞 Contact Page 🖥️Local development Or Check out Full Documentation. ⚙️Deployment and hosting Follow the steps. 🐞Reporting Issues We use GitHub Issues as the official bug tracker for the Airspace Template. Please Search existing issues. Someone may have already reported the same problem. If your problem or idea has not been addressed yet, feel free to open a new issue. 📱Submit Your Website To Our Showcase Are you using Airspace Hugo theme? Submit it to our showcase. Our showcase aims to demonstrate to the world what amazing websites people like you have created utilizing our Hugo themes and to show that Hugo has tremendous capabilities as a Static Site Generator. View all the websites powered by Airspace Hugo from here. Submit your Airspace Hugo powered website. 📄License Copyright &copy; Designed by Themefisher & Developed by Gethugothemes Code License: Released under the MIT license. Image license: The images are only for demonstration purposes. They have their licenses. We don't have permission to share those images. 🙏Special Thanks Bootstrap Jquery Ionicons Magnific Popup Shuffle Slick Slider Google Fonts All Contributors 👨‍💻Hire Us Besides developing unique, blazing-fast Hugo themes, we also provide customized services. We specialize in creating affordable, high-quality static websites based on Hugo. If you need to customize the theme or complete website development from scratch, you can hire us. Check Our Services 💎Premium Themes By Us | | | | |:---:|:---:|:---:| | Get 55+ Premium Hugo Themes Bundle | Bigspring | Navigator |

awesome-ai-in-finance
github
LLM Vibe Score0.58
Human Vibe Score1
georgezouqMar 28, 2025

awesome-ai-in-finance

Awesome AI in Finance There are millions of trades made in the global financial market every day. Data grows very quickly and people are hard to understand. With the power of the latest artificial intelligence research, people analyze & trade automatically and intelligently. This list contains the research, tools and code that people use to beat the market. [中文资源] Contents LLMs Papers Courses & Books Strategies & Research Time Series Data Portfolio Management High Frequency Trading Event Drive Crypto Currencies Strategies Technical Analysis Lottery & Gamble Arbitrage Data Sources Research Tools Trading System TA Lib Exchange API Articles Others LLMs 🌟🌟 MarS - A Financial Market Simulation Engine Powered by Generative Foundation Model. 🌟🌟 Financial Statement Analysis with Large Language Models - GPT-4 can outperform professional financial analysts in predicting future earnings changes, generating useful narrative insights, and resulting in superior trading strategies with higher Sharpe ratios and alphas, thereby suggesting a potential central role for LLMs in financial decision-making. PIXIU - An open-source resource providing a financial large language model, a dataset with 136K instruction samples, and a comprehensive evaluation benchmark. FinGPT - Provides a playground for all people interested in LLMs and NLP in Finance. MACD + RSI + ADX Strategy (ChatGPT-powered) by TradeSmart - Asked ChatGPT on which indicators are the most popular for trading. We used all of the recommendations given. A ChatGPT trading algorithm delivered 500% returns in stock market. My breakdown on what this means for hedge funds and retail investors Use chatgpt to adjust strategy parameters Hands-on LLMs: Train and Deploy a Real-time Financial Advisor - Train and deploy a real-time financial advisor chatbot with Falcon 7B and CometLLM. ChatGPT Strategy by OctoBot - Use ChatGPT to determine which cryptocurrency to trade based on technical indicators. Papers The Theory of Speculation L. Bachelier, 1900 - The influences which determine the movements of the Stock Exchange are. Brownian Motion in the Stock Market Osborne, 1959 - The common-stock prices can be regarded as an ensemble of decisions in statistical equilibrium. An Investigation into the Use of Reinforcement Learning Techniques within the Algorithmic Trading Domain, 2015 A Deep Reinforcement Learning Framework for the Financial Portfolio Management Problem Reinforcement Learning for Trading, 1994 Dragon-Kings, Black Swans and the Prediction of Crises Didier Sornette - The power laws in the distributions of event sizes under a broad range of conditions in a large variety of systems. Financial Trading as a Game: A Deep Reinforcement Learning Approach - Deep reinforcement learning provides a framework toward end-to-end training of such trading agent. Machine Learning for Trading - With an appropriate choice of the reward function, reinforcement learning techniques can successfully handle the risk-averse case. Ten Financial Applications of Machine Learning, 2018 - Slides review few important financial ML applications. FinRL: A Deep Reinforcement Learning Library for Automated Stock Trading in Quantitative Finance, 2020 - Introduce a DRL library FinRL that facilitates beginners to expose themselves to quantitative finance and to develop their own stock trading strategies. Deep Reinforcement Learning for Automated Stock Trading: An Ensemble Strategy, 2020 - Propose an ensemble strategy that employs deep reinforcement schemes to learn a stock trading strategy by maximizing investment return. Courses & Books & Blogs 🌟 QuantResearch - Quantitative analysis, strategies and backtests https://letianzj.github.io/ NYU: Overview of Advanced Methods of Reinforcement Learning in Finance Udacity: Artificial Intelligence for Trading AI in Finance - Learn Fintech Online. Advanced-Deep-Trading - Experiments based on "Advances in financial machine learning" book. Advances in Financial Machine Learning - Using advanced ML solutions to overcome real-world investment problems. Build Financial Software with Generative AI - Book about how to build financial software hands-on using generative AI tools like ChatGPT and Copilot. Mastering Python for Finance - Sources codes for: Mastering Python for Finance, Second Edition. MLSys-NYU-2022 - Slides, scripts and materials for the Machine Learning in Finance course at NYU Tandon, 2022. Train and Deploy a Serverless API to predict crypto prices - In this tutorial you won't build an ML system that will make you rich. But you will master the MLOps frameworks and tools you need to build ML systems that, together with tons of experimentation, can take you there. Strategies & Research Time Series Data Price and Volume process with Technology Analysis Indices 🌟🌟 stockpredictionai - A complete process for predicting stock price movements. 🌟 Personae - Implements and environment of Deep Reinforcement Learning & Supervised Learning for Quantitative Trading. 🌟 Ensemble-Strategy - Deep Reinforcement Learning for Automated Stock Trading. FinRL - A Deep Reinforcement Learning Library for Automated Stock Trading in Quantitative Finance. AutomatedStockTrading-DeepQ-Learning - Build a Deep Q-learning reinforcement agent model as automated trading robot. tfdeeprltrader - Trading environment(OpenAI Gym) + PPO(TensorForce). trading-gym - Trading agent to train with episode of short term trading itself. trading-rl - Deep Reinforcement Learning for Financial Trading using Price Trailing. deeprltrader - Trading environment(OpenAI Gym) + DDQN (Keras-RL). Quantitative-Trading - Papers and code implementing Quantitative-Trading. gym-trading - Environment for reinforcement-learning algorithmic trading models. zenbrain - A framework for machine-learning bots. DeepLearningNotes - Machine learning in quant analysis. stockmarketreinforcementlearning - Stock market trading OpenAI Gym environment with Deep Reinforcement Learning using Keras. Chaos Genius - ML powered analytics engine for outlier/anomaly detection and root cause analysis.. mlforecast - Scalable machine learning based time series forecasting. Portfolio Management Deep-Reinforcement-Stock-Trading - A light-weight deep reinforcement learning framework for portfolio management. qtrader - Reinforcement Learning for portfolio management. PGPortfolio - A Deep Reinforcement Learning framework for the financial portfolio management problem. DeepDow - Portfolio optimization with deep learning. skfolio - Python library for portfolio optimization built on top of scikit-learn. High Frequency Trading High-Frequency-Trading-Model-with-IB - A high-frequency trading model using Interactive Brokers API with pairs and mean-reversion. 🌟 SGX-Full-OrderBook-Tick-Data-Trading-Strategy - Solutions for high-frequency trading (HFT) strategies using data science approaches (Machine Learning) on Full Orderbook Tick Data. HFTBitcoin - Analysis of High Frequency Trading on Bitcoin exchanges. Event Drive 🌟🌟 stockpredictionai - Complete process for predicting stock price movements. 🌟 trump2cash - A stock trading bot powered by Trump tweets. Crypto Currencies Strategies LSTM-Crypto-Price-Prediction - Predicting price trends in crypto markets using an LSTM-RNN for trading. tforcebtctrader - TensorForce Bitcoin trading bot. Tensorflow-NeuroEvolution-Trading-Bot - A population model that trade cyrpto and breed and mutate iteratively. gekkoga - Genetic algorithm for solving optimization of trading strategies using Gekko. GekkoANNStrategies - ANN trading strategies for the Gekko trading bot. gekko-neuralnet - Neural network strategy for Gekko. bitcoinprediction - Code for "Bitcoin Prediction" by Siraj Raval on YouTube. Technical Analysis quant-trading - Python quantitative trading strategies. Gekko-Bot-Resources - Gekko bot resources. gekkotools - Gekko strategies, tools etc. gekko RSIWR - Gekko RSIWR strategies. gekko HL - Calculate down peak and trade on. EthTradingAlgorithm - Ethereum trading algorithm using Python 3.5 and the library ZipLine. gekkotradingstuff - Awesome crypto currency trading platform. forex.analytics - Node.js native library performing technical analysis over an OHLC dataset with use of genetic algorithmv. BitcoinMACDStrategy - Bitcoin MACD crossover trading strategy backtest. crypto-signal - Automated crypto trading & technical analysis (TA) bot for Bittrex, Binance, GDAX, and more. Gekko-Strategies - Strategies to Gekko trading bot with backtests results and some useful tools. gekko-gannswing - Gann's Swing trade strategy for Gekko trade bot. Lottery & Gamble LotteryPredict - Use LSTM to predict lottery. Arbitrage ArbitrageBot - Arbitrage bot that currently works on bittrex & poloniex. r2 - Automatic arbitrage trading system powered by Node.js + TypeScript. cryptocurrency-arbitrage - A crypto currency arbitrage opportunity calculator. Over 800 currencies and 50 markets. bitcoin-arbitrage - Bitcoin arbitrage opportunity detector. blackbird - Long / short market-neutral strategy. Data Sources Traditional Markets 🌟 Quandl - Get millions of financial and economic dataset from hundreds of publishers via a single free API. yahoo-finance - Python module to get stock data from Yahoo! Finance. Tushare - Crawling historical data of Chinese stocks. Financial Data - Stock Market and Financial Data API. Crypto Currencies CryptoInscriber - A live crypto currency historical trade data blotter. Download live historical trade data from any crypto exchange. Gekko-Datasets - Gekko trading bot dataset dumps. Download and use history files in SQLite format. Research Tools Synthical - AI-powered collaborative environment for Research. 🌟🌟 TensorTrade - Trade efficiently with reinforcement learning. ML-Quant - Quant resources from ArXiv (sanity), SSRN, RePec, Journals, Podcasts, Videos, and Blogs. JAQS - An open source quant strategies research platform. pyfolio - Portfolio and risk analytics in Python. alphalens - Performance analysis of predictive (alpha) stock factors. empyrical - Common financial risk and performance metrics. Used by Zipline and pyfolio. zvt - Zero vector trader. Trading System For Back Test & Live trading Traditional Market System 🌟🌟🌟 OpenBB - AI-powered opensource research and analytics workspace. 🌟🌟 zipline - A python algorithmic trading library. 🌟 TradingView - Get real-time information and market insights. rqalpha - A extendable, replaceable Python algorithmic backtest & trading framework. backtrader - Python backtesting library for trading strategies. kungfu - Kungfu Master trading system. lean - Algorithmic trading engine built for easy strategy research, backtesting and live trading. Combine & Rebuild pylivetrader - Python live trade execution library with zipline interface. CoinMarketCapBacktesting - As backtest frameworks for coin trading strategy. Crypto Currencies zenbot - Command-line crypto currency trading bot using Node.js and MongoDB. bot18 - High-frequency crypto currency trading bot developed by Zenbot. magic8bot - Crypto currency trading bot using Node.js and MongoDB. catalyst - An algorithmic trading library for Crypto-Assets in python. QuantResearchDev - Quant Research dev & Traders open source project. MACD - Zenbot MACD Auto-Trader. abu - A quant trading system base on python. Plugins CoinMarketCapBacktesting - Tests bt and Quantopian Zipline as backtesting frameworks for coin trading strategy. Gekko-BacktestTool - Batch backtest, import and strategy params optimalization for Gekko Trading Bot. TA Lib pandastalib - A Python Pandas implementation of technical analysis indicators. finta - Common financial technical indicators implemented in Python-Pandas (70+ indicators). tulipnode - Official Node.js wrapper for Tulip Indicators. Provides over 100 technical analysis overlay and indicator functions. techan.js - A visual, technical analysis and charting (Candlestick, OHLC, indicators) library built on D3. Exchange API Do it in real world! IbPy - Python API for the Interactive Brokers on-line trading system. HuobiFeeder - Connect HUOBIPRO exchange, get market/historical data for ABAT trading platform backtest analysis and live trading. ctpwrapper - Shanghai future exchange CTP api. PENDAX - Javascript SDK for Trading/Data API and Websockets for cryptocurrency exchanges like FTX, FTXUS, OKX, Bybit, & More Framework tf-quant-finance - High-performance TensorFlow library for quantitative finance. Visualizing playground - Play with neural networks. netron - Visualizer for deep learning and machine learning models. KLineChart - Highly customizable professional lightweight financial charts GYM Environment 🌟 TradingGym - Trading and Backtesting environment for training reinforcement learning agent. TradzQAI - Trading environment for RL agents, backtesting and training. btgym - Scalable, event-driven, deep-learning-friendly backtesting library. Articles The-Economist - The Economist. nyu-mlif-notes - NYU machine learning in finance notes. Using LSTMs to Turn Feelings Into Trades Others zipline-tensorboard - TensorBoard as a Zipline dashboard. gekko-quasar-ui - An UI port for gekko trading bot using Quasar framework. Floom AI gateway and marketplace for developers, enables streamlined integration and least volatile approach of AI features into products Other Resource 🌟🌟🌟 Stock-Prediction-Models - Stock-Prediction-Models, Gathers machine learning and deep learning models for Stock forecasting, included trading bots and simulations. 🌟🌟 Financial Machine Learning - A curated list of practical financial machine learning (FinML) tools and applications. This collection is primarily in Python. 🌟 Awesome-Quant-Machine-Learning-Trading - Quant / Algorithm trading resources with an emphasis on Machine Learning. awesome-quant - A curated list of insanely awesome libraries, packages and resources for Quants (Quantitative Finance). FinancePy - A Python Finance Library that focuses on the pricing and risk-management of Financial Derivatives, including fixed-income, equity, FX and credit derivatives. Explore Finance Service Libraries & Projects - Explore a curated list of Fintech popular & new libraries, top authors, trending project kits, discussions, tutorials & learning resources on kandi.

DownEdit
github
LLM Vibe Score0.491
Human Vibe Score0.032913669732192626
nxNullMar 28, 2025

DownEdit

DownEdit is a fast and powerful program for downloading and editing videos from top platforms like TikTok, Douyin, and Kuaishou. Effortlessly grab videos from user profiles, make bulk edits, throughout the entire directory with just one click. Plus, our advanced Chat & AI features let you download, edit, and generate videos, images, and sounds in bulk. Exciting new features are coming soon—stay tuned! ✨ Preview 🔥 Current Features Edit Video: Enhance videos with various functions designed to streamline editing tasks across entire directories. Edit Photo: Quickly enhance images in bulk with various functions, including AI-powered functions, Edit Sound: Improve audio in bulk using powerful functions, including cutting-edge AI-powered tools. Download all videos: Retrieve videos from users (TikTok, Kuaishou, Douyin, etc.) without watermarks. Bulk AI Generator: Generate images and videos in bulk using powerful generative AI. AI Editor: Enhance your content effortlessly with using AI editor designed for images, sounds and videos. 🌐 Service | Website| Provider| Single Video | User's Videos | Stream | Access | Status | | --- | --- | --- | --- | --- | --- | --- | | tiktok.com | None | ✔️ | ✔️ | ❌ | API (Cookie) | !Inactive | | douyin.com | None | ✔️ | ✔️ | ❌ | API (Cookie) | !Inactive | | kuaishou.com | None | ✔️ | ✔️ | ❌ | Login Required (Cookie) | !Active | | youtube.com | None | ✔️ | ✔️ | ❌ | (Public/Private) | !Active | 🤖 AI Cloud | Type | Model | Provider| Minimal | Bulk | Access | Status | | --- | --- | --- | --- | --- | --- | --- | | Image Generation | None | | None | ✔️ | API (Public) | !Active | | Video Generation | None | | None | ✔️ | | !Inactive | | Sound Generation | None | | None | ✔️ | | !Inactive | Local | Type | Model | Provider| Minimal | Bulk | Access | Status | | --- | --- | --- | --- | --- | --- | --- | | Image Generation | None | | None | ✔️ | | !Inactive | | Video Generation | None | | None | ✔️ | | !Inactive | | Sound Generation | None | | None | ✔️ | | !Inactive | 🚀 Usage Edit Video - Simply copy and paste (right click) whatever directory location you would like to process. Tutorial !EditVideoAdobeExpress Change it according to your desired video speed. Input your music file location Download douyin videos - Download all video from user by input user link. Tutorial Download tiktok videos - Download all video from user by input username with @. Tutorial Download kuaishou videos - Remember to input your own Cookie. Otherwise it won't work. Tutorial Step 1. Right click and select on Inspect element. Step 2. Copy your Cookie browser. Step 3. Copy user ID you want to download. Tips: If you still getting error, try changing your Browser, use Incognito/Private mode and reset your Internet/IP. Edit Photo - Simply copy and paste (right click) whatever directory location you would like to process. Tutorial Remove Background AI 🔎 Requirements Python [!NOTE] Version must be between 3.8 and 3.12. ⚙ Installation Step 1. Download and install python on your pc. Step 2. libraries installation You have three options to install the required libraries: Option 1: Manual Installation Option 2: Automatic installation & virtual environments Option 3: Terminal & virtual environments Step 3. Run the script For Regular Use: You can also download the application and use it on your PC without installing python. Windows: Download macOS: None [!TIP] Fix Terminal Font Issues Install the Microsoft Cascadia font on your computer if your terminal does not support the font, which is resulting in program error. 🔨 Module The following dependencies are required for the project: List Pystyle Requests Inquirer Colorama Moviepy Rich Playwright Rembg WMI Psutil Httpx Aiofiles Author 👤 Sokun Heng Github: @SokunHeng Show your support Please ⭐️ this repository if this project helped you! 📚 Reference Documentation 📝 License Copyright © 2022 SokunHeng.

eiten
github
LLM Vibe Score0.549
Human Vibe Score0.754375921646308
tradyticsMar 27, 2025

eiten

Eiten - Algorithmic Investing Strategies for Everyone Eiten is an open source toolkit by Tradytics that implements various statistical and algorithmic investing strategies such as Eigen Portfolios, Minimum Variance Portfolios, Maximum Sharpe Ratio Portfolios, and Genetic Algorithms based Portfolios. It allows you to build your own portfolios with your own set of stocks that can beat the market. The rigorous testing framework included in Eiten enables you to have confidence in your portfolios. If you are looking to discuss these tools in depth and talk about more tools that we are working on, please feel free to join our Discord channel where we have a bunch of more tools too. Files Description | Path | Description | :--- | :---------- | eiten | Main folder. | &boxur; figures | Figures for this github repositories. | &boxur; stocks | Folder to keep your stock lists that you want to use to create your portfolios. | &boxur; strategies | A bunch of strategies implemented in python. | backtester.py | Backtesting module that both backtests and forward tests all portfolios. | data_loader.py | Module for loading data from yahoo finance. | portfolio_manager.py | Main file that takes in a bunch of arguments and generates several portfolios for you. | simulator.py | Simulator that uses historical returns and monte carlo to simulate future prices for the portfolios. | strategy_manager.py | Manages the strategies implemented in the 'strategies' folder. Required Packages You will need to install the following package to train and test the models. Scikit-learn Numpy Tqdm Yfinance Pandas Scipy You can install all packages using the following command. Please note that the script was written using python3. Build your portfolios Let us see how we can use all the strategies given in the toolkit to build our portfolios. The first thing you need to do is modify the stocks.txt file in the stocks folder and add the stocks of your choice. It is recommended to keep the list small i.e anywhere between 5 to 50 stocks should be fine. We have already put a small stocks list containing a bunch of tech stocks like AAPL, MSFT, TSLA etc. Let us build our portfolios now. This is the main command that you need to run. This command will use last 5 years of daily data excluding the last 90 days and build several portfolios for you. Based on those portfolios, it will then test them on the out of sample data of 90 days and show you the performance of each portfolio. Finally, it will also compare the performance with your choice of market index which is QQQ here. Let's dive into each of the parameters in detail. istest: The value determined if the program is going to keep some separate data for future testing. When this is enabled, the value of futurebars should be larger than 5. future_bars: These are the bars that the tool will exclude during portfolio building and will forward test the portfolios on the excluded set. This is also called out of sample data. datagranularityminutes: How much granular data do you want to use to build your portfolios. For long term portfolios, you should use daily data but for short term, you can use hourly or minute level data. The possible values here are 3600, 60, 30, 15, 5, 1. 3600 means daily. historytouse: Whether to use a specific number of historical bars or use everything that we receive from yahoo finance. For minute level data, we only receive up to one month of historical data. For daily, we receive 5 years worth of historical data. If you want to use all available data, the value should be all but if you want to use smaller history, you can set it to an integer value e.g 100 which will only use the last 100 bars to build the portfolios. applynoisefiltering: This uses random matrix theory to filter out the covariance matrix from randomness thus yielding better portfolios. A value of 1 will enable it and 0 will disable it. market_index: Which index do you want to use to compare your portfolios. This should mostly be SPY but since we analyzed tech stocks, we used QQQ. only_long: Whether to use long only portfolio or enable short selling as well. Long only portfolios have shown to have better performance using algorithmic techniques. eigenportfolionumber: Which eigen portfolio to use. Any value between 1-5 should work. The first eigen portfolio (1) represents the market portfolio and should act just like the underlying index such as SPY or QQQ. The second one is orthogonal and uncorrelated to the market and poses the greatest risk and reward. The following ones have reduced risk and reward. Read more on eigen-portfolios. stocksfilepath: File that contains the list of stocks that you want to use to build your portfolio. Some Portfolio Building Examples Here are a few examples for building different types of portfolios. Both long and short portfolios by analyzing last 90 days data and keeping the last 30 days as testing data. This will give us 60 days of portfolio construction data and 30 days of testing. Only long portfolio on 60 minute bars of the last 30 days. No future testing. Compare the results with SPY index instead of QQQ. Do not apply noise filtering on the covariance matrix. Use the first eigen portfolio (market portfolio) and compare with SQQQ, Portfolio Strategies Four different portfolio strategies are currently supported by the toolkit. Eigen Portfolios These portfolios are orthogonal and uncorrelated to the market in general thus yielding high reward and alpha. However, since they are uncorrelated to the market, they can also provide great risk. The first eigen portfolio is considered to be a market portfolio which is often ignored. The second one is uncorrelated to the others and provides the highest risk and reward. As we go down the numbering, the risk as well as the reward are reduced. Minimum Variance Portfolio (MVP) MVP tries to minimize the variance of the portfolio. These portfolios are lowest risk and reward. Maximum Sharpe Ratio Portfolio (MSR) MSR solves an optimization problem that tries to maximize the sharpe ratio of the portfolio. It uses past returns during the optimization process which means if past returns are not the same as future returns, the results can vary in future. Genetic Algorithm (GA) based Portfolio This is our own implementation of a GA based portfolio that again tries to maximize the sharpe ratio but in a slightly more robust way. This usually provides more robust portfolios than the others. When you run the command above, our tool will generate portfolios from all these strategies and give them to you. Let us look at some resulting portfolios. Resulting Portfolios For the purpose these results, we will use the 9 stocks in the stocks/stocks.txt file. When we run the above command, we first get the portfolio weights for all four strategies. For testing purposes, the above command used last five years of daily data up till April 29th. The remaining data for this year was used for forward testing i.e the portfolio strategies had no access to it when building the portfolios. What if my portfolio needs different stocks?: All you need to do is change the stocks in the stocks.txt file and run the tool again. Here is the final command again that we run in order to get our portfolios: Portfolio Weights We can see that the eigen portfolio is giving a large weight to TSLA while the others are dividing their weights more uniformly. An interesting phenomena happening here is the hedging with SQQQ that all the strategies have learned automatically. Every tool is assigning some positive weight to SQQQ while also assigning positive weights to other stocks which indicates that the strategies are automatically trying to hedge the portfolios from risk. Obviously this is not perfect, but just the fact that it's happening is fascinating. Let us look at the backtest results on the last five years prior to April 29, 2020. Backtest Results The backtests look pretty encouraging. The black dotted line is the market index i.e QQQ. Other lines are the strategies. Our custom genetic algorithm implementation seems to have the best backtest results because it's an advanced version of other strategies. The eigen portfolio that weighed TSLA the most have the most volatility but its profits are also very high. Finally, as expected, the MVP has the minimum variance and ultimately the least profits. However, since the variance is extremely low, it is a good portfolio for those who want to stay safe. The most interesting part comes next, let us look at the forward or future test results for these portfolios. Forward Test Results These results are from April 29th, 2020 to September 4th, 2020. The eigen portfolio performed the best but it also had a lot of volatility. Moreover, most of those returns are due to TSLA rocketing in the last few months. After that, our GA algorithm worked quite effectively as it beat the market index. Again, as expected, the MVP had the lowest risk and reward and slowly went up in 4-5 months. This shows the effectiveness and power of these algorithmic portfolio optimization strategies where we've developed different portfolios for different kinds of risk and reward profiles. Conclusion and Discussion We are happy to share this toolkit with the trading community and hope that people will like and contribute to it. As is the case with everything in trading, these strategies are not perfect but they are based on rigorous theory and some great empirical results. Please take care when trading with these strategies and always manage your risk. The above results were not cherry picked but the market has been highly bullish in the last few months which has led to the strong results shown above. We would love for the community to try out different strategies and share them with us. Special Thanks Special thanks to Scott Rome's blog. The eigen portfolios and minimum variance portfolio concepts came from his blog posts. The code for filtering eigen values of the covariance matrix was also mostly obtained from one of his posts. License A product by Tradytics Copyright (c) 2020-present, Tradytics.com

machine-learning-blackjack-solution
github
LLM Vibe Score0.42
Human Vibe Score0.022610872675250356
GregSommervilleMar 27, 2025

machine-learning-blackjack-solution

machine-learning-blackjack-solution Introduction A genetic algorithm is a type of artificial intelligence programming that uses ideas from evolution to solve complex problems. It works by creating a population of (initially random) candidate solutions, then repeatedly selecting pairs of candidates and combining their solutions using a process similar to genetic crossover. Sometimes candidate solutions even go through mutation, just to introduce new possibilities into the population. After a large number of generations, the best solution found up to that point is often the optimal, best solution possible. Genetic algorithms are particularly well-suited for combinatorial problems, where there are huge numbers of potential solutions to a problem. The evolutionary process they go through is, in essence, a search through a huge solution space. A solution space so large that you simply could never use a brute force approach. This project is a demonstration of using a genetic algorithm to find an optimal strategy for playing the casino game Blackjack. Please see this article for a story about how this program was used, and what the results were. The article describes some of the available settings, and shows how different values for those settings affect the final result. The source code is for a Windows application written in Cthat allows you to play with different settings like population size, selection style and mutation rate. Each generation's best solution is displayed, so you can watch the program literally evolve a solution. !blackjack strategy tester screenshot The property grid located at the upper left of the screen is where you adjust settings. There's an informational area below that, and the right side of the screen is the display area for the three tables that represent a strategy for playing Blackjack. The tall table on the left is for hard hands, the table in the upper right is for soft hands, and the table in the lower right is for pairs. We'll talk more about how to interpret this strategy in a bit. The columns along the tops of the three tables are for the dealer upcard. When you play Blackjack the dealer has one of his two cards initially turned face up, and the rank of that card has a big impact on recommended strategy. Notice that the upcard ranks don't include Jack, Queen or King. That's because those cards all count 10, so we group them and the Ten together and simplify the tables. To use the tables, first, determine if you have a pair, soft hand, or hard hand. Then look in the appropriate table, with the correct dealer upcard column. The cell in the table will be "H" when the correct strategy is to hit, "S" when the correct strategy is to stand, "D" for double-down, and (in the pairs table only) "P" for split. A Word About This "Optimal" Strategy Before we go any further, it needs to be stated that this problem of finding an optimal Blackjack strategy has already been solved. Back in the 1960s, a mathematician named Edward O. Thorp authored a book called Beat the Dealer, which included charts showing the optimal "Basic" strategy. That strategy looks like this: !optimal blackjack strategy So we're solving a problem that has already been solved, but that's actually good. That means we can compare our results to the known best solution. For example, if our result strategy tells us to do anything but stand when holding a pair of Tens, Jacks, Queens or Kings, we know there's a problem. There's one other thing to get out of the way before we go any further, and that's the idea of nondeterministic code. That means that if we run the same code twice in a row, we're likely to get two different results. That's something that happens with genetic algorithms due to their inherent randomness. There's no guarantee you'll find the absolute optimal solution, but it is assured that you will find an optimal or near-optimal solution. It's something that isn't typical when writing code, so it takes some adjustment for most programmers. Genetic Algorithms Now let's talk about the details of a genetic algorithm. Fitness Scores First of all, we need a way to evaluate candidates so we can compare them to each other. That means a numeric fitness score, which in this case is quite simple: you simulate playing a certain number of hands using the strategy, and then count the number of chips you have at the end. The big question is, how many hands should we test with? The challenge of trying to test a strategy is that due to the innate randomness of Blackjack, you could use the same strategy ten times and get ten completely different results. Obviously, the more hands you play, the more the randomness gets smoothed out, and the quality of the underlying strategy starts to emerge. If you doubt this, just think about flipping a coin. If you only flip it five times, there's certainly a possibility that it'll come up heads all five times (in fact, that happens just over 3% of the time). However, if you flip it 500 times, there's no way it's going to end up all heads - the odds of it happening are 0.5500, which works out to be roughly once every 3 x 10150 times you try it. After some testing and analysis, it was determined that a minimum of 100,000 hands per test is needed for a reasonable level of accuracy. There's still variance even at that number, but in order to cut the variance in half, you'd need to bump the number of hands to 500,000. One reason this accuracy is important is that in the later generations, the differences between candidates are very small. Evolution has caused the main parts of the strategy to converge on a particular approach, and towards the end all it's doing is refining the minor details. In those cases it's important to accurately determine the difference between two similar candidates. Representation Representation is simply the idea that we need to use a data structure for a candidate solution that can be combined via crossover, and possibly mutated. In this case, that's also quite simple because the way that human beings represent a Blackjack strategy is to use three tables, as we've seen. Representing those in code with three two-dimensional arrays is the obvious approach. Each cell in those three tables will have "Hit", "Stand", "Double-Down", or (only for pairs) "Split". By the way, since there are 160 cells in the hard hands table, and 80 cells in the soft hands table, and 100 cells in the pairs table, we can calculate exactly how many possible distinct strategies there are for Blackjack: 4100 x 380 x 3160 = 5 x 10174 possible Blackjack strategies That's a big number, which is obviously impossible to search using brute force. Genetic algorithms (GAs) are extremely helpful when trying to find an optimal solution from a very large set of possible solutions like this. Blackjack Rules and Strategies The rules of Blackjack are fairly simple. The dealer and the player both are dealt two cards. The player sees both of their cards (they are usually dealt face up), and one of the dealer's cards is dealt face up. Each card has a value - for cards between 2 and 10, the value is the same as the card's rank (so an Eight of Spades counts as 8, for example). All face cards count as 10, and an Ace can either be 1 or 11 (it counts as 11 only when that does not result in a hand that exceeds 21). The suit of a card does not matter. After the cards are dealt, if the player has Blackjack (a total of 21) and the dealer does not, the player is immediately paid 1.5 times their original bet, and a new hand is dealt. If the player has 21 and the dealer does also, then it's a tie and the player gets their original bet back, and a new hand is dealt. If the player wasn't dealt a Blackjack, then play continues with the player deciding whether to Stand (not get any more cards), Hit (receive an additional card), Double-down (place an additional bet, and receive one and only one more card), or, in the case of holding a pair, splitting the hand, which means placing an additional bet and receiving two new cards, so the end result is that the player is now playing two (or, in the case of multiple splits, more than two) hands simultaneously. If the player hits or double-downs and has a resulting hand that exceeds 21, then they lose and play continues with the next hand. If not, then the dealer draws until their hand totals at least 17. If the dealer exceeds 21 at this point, the player receives a payment equal to twice their original bet. If the dealer doesn't exceed 21, then the hands are compared and the player with the highest total that doesn't exceed 21 wins. Because of these rules, certain effective strategies emerge. One common strategy is that if you hold a hard hand with a value of 20, 19 or 18, you should Stand, since you avoid busting by going over 21, and you have a nice hand total that might win in a showdown with the dealer. Another common strategy is to split a pair of Aces, since Aces are so powerful (due to the fact that count as 11 or 1, you can often Hit a hand with a soft Ace with no risk of busting). Likewise, splitting a pair of 8s is a good idea because with a hard total of 16, it's likely you will bust if you take a Hit (since so many cards count as 10). As a human being, all it takes is a little knowledge about the rules in order to construct a strategy. The GA program doesn't have that advantage, and operates completely without any pre-programmed knowledge of Blackjack. It simply uses the relative fitness scores and the mechanism of evolution to find the solution. GA Settings There are many variables or settings for a GA. You can adjust population size, how parent candidates are selected, how the resulting children may be mutated, and several other items. The following sections describe some of these settings: Setting: Selection Style Once we've solved representation and have a fitness function, the next step is to select two candidates for crossover during the process of building a new generation. There are three common styles for selection, and this program supports all of them. First, you can choose Roulette Wheel selection. It's named for a Roulette wheel because you can imagine each candidate's fitness score being a wedge in a pie chart, with a size proportionate to its relative fitness compared to the other candidates. (Of course, this assumes that all fitness scores are positive, which we will talk about shortly). The main benefit of Roulette Wheel selection is that selection is fitness-proportionate. Imagine if you had only three candidates, with fitness scores of 1, 3, and 8. The relative selection probabilities for those candidates will be 1/12, 3/12, and 8/12. The downside of Roulette Wheel selection is that it tends to be somewhat slow in terms of processing. The selection process is done by iterating through the candidates until a particular condition is matched - in other words, O(N) performance. Another potential problem with Roulette Wheel selection is that there may be situations where fitness scores vary widely, to such an extent that only certain candidates have any reasonable chance of being selected. This happens frequently in early generations, since the majority of candidates are mostly random. Although this might sound like a positive (since you ultimately want to select candidates with high fitness scores), it also results in a loss of genetic diversity. In other words, even though a particular candidate may have a low fitness score in an early generation, it may contain elements that are needed to find the ultimate solution in later generations. Ranked Selection is the solution to this problem. Instead of using raw fitness scores during the selection process, the candidates are sorted by fitness, with the worst candidate receiving a score of 0, the second worse receiving 1, and so forth, all the way to the best candidate, which has a score equal to the population size - 1. Ranked Selection is quite slow, since it combines the O(N) performance of Roulette Wheel, with the additional requirement that the candidates be sorted before selection. However, there may be circumstances where it performs better than other selection approaches. Finally, the fastest selection method of all is called Tournament Selection. This method simply selects N random candidates from the current generation, and then uses the one with the best fitness score. A tournament size of 2 means two random candidates are selected, and the best of those two is used. If you have a large tournament size (like 10), then 10 different candidates will be selected, with the best of those being the ultimate selection. That obviously tilts the balance between randomness and quality. Tournament selection works well in most cases, but it does require some experimentation to find the best tourney size. Setting: Elitism Elitism is a technique that helps ensure that the best candidates are always maintained. Since all selection methods are random to some degree, it is possible to completely lose the best candidates from one generation to another. By using Elitism, we automatically advance a certain percentage of the best candidates to the next generation. Elitism does have a negative impact on performance since all of the candidates must be sorted by fitness score. Typically Elitism is done before filling the rest of a new generation with new candidates created by crossover. Crossover Details Once two candidate solutions have been selected, the next step in building a new generation is to combine those two into a single new candidate, hopefully using the best of both parent strategies. There are a number of ways to do crossover, but the method used in this program is quite straightforward - the two fitness scores are compared, and crossover happens in a relatively proportionate way. If one candidate has a fitness of 10, and the other has a fitness of 5, then the one with fitness 10 contributes twice as much to the child as the parent with a fitness of 5. Since the fitness scores in this program are based on how much the strategy would win over thousands of hands, almost all fitness scores will be negative. (This is obviously because the rules are set up so the house always wins.) This makes it difficult to calculate relative fitnesses (how do you compare a positive number with a negative, and find relative proportions?), and also causes problems with selection methods like Roulette Wheel or Ranked. To solve this, we find the lowest fitness score of the generation and add that value to each candidate. This results in an adjusted fitness score of 0 for the very worse candidate, so it never gets selected. Mutation As has been mentioned a few times, maintaining genetic diversity in our population of candidate solutions is a good thing. It helps the GA ultimately find the very best solution, by occasionally altering a candidate in a positive direction. There are two settings for mutation. MutationRate controls what percentage of new candidates have mutation done on them. MutationImpact controls what percentage of their strategy is randomized. Population Size Population size has a significant impact on performance. The smaller the population size, the faster the GA will execute. On the other hand, if the size is too low the population may not have enough genetic diversity to find the ultimate solution. During testing, it looks like 700 to 1000 is a good balance between speed and correctness. Performance Notes This program consumes a lot of processing power. Running tests of hundreds of thousands of hands of Blackjack for hundreds or thousands of candidates consumes a lot of time. It's really imperative to write the code so that it works as efficiently as possible. If your CPU isn't consistently at or above 95% usage, there's still room for improvement. Multi-threading is a natural fit for genetic algorithms because we often want to perform the same action on each candidate. The best example of this is when we calculate fitness scores. This is often an operation that takes quite a bit of time. In our case, we're dealing out 100,000 hands, and each hand has to be played until the end. If we're single-threading that code, it's going to take a long time. Multi-threading is really the way to go. Luckily, there's a ridiculously simple way to efficiently use all of your processors for an operation like this. This code loops over all of the candidates in the currentGeneration list, calls the fitness function and sets the fitness property for each: Regardless of the number of items in the list or the number of processors on your machine, the code will efficiently run the code in a multi-threaded manner, and continue only when all of the threads are complete. One of the side effects of making this code multi-threaded is that all of the code relating to evaluating a candidate must be thread-safe, including any Singleton objects. When making code thread-safe, pay attention that you don't accidentally introduce code that will slow your program down unintentionally, because sometimes it can be quite subtle. Random numbers are central to how genetic algorithms work, so it's critical that they can be used correctly from a multithreaded environment. That means that each random number generator must be separate from the others, and it also means that each must produce a distinct series of random numbers. Random number generators use seed values which are usually time-based, like the number of milliseconds the computer has been turned on. Starting with that seed, subsequent calls will return a series of numbers that look random, but really aren't. If you start with the same seed, you get the same sequence. And that's a problem because if you create multiple random number generator objects in a loop using the default time-based seed, several of them will have the same time-based initial seed value, which will result in the same sequence of "random" numbers. That's a bug, because it can reduce the true randomness of the program a great deal, and that's vital to a genetic algorithm. There are a couple of ways to solve this problem. First, you can make the random object truly a singleton, and restrict access to it by using a Clock statement. The makes all access serialized for any random number need, which reduces performance. Another approach is to make the variable static per thread. By declaring the variable as static and also marking it with the [ThreadStatic] attribute, the .NET runtime allocates one static variable per thread. That eliminates the locking/serialization, but also has performance issues. The approach used in this application is to use a non-default seed value. In this case we call Guid.NewGuid().GetHashCode(), which generates a new, unique GUID, then gets an integer hashcode value that should be unique, depending on how GetHashCode is implemented. While multithreading really helps performance, there are also other things we can do to improve performance. For example, when dealing with large populations, the hundreds or thousands of objects that will be generated each generation can quickly turn into a huge problem related to garbage collection. In the end, the easiest way to solve that is to look through the code and find objects being allocate inside a loop. It's better to declare the variable outside of the loop, and then clear it in the loop, rather than reallocate it. In a program like this one where you could be looping hundreds of thousands of times, this can result in a very significant performance boost. For example, in an early version of this code, a Deck object was created for each hand. Since there are hundreds of candidate solutions running hundreds of thousands of trial hands, this was a huge inefficiency. The code was changed to allocate one deck per test sequence. The deck was shuffled as needed, so it never needs to be reallocated. Beyond the cards in the deck, another object type that was repeatedly created and destroyed were the candidate strategies. To mitigate this problem, a StrategyPool class was created that handles allocation and deallocation. This means that strategy objects are reused, rather than dynamically created when needed. The pool class has to be thread-safe, so it does serialize access to its methods via a Clock statement, but overall using the pool approach produced a good performance increase. Finally, a subtle form of object allocation is conversion. In an early version of the code, a utility card function used Convert.ToInt32(rankEnum). Obviously, the easiest way to convert from an enum to an int is simply to cast it, like (int)rankEnum. But it's hard to know exactly what the difference is between that approach, int.Parse(), int.TryParse(), or Convert.ToInt32(), since they can all be used and are roughly equivalent. Perhaps the compiler was boxing the enum value before passing it to Convert.ToInt32(), because the profiler identified this as a function that had large amounts of thread contention waiting - and the problem got much, much worse as the generations passed. By rewriting the conversion to use a simple cast, the program performance increased threefold (3x). Contributing Please read CONTRIBUTING.md for details on our code of conduct, and the process for submitting pull requests to us. Author Greg Sommerville - Initial work* License This project is licensed under the Apache 2.0 License - see the LICENSE.md file for details

With Vibe Coding Say Goodbye to Boring Coding!
youtube
LLM Vibe Score0.321
Human Vibe Score0.44
GeeksforGeeksMar 27, 2025

With Vibe Coding Say Goodbye to Boring Coding!

Coding doesn’t have to be boring anymore! With the rise of AI-powered tools and innovative development approaches, the way we write code is changing drastically. Are you ready to embrace this new era of vibe coding? 🚀 💡 Want to level up your coding and problem-solving skills? Join the Three 90 Challenge by GeeksforGeeks—ending on 31st March! ✅ Complete 90% of your course in 90 days ✅ Get 90% of your fee refunded! Yes, you read that right! 🌟 Over ₹5 CRORE in refunds already processed—yours could be next! 👉 Start the challenge now: https://gfgcdn.com/tu/U4a/ 📌 Stay Connected for More Coding Challenges & Learning Resources: 📱 Download the GeeksforGeeks App: https://play.google.com/store/apps/details?id=free.programming.programming 💬 Twitter: https://twitter.com/geeksforgeeks 🧑‍💼 LinkedIn: https://www.linkedin.com/company/geeksforgeeks 📷 Instagram: https://www.instagram.com/geeksforgeeks/ 💌 Telegram: https://t.me/geeksforgeeks_official 📌 Pinterest: https://in.pinterest.com/geeksforgeeks/ 🎮 Discord: https://discord.gg/geeksforgeeks 🔍 Tags: AI Coding, AI-Powered Development, Vibe Coding, Future of Programming, Software Development Trends, Coding with AI, AI-Assisted Programming, Tech Innovations, Machine Learning in Coding, AI Coding Assistants, Software Engineering Revolution, AI for Developers, ChatGPT Coding, AI Coding Tools, gfg, gfg courses, gfg classes, it jobs, it job market, ai trends, ai news, ai vs software developers 🔥 Hashtags: #AICoding #FutureOfProgramming #VibeCoding #SoftwareDevelopment #TechTrends #CodingWithAI #AIRevolution #AIInTech #MachineLearning #CodingFuture #GeeksforGeeks #CodeSmarter #AIforDevelopers

yoha
github
LLM Vibe Score0.556
Human Vibe Score0.3408299306652369
handtracking-ioMar 27, 2025

yoha

Yoha A practical hand tracking engine. Note: Yoha is currently unmaintained. Quick Links: Demo (Code) Docs Website npm Installation npm install @handtracking.io/yoha Please note: You need to serve the files from node_modules/@handtracking.io/yoha since the library needs to download the model files from here. (Webpack Example) You need to serve your page with https for webcam access. (Webpack Example) You should use cross-origin isolation as it improves the engine's performance in certain scenarios. (Webpack Example) Description Yoha is a hand tracking engine that is built with the goal of being a versatile solution in practical scenarios where hand tracking is employed to add value to an application. While ultimately the goal is to be a general purpose hand tracking engine supporting any hand pose, the engine evolves around specific hand poses that users/developers find useful. These poses are detected by the engine which allows to build applications with meaningful interactions. See the demo for an example. Yoha is currently in beta. About the name: Yoha is short for ("Your Hand Tracking"). Language Support Yoha is currently available for the web via JavaScript. More languages will be added in the future. If you want to port Yoha to another language and need help feel free reach out. Technical Details Yoha was built from scratch. It uses a custom neural network trained using a custom dataset. The backbone for the inference in the browser is currently TensorFlow.js Features: Detection of 21 2D-landmark coordinates (single hand). Hand presence detection. Hand orientation (left/right hand) detection. Inbuilt pose detection. Supported Hand Poses: Pinch (index finger and thumb touch) Fist Your desired pose is not on this list? Feel free to create an issue for it. Performance Yoha was built with performance in mind. It is able to provide realtime user experience on a broad range of laptops and desktop devices. The performance on mobile devices is not great which hopefuly will change with the further development of inference frameworks like TensorFlow.js Please note that native inference speed can not be compared with the web inference speed. Differently put, if you were to run Yoha natively it would be much faster than via the web browser. Minimal Example Source Running locally: Drawing Demo Live Version Source Running locally:

obsei
github
LLM Vibe Score0.545
Human Vibe Score0.10175553624190911
obseiMar 27, 2025

obsei

Note: Obsei is still in alpha stage hence carefully use it in Production. Also, as it is constantly undergoing development hence master branch may contain many breaking changes. Please use released version. Obsei (pronounced "Ob see" | /əb-'sē/) is an open-source, low-code, AI powered automation tool. Obsei consists of - Observer: Collect unstructured data from various sources like tweets from Twitter, Subreddit comments on Reddit, page post's comments from Facebook, App Stores reviews, Google reviews, Amazon reviews, News, Website, etc. Analyzer: Analyze unstructured data collected with various AI tasks like classification, sentiment analysis, translation, PII, etc. Informer: Send analyzed data to various destinations like ticketing platforms, data storage, dataframe, etc so that the user can take further actions and perform analysis on the data. All the Observers can store their state in databases (Sqlite, Postgres, MySQL, etc.), making Obsei suitable for scheduled jobs or serverless applications. !Obsei diagram Future direction - Text, Image, Audio, Documents and Video oriented workflows Collect data from every possible private and public channels Add every possible workflow to an AI downstream application to automate manual cognitive workflows Use cases Obsei use cases are following, but not limited to - Social listening: Listening about social media posts, comments, customer feedback, etc. Alerting/Notification: To get auto-alerts for events such as customer complaints, qualified sales leads, etc. Automatic customer issue creation based on customer complaints on Social Media, Email, etc. Automatic assignment of proper tags to tickets based content of customer complaint for example login issue, sign up issue, delivery issue, etc. Extraction of deeper insight from feedbacks on various platforms Market research Creation of dataset for various AI tasks Many more based on creativity 💡 Installation Prerequisite Install the following (if not present already) - Install Python 3.7+ Install PIP Install Obsei You can install Obsei either via PIP or Conda based on your preference. To install latest released version - Install from master branch (if you want to try the latest features) - Note: all option will install all the dependencies which might not be needed for your workflow, alternatively following options are available to install minimal dependencies as per need - pip install obsei[source]: To install dependencies related to all observers pip install obsei[sink]: To install dependencies related to all informers pip install obsei[analyzer]: To install dependencies related to all analyzers, it will install pytorch as well pip install obsei[twitter-api]: To install dependencies related to Twitter observer pip install obsei[google-play-scraper]: To install dependencies related to Play Store review scrapper observer pip install obsei[google-play-api]: To install dependencies related to Google official play store review API based observer pip install obsei[app-store-scraper]: To install dependencies related to Apple App Store review scrapper observer pip install obsei[reddit-scraper]: To install dependencies related to Reddit post and comment scrapper observer pip install obsei[reddit-api]: To install dependencies related to Reddit official api based observer pip install obsei[pandas]: To install dependencies related to TSV/CSV/Pandas based observer and informer pip install obsei[google-news-scraper]: To install dependencies related to Google news scrapper observer pip install obsei[facebook-api]: To install dependencies related to Facebook official page post and comments api based observer pip install obsei[atlassian-api]: To install dependencies related to Jira official api based informer pip install obsei[elasticsearch]: To install dependencies related to elasticsearch informer pip install obsei[slack-api]:To install dependencies related to Slack official api based informer You can also mix multiple dependencies together in single installation command. For example to install dependencies Twitter observer, all analyzer, and Slack informer use following command - How to use Expand the following steps and create a workflow - Step 1: Configure Source/Observer Twitter Youtube Scrapper Facebook Email Google Maps Reviews Scrapper AppStore Reviews Scrapper Play Store Reviews Scrapper Reddit Reddit Scrapper Note: Reddit heavily rate limit scrappers, hence use it to fetch small data during long period Google News Web Crawler Pandas DataFrame Step 2: Configure Analyzer Note: To run transformers in an offline mode, check transformers offline mode. Some analyzer support GPU and to utilize pass device parameter. List of possible values of device parameter (default value auto): auto: GPU (cuda:0) will be used if available otherwise CPU will be used cpu: CPU will be used cuda:{id} - GPU will be used with provided CUDA device id Text Classification Text classification: Classify text into user provided categories. Sentiment Analyzer Sentiment Analyzer: Detect the sentiment of the text. Text classification can also perform sentiment analysis but if you don't want to use heavy-duty NLP model then use less resource hungry dictionary based Vader Sentiment detector. NER Analyzer NER (Named-Entity Recognition) Analyzer: Extract information and classify named entities mentioned in text into pre-defined categories such as person names, organizations, locations, medical codes, time expressions, quantities, monetary values, percentages, etc Translator PII Anonymizer Dummy Analyzer Dummy Analyzer: Does nothing. Its simply used for transforming the input (TextPayload) to output (TextPayload) and adding the user supplied dummy data. Step 3: Configure Sink/Informer Slack Zendesk Jira ElasticSearch Http Pandas DataFrame Logger This is useful for testing and dry running the pipeline. Step 4: Join and create workflow source will fetch data from the selected source, then feed it to the analyzer for processing, whose output we feed into a sink to get notified at that sink. Step 5: Execute workflow Copy the code snippets from Steps 1 to 4 into a python file, for example example.py and execute the following command - Demo We have a minimal streamlit based UI that you can use to test Obsei. !Screenshot Watch UI demo video Check demo at (Note: Sometimes the Streamlit demo might not work due to rate limiting, use the docker image (locally) in such cases.) To test locally, just run To run Obsei workflow easily using GitHub Actions (no sign ups and cloud hosting required), refer to this repo. Companies/Projects using Obsei Here are some companies/projects (alphabetical order) using Obsei. To add your company/project to the list, please raise a PR or contact us via email. Oraika: Contextually understand customer feedback 1Page: Giving a better context in meetings and calls Spacepulse: The operating system for spaces Superblog: A blazing fast alternative to WordPress and Medium Zolve: Creating a financial world beyond borders Utilize: No-code app builder for businesses with a deskless workforce Articles Sr. No. Title Author 1 AI based Comparative Customer Feedback Analysis Using Obsei Reena Bapna 2 LinkedIn App - User Feedback Analysis Himanshu Sharma Tutorials Sr. No. Workflow Colab Binder 1 Observe app reviews from Google play store, Analyze them by performing text classification and then Inform them on console via logger PlayStore Reviews → Classification → Logger 2 Observe app reviews from Google play store, PreProcess text via various text cleaning functions, Analyze them by performing text classification, Inform them to Pandas DataFrame and store resultant CSV to Google Drive PlayStore Reviews → PreProcessing → Classification → Pandas DataFrame → CSV in Google Drive 3 Observe app reviews from Apple app store, PreProcess text via various text cleaning function, Analyze them by performing text classification, Inform them to Pandas DataFrame and store resultant CSV to Google Drive AppStore Reviews → PreProcessing → Classification → Pandas DataFrame → CSV in Google Drive 4 Observe news article from Google news, PreProcess text via various text cleaning function, Analyze them via performing text classification while splitting text in small chunks and later computing final inference using given formula Google News → Text Cleaner → Text Splitter → Classification → Inference Aggregator 💡Tips: Handle large text classification via Obsei Documentation For detailed installation instructions, usages and examples, refer to our documentation. Support and Release Matrix Linux Mac Windows Remark Tests ✅ ✅ ✅ Low Coverage as difficult to test 3rd party libs PIP ✅ ✅ ✅ Fully Supported Conda ❌ ❌ ❌ Not Supported Discussion forum Discussion about Obsei can be done at community forum Changelogs Refer releases for changelogs Security Issue For any security issue please contact us via email Stargazers over time Maintainers This project is being maintained by Oraika Technologies. Lalit Pagaria and Girish Patel are maintainers of this project. License Copyright holder: Oraika Technologies Overall Apache 2.0 and you can read License file. Multiple other secondary permissive or weak copyleft licenses (LGPL, MIT, BSD etc.) for third-party components refer Attribution. To make project more commercial friendly, we void third party components which have strong copyleft licenses (GPL, AGPL etc.) into the project. Attribution This could not have been possible without these open source softwares. Contribution First off, thank you for even considering contributing to this package, every contribution big or small is greatly appreciated. Please refer our Contribution Guideline and Code of Conduct. Thanks so much to all our contributors

CollabAI
github
LLM Vibe Score0.449
Human Vibe Score0.07795191529604462
sjinnovationMar 27, 2025

CollabAI

CollabAI About Welcome to Collabai.software, where we've taken the world of AI to new heights. We've been working tirelessly to bring you the most advanced, user-friendly platform that seamlessly integrates with the powerful OpenAI API, Gemini, and Claude. Imagine running your own ChatGPT on your server, with the ability to manage access for your entire team. Picture creating custom AI assistants that cater to your unique needs, and organizing your employees into groups for streamlined collaboration. With Collabai.software, this is not just a dream, but a reality. Collabai.software Features: Self-Hosting on Your Cloud: Gain full control by hosting the platform on your private cloud. Ensure data privacy by using your API codes, allowing for secure data handling. Enhanced Team Management: Manage teams with private accounts and customizable access levels (Departments). Prompt Templates: Utilize generic templates to streamline team usage. Departmental Access & Assistant Assignment: Assign AI assistants to specific departments for shared team access. Customizable AI Assistants: Create personalized AI assistants for users or organizations. Tagging Feature in Chats: Organize and retrieve chat data efficiently with custom tags. Chat Storage and Retrieval: Save all chats and replies for future analysis, with an option to restore accidentally deleted chats from Trash. Optimized Performance: Experience our high-speed, efficient platform. Our clients have been using it for over a year, with some spending $1500-$2000 per month on the API. File Upload & GPT-4 Vision Integration: Enhance interactions by uploading files for analysis and sending pictures for AI description. OpenAI API, Gemini, and Claude Integration: Seamlessly integrate with the powerful OpenAI API, Gemini, and Claude for a comprehensive suite of AI capabilities. API-Based Function Calls: Execute custom functions and automate tasks directly through the API. Usage Monitoring: Track your daily and monthly API usage costs to optimize spending. Day and Night Mode: Switch between light and dark themes to enhance visual comfort. Additional Features: Private Accounts: Ensure the security and privacy of your team members' data. Customizable Access Levels: Tailor access permissions to meet the specific needs of your organization. Shared Team Access: Foster collaboration by assigning AI assistants to specific departments or teams. AI-Powered File Analysis: Gain insights and automate tasks by uploading files for AI analysis. AI-Generated Image Descriptions: Enhance communication and understanding by sending pictures for AI-powered descriptions. !image !image !image Folder Structure Client The client folder contains the React-based frontend code for the application. This includes JSX, CSS, and JavaScript files, as well as any additional assets such as images or fonts. Below is a brief overview of the main subdirectories within the client folder: src: This directory contains the React components, styles, and scripts for the frontend application. public: Static assets, such as images or favicon.ico, go here. This folder is served as-is and not processed by the build system. Server The server folder contains all the backend-related code for the application, following a Model-View-Controller (MVC) pattern. Here is a breakdown of the main subdirectories within the server folder: controllers: This directory holds the controller files responsible for handling requests, processing data, and interacting with models. models: Data models and database-related code are organized in this folder. config: Configuration files for the backend, such as database configuration or any other service configuration should be stored here, can be stored in this directory. Getting Started Follow the steps below to get the project up and running. Prerequisites Node.js (Version: >=20.x) MongoDB NPM Development Setup Clone the Repository bash cd client Install Dependencies bash cd ../server Install Backend Dependencies bash npm start To initialize the application data and create a superadmin user, you can use either cURL or Postman: Using cURL If you prefer command-line tools, you can use curl to make a POST request to the /init-setup endpoint. Open your terminal and run the following command: curl -X POST http://localhost:8011/api/init -H "Content-Type: application/json" -d '{ "fname": "Super", "lname": "Admin", "email": "superadmin@example.com", "password": "yourSecurePassword", "employeeCount": 100, "companyName": "INIT_COMPANY" }' Initializing Setup with Postman Open Postman: Launch the Postman application. Create a New Request: Click on the '+' or 'New' button to create a new request. Set HTTP Method to POST: Ensure that the HTTP method is set to POST. Enter URL: Enter the URL http://localhost:8011/api/init. Set Headers: Go to the 'Headers' tab. Set Content-Type to application/json. Set Request Body: Switch to the 'Body' tab. Select the 'raw' radio button. Enter the JSON data for your superadmin user: Send Request: Click the 'Send' button to make the request. This will send a POST request to http://localhost:8011/api/init with the provided JSON payload, creating a superadmin user with the specified details. Site Setup: Login with the superadmin credentials and set up your site by adding configs from your settings page, for ex. API keys, etc. Reference CollaborativeAI Reference Guide Contributing If you would like to contribute to the project, we welcome your contributions! Please follow the guidelines outlined in the CONTRIBUTING.md file. Feel free to raise issues, suggest new features, or send pull requests to help improve the project. Your involvement is greatly appreciated! Thank you for contributing to our project! License MIT

AI-PhD-S24
github
LLM Vibe Score0.472
Human Vibe Score0.0922477795435268
rphilipzhangMar 25, 2025

AI-PhD-S24

Artificial Intelligence for Business Research (Spring 2024) Scribed Lecture Notes Class Recordings (You need to apply for access.) Teaching Team Instructor*: Renyu (Philip) Zhang, Associate Professor, Department of Decisions, Operations and Technology, CUHK Business School, philipzhang@cuhk.edu.hk, @911 Cheng Yu Tung Building. Teaching Assistant*: Leo Cao, Full-time TA, Department of Decisions, Operations and Technology, CUHK Business School, yinglyucao@cuhk.edu.hk. Please be noted that Leo will help with any issues related to the logistics, but not the content, of this course. Tutorial Instructor*: Qiansiqi Hu, MSBA Student, Department of Decisions, Operations and Technology, CUHK Business School, 1155208353@link.cuhk.edu.hk. BS in ECE, Shanghai Jiaotong University Michigan Institute. Basic Information Website: https://github.com/rphilipzhang/AI-PhD-S24 Time: Tuesday, 12:30pm-3:15pm, from Jan 9, 2024 to Apr 16, 2024, except for Feb 13 (Chinese New Year) and Mar 5 (Final Project Discussion) Location: Cheng Yu Tung Building (CYT) LT5 About Welcome to the mono-repo of the PhD course AI for Business Research (DSME 6635) at CUHK Business School in Spring 2024. You may download the Syllabus of this course first. The purpose of this course is to learn the following: Have a basic understanding of the fundamental concepts/methods in machine learning (ML) and artificial intelligence (AI) that are used (or potentially useful) in business research. Understand how business researchers have utilized ML/AI and what managerial questions have been addressed by ML/AI in the recent decade. Nurture a taste of what the state-of-the-art AI/ML technologies can do in the ML/AI community and, potentially, in your own research field. We will meet each Tuesday at 12:30pm in Cheng Yu Tung Building (CYT) LT5 (please pay attention to this room change). Please ask for my approval if you need to join us via the following Zoom links: Zoom link, Meeting ID 996 4239 3764, Passcode 386119. Most of the code in this course will be distributed through the Google CoLab cloud computing environment to avoid the incompatibility and version control issues on your local individual computer. On the other hand, you can always download the Jupyter Notebook from CoLab and run it your own computer. The CoLab files of this course can be found at this folder. The Google Sheet to sign up for groups and group tasks can be found here. The overleaf template for scribing the lecture notes of this course can be found here. If you have any feedback on this course, please directly contact Philip at philipzhang@cuhk.edu.hk and we will try our best to address it. Brief Schedule Subject to modifications. All classes start at 12:30pm and end at 3:15pm. |Session|Date |Topic|Key Words| |:-------:|:-------------:|:----:|:-:| |1|1.09|AI/ML in a Nutshell|Course Intro, ML Models, Model Evaluations| |2|1.16|Intro to DL|DL Intro, Neural Nets, Computational Issues in DL| |3|1.23|Prediction and Traditional NLP|Prediction in Biz Research, Pre-processing| |4|1.30|NLP (II): Traditional NLP|$N$-gram, NLP Performance Evaluations, Naïve Bayes| |5|2.06|NLP (III): Word2Vec|CBOW, Skip Gram| |6|2.20|NLP (IV): RNN|Glove, Language Model Evaluation, RNN| |7|2.27|NLP (V): Seq2Seq|LSTM, Seq2Seq, Attention Mechanism| |7.5|3.05|NLP (V.V): Transformer|The Bitter Lesson, Attention is All You Need| |8|3.12|NLP (VI): Pre-training|Computational Tricks in DL, BERT, GPT| |9|3.19|NLP (VII): LLM|Emergent Abilities, Chain-of-Thought, In-context Learning, GenAI in Business Research| |10|3.26|CV (I): Image Classification|CNN, AlexNet, ResNet, ViT| |11|4.02|CV (II): Image Segmentation and Video Analysis|R-CNN, YOLO, 3D-CNN| |12|4.09|Unsupervised Learning (I): Clustering & Topic Modeling|GMM, EM Algorithm, LDA| |13|4.16|Unsupervised Learning (II): Diffusion Models|VAE, DDPM, LDM, DiT| Important Dates All problem sets are due at 12:30pm right before class. |Date| Time|Event|Note| |:--:|:-:|:---:|:--:| |1.10| 11:59pm|Group Sign-Ups|Each group has at most two students.| |1.12| 7:00pm-9:00pm|Python Tutorial|Given by Qiansiqi Hu, Python Tutorial CoLab| |1.19| 7:00pm-9:00pm|PyTorch Tutorial|Given by Qiansiqi Hu, PyTorch Tutorial CoLab| |3.05|9:00am-6:00pm|Final Project Discussion|Please schedule a meeting with Philip.| |3.12| 12:30pm|Final Project Proposal|1-page maximum| |4.30| 11:59pm|Scribed Lecture Notes|Overleaf link| |5.12|11:59pm|Project Paper, Slides, and Code|Paper page limit: 10| Useful Resources Find more on the Syllabus. Books: ESL, Deep Learning, Dive into Deep Learning, ML Fairness, Applied Causal Inference Powered by ML and AI Courses: ML Intro by Andrew Ng, DL Intro by Andrew Ng, NLP (CS224N) by Chris Manning, CV (CS231N) by Fei-Fei Li, Deep Unsupervised Learning by Pieter Abbeel, DLR by Sergey Levine, DL Theory by Matus Telgarsky, LLM by Danqi Chen, Generative AI by Andrew Ng, Machine Learning and Big Data by Melissa Dell and Matthew Harding, Digital Economics and the Economics of AI by Martin Beraja, Chiara Farronato, Avi Goldfarb, and Catherine Tucker Detailed Schedule The following schedule is tentative and subject to changes. Session 1. Artificial Intelligence and Machine Learning in a Nutshell (Jan/09/2024) Keywords: Course Introduction, Machine Learning Basics, Bias-Variance Trade-off, Cross Validation, $k$-Nearest Neighbors, Decision Tree, Ensemble Methods Slides: Course Introduction, Machine Learning Basics CoLab Notebook Demos: k-Nearest Neighbors, Decision Tree Homework: Problem Set 1: Bias-Variance Trade-Off Online Python Tutorial: Python Tutorial CoLab, 7:00pm-9:00pm, Jan/12/2024 (Friday), given by Qiansiqi Hu, 1155208353@link.cuhk.edu.hk. Zoom Link, Meeting ID: 923 4642 4433, Pass code: 178146 References: The Elements of Statistical Learning (2nd Edition), 2009, by Trevor Hastie, Robert Tibshirani, Jerome Friedman, https://hastie.su.domains/ElemStatLearn/. Probabilistic Machine Learning: An Introduction, 2022, by Kevin Murphy, https://probml.github.io/pml-book/book1.html. Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106. Athey, Susan, and Guido W. Imbens. 2019. Machine learning methods that economists should know about. Annual Review of Economics 11: 685-725. Hofman, Jake M., et al. 2021. Integrating explanation and prediction in computational social science. Nature 595.7866: 181-188. Bastani, Hamsa, Dennis Zhang, and Heng Zhang. 2022. Applied machine learning in operations management. Innovative Technology at the Interface of Finance and Operations. Springer: 189-222. Kelly, Brian, and Dacheng Xiu. 2023. Financial machine learning, SSRN, https://ssrn.com/abstract=4501707. The Bitter Lesson, by Rich Sutton, which develops so far the most critical insight of AI: "The biggest lesson that can be read from 70 years of AI research is that general methods that leverage computation are ultimately the most effective, and by a large margin." Session 2. Introduction to Deep Learning (Jan/16/2024) Keywords: Random Forests, eXtreme Gradient Boosting Trees, Deep Learning Basics, Neural Nets Models, Computational Issues of Deep Learning Slides: Machine Learning Basics, Deep Learning Basics CoLab Notebook Demos: Random Forest, Extreme Gradient Boosting Tree, Gradient Descent, Chain Rule Presentation: By Xinyu Li and Qingyu Xu. Gu, Shihao, Brian Kelly, and Dacheng Xiu. 2020. Empirical asset pricing via machine learning. Review of Financial Studies 33: 2223-2273. Link to the paper. Homework: Problem Set 2: Implementing Neural Nets Online PyTorch Tutorial: PyTorch Tutorial CoLab, 7:00pm-9:00pm, Jan/19/2024 (Friday), given by Qiansiqi Hu, 1155208353@link.cuhk.edu.hk. Zoom Link, Meeting ID: 923 4642 4433, Pass code: 178146 References: Deep Learning, 2016, by Ian Goodfellow, Yoshua Bengio and Aaron Courville, https://www.deeplearningbook.org/. Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. Probabilistic Machine Learning: Advanced Topics, 2023, by Kevin Murphy, https://probml.github.io/pml-book/book2.html. Deep Learning with PyTorch, 2020, by Eli Stevens, Luca Antiga, and Thomas Viehmann. Gu, Shihao, Brian Kelly, and Dacheng Xiu. 2020. Empirical asset pricing with machine learning. Review of Financial Studies 33: 2223-2273. Session 3. DL Basics, Predictions in Business Research, and Traditonal NLP (Jan/23/2024) Keywords: Optimization and Computational Issues of Deep Learning, Prediction Problems in Business Research, Pre-processing and Word Representations in Traditional Natural Language Processing Slides: Deep Learning Basics, Prediction Problems in Business Research, NLP(I): Pre-processing and Word Representations.pdf) CoLab Notebook Demos: He Initialization, Dropout, Micrograd, NLP Pre-processing Presentation: By Letian Kong and Liheng Tan. Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106. Link to the paper. Homework: Problem Set 2: Implementing Neural Nets, due at 12:30pm, Jan/30/2024 (Tuesday). References: Kleinberg, Jon, Jens Ludwig, Sendhil Mullainathan, and Ziad Obermeyer. 2015. Prediction policy problems. American Economic Review 105(5): 491-495. Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106. Kleinberg, Jon, Himabindu Lakkaraju, Jure Leskovec, Jens Ludwig, and Sendhil Mullainathan. 2018. Human decisions and machine predictions. Quarterly Journal of Economics 133(1): 237-293. Bajari, Patrick, Denis Nekipelov, Stephen P. Ryan, and Miaoyu Yang. 2015. Machine learning methods for demand estimation. American Economic Review, 105(5): 481-485. Farias, Vivek F., and Andrew A. Li. 2019. Learning preferences with side information. Management Science 65(7): 3131-3149. Cui, Ruomeng, Santiago Gallino, Antonio Moreno, and Dennis J. Zhang. 2018. The operational value of social media information. Production and Operations Management, 27(10): 1749-1769. Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574. Chapter 2, Introduction to Information Retrieval, 2008, Cambridge University Press, by Christopher D. Manning, Prabhakar Raghavan and Hinrich Schutze, https://nlp.stanford.edu/IR-book/information-retrieval-book.html. Chapter 2, Speech and Language Processing (3rd ed. draft), 2023, by Dan Jurafsky and James H. Martin, https://web.stanford.edu/~jurafsky/slp3/. Parameter Initialization and Batch Normalization (in Chinese) GPU Comparisons-vs-NVIDIA-H100-(PCIe)-vs-NVIDIA-RTX-6000-Ada/624vs632vs640) GitHub Repo for Micrograd, by Andrej Karpathy. Hand Written Notes Session 4. Traditonal NLP (Jan/30/2024) Keywords: Pre-processing and Word Representations in NLP, N-Gram, Naïve Bayes, Language Model Evaluation, Traditional NLP Applied to Business/Econ Research Slides: NLP(I): Pre-processing and Word Representations.pdf), NLP(II): N-Gram, Naïve Bayes, and Language Model Evaluation.pdf) CoLab Notebook Demos: NLP Pre-processing, N-Gram, Naïve Bayes Presentation: By Zhi Li and Boya Peng. Hansen, Stephen, Michael McMahon, and Andrea Prat. 2018. Transparency and deliberation within the FOMC: A computational linguistics approach. Quarterly Journal of Economics, 133(2): 801-870. Link to the paper. Homework: Problem Set 3: Implementing Traditional NLP Techniques, due at 12:30pm, Feb/6/2024 (Tuesday). References: Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574. Hansen, Stephen, Michael McMahon, and Andrea Prat. 2018. Transparency and deliberation within the FOMC: A computational linguistics approach. Quarterly Journal of Economics, 133(2): 801-870. Chapters 2, 12, & 13, Introduction to Information Retrieval, 2008, Cambridge University Press, by Christopher D. Manning, Prabhakar Raghavan and Hinrich Schutze, https://nlp.stanford.edu/IR-book/information-retrieval-book.html. Chapter 2, 3 & 4, Speech and Language Processing (3rd ed. draft), 2023, by Dan Jurafsky and James H. Martin, https://web.stanford.edu/~jurafsky/slp3/. Natural Language Tool Kit (NLTK) Documentation Hand Written Notes Session 5. Deep-Learning-Based NLP: Word2Vec (Feb/06/2024) Keywords: Traditional NLP Applied to Business/Econ Research, Word2Vec: Continuous Bag of Words and Skip-Gram Slides: NLP(II): N-Gram, Naïve Bayes, and Language Model Evaluation.pdf), NLP(III): Word2Vec.pdf) CoLab Notebook Demos: Word2Vec: CBOW, Word2Vec: Skip-Gram Presentation: By Xinyu Xu and Shu Zhang. Timoshenko, Artem, and John R. Hauser. 2019. Identifying customer needs from user-generated content. Marketing Science, 38(1): 1-20. Link to the paper. Homework: No homework this week. Probably you should think about your final project when enjoying your Lunar New Year Holiday. References: Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574. Tetlock, Paul. 2007. Giving content to investor sentiment: The role of media in the stock market. Journal of Finance, 62(3): 1139-1168. Baker, Scott, Nicholas Bloom, and Steven Davis, 2016. Measuring economic policy uncertainty. Quarterly Journal of Economics, 131(4): 1593-1636. Gentzkow, Matthew, and Jesse Shapiro. 2010. What drives media slant? Evidence from US daily newspapers. Econometrica, 78(1): 35-71. Timoshenko, Artem, and John R. Hauser. 2019. Identifying customer needs from user-generated content. Marketing Science, 38(1): 1-20. Mikolov, Tomas, Kai Chen, Greg Corrado, and Jeff Dean. 2013. Efficient estimation of word representations in vector space. ArXiv Preprint, arXiv:1301.3781. Mikolov, Tomas, Ilya Sutskever, Kai Chen, Greg Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. Advances in Neural Information Processing Systems (NeurIPS) 26. Parts I - II, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Word Embeddings Trained on Google News Corpus Hand Written Notes Session 6. Deep-Learning-Based NLP: RNN and Seq2Seq (Feb/20/2024) Keywords: Word2Vec: GloVe, Word Embedding and Language Model Evaluations, Word2Vec and RNN Applied to Business/Econ Research, RNN Slides: Guest Lecture Announcement, NLP(III): Word2Vec.pdf), NLP(IV): RNN & Seq2Seq.pdf) CoLab Notebook Demos: Word2Vec: CBOW, Word2Vec: Skip-Gram Presentation: By Qiyu Dai and Yifan Ren. Huang, Allen H., Hui Wang, and Yi Yang. 2023. FinBERT: A large language model for extracting information from financial text. Contemporary Accounting Research, 40(2): 806-841. Link to the paper. Link to GitHub Repo. Homework: Problem Set 4 - Word2Vec & LSTM for Sentiment Analysis References: Ash, Elliot, and Stephen Hansen. 2023. Text algorithms in economics. Annual Review of Economics, 15: 659-688. Associated GitHub with Code Demonstrations. Li, Kai, Feng Mai, Rui Shen, and Xinyan Yan. 2021. Measuring corporate culture using machine learning. Review of Financial Studies, 34(7): 3265-3315. Chen, Fanglin, Xiao Liu, Davide Proserpio, and Isamar Troncoso. 2022. Product2Vec: Leveraging representation learning to model consumer product choice in large assortments. Available at SSRN 3519358. Pennington, Jeffrey, Richard Socher, and Christopher Manning. 2014. Glove: Global vectors for word representation. Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP) (pp. 1532-1543). Parts 2 and 5, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Chapters 9 and 10, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. RNN and LSTM Visualizations Hand Written Notes Session 7. Deep-Learning-Based NLP: Attention and Transformer (Feb/27/2024) Keywords: RNN and its Applications to Business/Econ Research, LSTM, Seq2Seq, Attention Mechanism Slides: Final Project, NLP(IV): RNN & Seq2Seq.pdf), NLP(V): Attention & Transformer.pdf) CoLab Notebook Demos: RNN & LSTM, Attention Mechanism Presentation: By Qinghe Gui and Chaoyuan Jiang. Zhang, Mengxia and Lan Luo. 2023. Can consumer-posted photos serve as a leading indicator of restaurant survival? Evidence from Yelp. Management Science 69(1): 25-50. Link to the paper. Homework: Problem Set 4 - Word2Vec & LSTM for Sentiment Analysis References: Qi, Meng, Yuanyuan Shi, Yongzhi Qi, Chenxin Ma, Rong Yuan, Di Wu, Zuo-Jun (Max) Shen. 2023. A Practical End-to-End Inventory Management Model with Deep Learning. Management Science, 69(2): 759-773. Sarzynska-Wawer, Justyna, Aleksander Wawer, Aleksandra Pawlak, Julia Szymanowska, Izabela Stefaniak, Michal Jarkiewicz, and Lukasz Okruszek. 2021. Detecting formal thought disorder by deep contextualized word representations. Psychiatry Research, 304, 114135. Hansen, Stephen, Peter J. Lambert, Nicholas Bloom, Steven J. Davis, Raffaella Sadun, and Bledi Taska. 2023. Remote work across jobs, companies, and space (No. w31007). National Bureau of Economic Research. Sutskever, Ilya, Oriol Vinyals, and Quoc V. Le. 2014. Sequence to sequence learning with neural networks. Advances in neural information processing systems, 27. Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. ICLR Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... and Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30. Parts 5, 6, and 8, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Chapters 9, 10, and 11, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. RNN and LSTM Visualizations PyTorch's Tutorial of Seq2Seq for Machine Translation Illustrated Transformer Transformer from Scratch, with the Code on GitHub Hand Written Notes Session 7.5. Deep-Learning-Based NLP: Attention is All You Need (Mar/05/2024) Keywords: Bitter Lesson: Power of Computation in AI, Attention Mechanism, Transformer Slides: The Bitter Lesson, NLP(V): Attention & Transformer.pdf) CoLab Notebook Demos: Attention Mechanism, Transformer Homework: One-page Proposal for Your Final Project References: The Bitter Lesson, by Rich Sutton Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. ICLR Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... and Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30. Part 8, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Chapter 11, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. Illustrated Transformer Transformer from Scratch, with the Code on GitHub Andrej Karpathy's Lecture to Build Transformers Hand Written Notes Session 8. Deep-Learning-Based NLP: Pretraining (Mar/12/2024) Keywords: Computations in AI, BERT (Bidirectional Encoder Representations from Transformers), GPT (Generative Pretrained Transformers) Slides: Guest Lecture by Dr. Liubo Li on Deep Learning Computation, Pretraining.pdf) CoLab Notebook Demos: Crafting Intelligence: The Art of Deep Learning Modeling, BERT API @ Hugging Face Presentation: By Zhankun Chen and Yiyi Zhao. Noy, Shakked and Whitney Zhang. 2023. Experimental evidence on the productivity effects of generative artificial intelligence. Science, 381: 187-192. Link to the Paper Homework: Problem Set 5 - Sentiment Analysis with Hugging Face, due at 12:30pm, March 26, Tuesday. References: Devlin, Jacob, Ming-Wei Chang, Kenton Lee, Kristina Toutanova. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. ArXiv preprint arXiv:1810.04805. GitHub Repo Radford, Alec, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. Improving language understanding by generative pre-training, (GPT-1) PDF link, GitHub Repo Radford, Alec, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI blog, 1(8), 9. (GPT-2) PDF Link, GitHub Repo Brown, Tom, et al. 2020. Language models are few-shot learners. Advances in neural information processing systems, 33, 1877-1901. (GPT-3) GitHub Repo Huang, Allen H., Hui Wang, and Yi Yang. 2023. FinBERT: A large language model for extracting information from financial text. Contemporary Accounting Research, 40(2): 806-841. GitHub Repo Part 9, Lecture Notes and Slides for CS 224N: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto. Link to CS 224N Part 2 & 4, Slides for COS 597G: Understanding Large Language Models, by Danqi Chen. Link to COS 597G A Visual Guide to BERT, How GPT-3 Works Andrej Karpathy's Lecture to Build GPT-2 (124M) from Scratch Hand Written Notes Session 9. Deep-Learning-Based NLP: Large Language Models (Mar/19/2024) Keywords: Large Language Models, Generative AI, Emergent Ababilities, Instruction Fine-Tuning (IFT), Reinforcement Learning with Human Feedback (RLHF), In-Context Learning, Chain-of-Thought (CoT) Slides: What's Next, Pretraining.pdf), Large Language Models.pdf) CoLab Notebook Demos: BERT API @ Hugging Face Presentation: By Jia Liu. Liu, Liu, Dzyabura, Daria, Mizik, Natalie. 2020. Visual listening in: Extracting brand image portrayed on social media. Marketing Science, 39(4): 669-686. Link to the Paper Homework: Problem Set 5 - Sentiment Analysis with Hugging Face, due at 12:30pm, March 26, Tuesday (soft-deadline). References: Wei, Jason, et al. 2021. Finetuned language models are zero-shot learners. ArXiv preprint arXiv:2109.01652, link to the paper. Wei, Jason, et al. 2022. Emergent abilities of large language models. ArXiv preprint arXiv:2206.07682, link to the paper. Ouyang, Long, et al. 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35, 27730-27744. Wei, Jason, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35, 24824-24837. Kaplan, Jared. 2020. Scaling laws for neural language models. ArXiv preprint arXiv:2001.08361, link to the paper. Hoffmann, Jordan, et al. 2022. Training compute-optimal large language models. ArXiv preprint arXiv:2203.15556, link to the paper. Shinn, Noah, et al. 2023. Reflexion: Language agents with verbal reinforcement learning. ArXiv preprint arXiv:2303.11366, link to the paper. Reisenbichler, Martin, Thomas Reutterer, David A. Schweidel, and Daniel Dan. 2022. Frontiers: Supporting content marketing with natural language generation. Marketing Science, 41(3): 441-452. Romera-Paredes, B., Barekatain, M., Novikov, A. et al. 2023. Mathematical discoveries from program search with large language models. Nature, link to the paper. Part 10, Lecture Notes and Slides for CS224N: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto. Link to CS 224N COS 597G: Understanding Large Language Models, by Danqi Chen. Link to COS 597G Andrej Karpathy's 1-hour Talk on LLM CS224n, Hugging Face Tutorial Session 10. Deep-Learning-Based CV: Image Classification (Mar/26/2024) Keywords: Large Language Models Applications, Convolution Neural Nets (CNN), LeNet, AlexNet, VGG, ResNet, ViT Slides: What's Next, Large Language Models.pdf), Image Classification.pdf) CoLab Notebook Demos: CNN, LeNet, & AlexNet, VGG, ResNet, ViT Presentation: By Yingxin Lin and Zeshen Ye. Netzer, Oded, Alain Lemaire, and Michal Herzenstein. 2019. When words sweat: Identifying signals for loan default in the text of loan applications. Journal of Marketing Research, 56(6): 960-980. Link to the Paper Homework: Problem Set 6 - AlexNet and ResNet, due at 12:30pm, April 9, Tuesday. References: Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton. 2012. Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems, 25. He, Kaiming, Xiangyu Zhang, Shaoqing Ren and Jian Sun. 2016. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition, 770-778. Dosovitskiy, Alexey, et al. 2020. An image is worth 16x16 words: Transformers for image recognition at scale. ArXiv preprint, arXiv:2010.11929, link to the paper, link to the GitHub repo. Jean, Neal, Marshall Burke, Michael Xie, Matthew W. Davis, David B. Lobell, and Stefand Ermon. 2016. Combining satellite imagery and machine learning to predict poverty. Science, 353(6301), 790-794. Zhang, Mengxia and Lan Luo. 2023. Can consumer-posted photos serve as a leading indicator of restaurant survival? Evidence from Yelp. Management Science 69(1): 25-50. Course Notes (Lectures 5 & 6) for CS231n: Deep Learning for Computer Vision, by Fei-Fei Li, Ruohan Gao, & Yunzhu Li. Link to CS231n. Chapters 7 and 8, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola. Link to the book. Fine-Tune ViT for Image Classification with Hugging Face 🤗 Transformers Hugging Face 🤗 ViT CoLab Tutorial Session 11. Deep-Learning-Based CV (II): Object Detection & Video Analysis (Apr/2/2024) Keywords: Image Processing Applications, Localization, R-CNNs, YOLOs, Semantic Segmentation, 3D CNN, Video Analysis Applications Slides: What's Next, Image Classification.pdf), Object Detection and Video Analysis.pdf) CoLab Notebook Demos: Data Augmentation, Faster R-CNN & YOLO v5 Presentation: By Qinlu Hu and Yilin Shi. Yang, Jeremy, Juanjuan Zhang, and Yuhan Zhang. 2023. Engagement that sells: Influencer video advertising on TikTok. Available at SSRN Link to the Paper Homework: Problem Set 6 - AlexNet and ResNet, due at 12:30pm, April 9, Tuesday. References: Girshick, R., Donahue, J., Darrell, T. and Malik, J., 2014. Rich feature hierarchies for accurate object detection and semantic segmentation. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 580-587). Redmon, Joseph, Santosh Divvala, Ross Girshick, and Ali Farhadi. 2016. You only look once: Unified, real-time object detection. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 779-788). Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R. and Fei-Fei, L., 2014. Large-scale video classification with convolutional neural networks. Proceedings of the IEEE conference on Computer Vision and Pattern Recognition (pp. 1725-1732). Glaeser, Edward L., Scott D. Kominers, Michael Luca, and Nikhil Naik. 2018. Big data and big cities: The promises and limitations of improved measures of urban life. Economic Inquiry, 56(1): 114-137. Zhang, S., Xu, K. and Srinivasan, K., 2023. Frontiers: Unmasking Social Compliance Behavior During the Pandemic. Marketing Science, 42(3), pp.440-450. Course Notes (Lectures 10 & 11) for CS231n: Deep Learning for Computer Vision, by Fei-Fei Li, Ruohan Gao, & Yunzhu Li. Link to CS231n. Chapter 14, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola. Link to the book. Hand Written Notes Session 12. Unsupervised Learning: Clustering, Topic Modeling & VAE (Apr/9/2024) Keywords: K-Means, Gaussian Mixture Models, EM-Algorithm, Latent Dirichlet Allocation, Variational Auto-Encoder Slides: What's Next, Clustering, Topic Modeling & VAE.pdf) CoLab Notebook Demos: K-Means, LDA, VAE Homework: Problem Set 7 - Unsupervised Learning (EM & LDA), due at 12:30pm, April 23, Tuesday. References: Blei, David M., Ng, Andrew Y., and Jordan, Michael I. 2003. Latent Dirichlet allocation. Journal of Machine Learning Research, 3(Jan): 993-1022. Kingma, D.P. and Welling, M., 2013. Auto-encoding Variational Bayes. arXiv preprint arXiv:1312.6114. Kingma, D.P. and Welling, M., 2019. An introduction to variational autoencoders. Foundations and Trends® in Machine Learning, 12(4), pp.307-392. Bandiera, O., Prat, A., Hansen, S., & Sadun, R. 2020. CEO behavior and firm performance. Journal of Political Economy, 128(4), 1325-1369. Liu, Jia and Olivier Toubia. 2018. A semantic approach for estimating consumer content preferences from online search queries. Marketing Science, 37(6): 930-952. Mueller, Hannes, and Christopher Rauh. 2018. Reading between the lines: Prediction of political violence using newspaper text. American Political Science Review, 112(2): 358-375. Tian, Z., Dew, R. and Iyengar, R., 2023. Mega or Micro? Influencer Selection Using Follower Elasticity. Journal of Marketing Research. Chapters 8.5 and 14, The Elements of Statistical Learning (2nd Edition), 2009, by Trevor Hastie, Robert Tibshirani, Jerome Friedman, Link to Book. Course Notes (Lectures 1 & 4) for CS294-158-SP24: Deep Unsupervised Learning, taught by Pieter Abbeel, Wilson Yan, Kevin Frans, Philipp Wu. Link to CS294-158-SP24. Hand Written Notes Session 13. Unsupervised Learning: Diffusion Models (Apr/16/2024) Keywords: VAE, Denoised Diffusion Probabilistic Models, Latent Diffusion Models, CLIP, Imagen, Diffusion Transformers Slides: Clustering, Topic Modeling & VAE.pdf), Diffusion Models.pdf), Course Summary CoLab Notebook Demos: VAE, DDPM, DiT Homework: Problem Set 7 - Unsupervised Learning (EM & LDA), due at 12:30pm, April 23, Tuesday. References: Kingma, D.P. and Welling, M., 2013. Auto-encoding Variational Bayes. arXiv preprint arXiv:1312.6114. Kingma, D.P. and Welling, M., 2019. An introduction to variational autoencoders. Foundations and Trends® in Machine Learning, 12(4), pp.307-392. Ho, J., Jain, A. and Abbeel, P., 2020. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33, 6840-6851. Chan, S.H., 2024. Tutorial on Diffusion Models for Imaging and Vision. arXiv preprint arXiv:2403.18103. Peebles, W. and Xie, S., 2023. Scalable diffusion models with transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 4195-4205. Link to GitHub Repo. Tian, Z., Dew, R. and Iyengar, R., 2023. Mega or Micro? Influencer Selection Using Follower Elasticity. Journal of Marketing Research. Ludwig, J. and Mullainathan, S., 2024. Machine learning as a tool for hypothesis generation. Quarterly Journal of Economics, 139(2), 751-827. Burnap, A., Hauser, J.R. and Timoshenko, A., 2023. Product aesthetic design: A machine learning augmentation. Marketing Science, 42(6), 1029-1056. Course Notes (Lecture 6) for CS294-158-SP24: Deep Unsupervised Learning, taught by Pieter Abbeel, Wilson Yan, Kevin Frans, Philipp Wu. Link to CS294-158-SP24. CVPR 2022 Tutorial: Denoising Diffusion-based Generative Modeling: Foundations and Applications, by Karsten Kreis, Ruiqi Gao, and Arash Vahdat Link to the Tutorial Lilian Weng (OpenAI)'s Blog on Diffusion Models Lilian Weng (OpenAI)'s Blog on Diffusion Models for Video Generation Hugging Face Diffusers 🤗 Library Hand Written Notes

voicefilter
github
LLM Vibe Score0.496
Human Vibe Score0.029786815978503328
maum-aiMar 24, 2025

voicefilter

VoiceFilter Note from Seung-won (2020.10.25) Hi everyone! It's Seung-won from MINDs Lab, Inc. It's been a long time since I've released this open-source, and I didn't expect this repository to grab such a great amount of attention for a long time. I would like to thank everyone for giving such attention, and also Mr. Quan Wang (the first author of the VoiceFilter paper) for referring this project in his paper. Actually, this project was done by me when it was only 3 months after I started studying deep learning & speech separation without a supervisor in the relevant field. Back then, I didn't know what is a power-law compression, and the correct way to validate/test the models. Now that I've spent more time on deep learning & speech since then (I also wrote a paper published at Interspeech 2020 😊), I can observe some obvious mistakes that I've made. Those issues were kindly raised by GitHub users; please refer to the Issues and Pull Requests for that. That being said, this repository can be quite unreliable, and I would like to remind everyone to use this code at their own risk (as specified in LICENSE). Unfortunately, I can't afford extra time on revising this project or reviewing the Issues / Pull Requests. Instead, I would like to offer some pointers to newer, more reliable resources: VoiceFilter-Lite: This is a newer version of VoiceFilter presented at Interspeech 2020, which is also written by Mr. Quan Wang (and his colleagues at Google). I highly recommend checking this paper, since it focused on a more realistic situation where VoiceFilter is needed. List of VoiceFilter implementation available on GitHub: In March 2019, this repository was the only available open-source implementation of VoiceFilter. However, much better implementations that deserve more attention became available across GitHub. Please check them, and choose the one that meets your demand. PyTorch Lightning: Back in 2019, I could not find a great deep-learning project template for myself, so I and my colleagues had used this project as a template for other new projects. For people who are searching for such project template, I would like to strongly recommend PyTorch Lightning. Even though I had done a lot of effort into developing my own template during 2019 (VoiceFilter -> RandWireNN -> MelNet -> MelGAN), I found PyTorch Lightning much better than my own template. Thanks for reading, and I wish everyone good health during the global pandemic situation. Best regards, Seung-won Park Unofficial PyTorch implementation of Google AI's: VoiceFilter: Targeted Voice Separation by Speaker-Conditioned Spectrogram Masking. Result Training took about 20 hours on AWS p3.2xlarge(NVIDIA V100). Audio Sample Listen to audio sample at webpage: http://swpark.me/voicefilter/ Metric | Median SDR | Paper | Ours | | ---------------------- | ----- | ---- | | before VoiceFilter | 2.5 | 1.9 | | after VoiceFilter | 12.6 | 10.2 | SDR converged at 10, which is slightly lower than paper's. Dependencies Python and packages This code was tested on Python 3.6 with PyTorch 1.0.1. Other packages can be installed by: Miscellaneous ffmpeg-normalize is used for resampling and normalizing wav files. See README.md of ffmpeg-normalize for installation. Prepare Dataset Download LibriSpeech dataset To replicate VoiceFilter paper, get LibriSpeech dataset at http://www.openslr.org/12/. train-clear-100.tar.gz(6.3G) contains speech of 252 speakers, and train-clear-360.tar.gz(23G) contains 922 speakers. You may use either, but the more speakers you have in dataset, the more better VoiceFilter will be. Resample & Normalize wav files First, unzip tar.gz file to desired folder: Next, copy utils/normalize-resample.sh to root directory of unzipped data folder. Then: Edit config.yaml Preprocess wav files In order to boost training speed, perform STFT for each files before training by: This will create 100,000(train) + 1000(test) data. (About 160G) Train VoiceFilter Get pretrained model for speaker recognition system VoiceFilter utilizes speaker recognition system (d-vector embeddings). Here, we provide pretrained model for obtaining d-vector embeddings. This model was trained with VoxCeleb2 dataset, where utterances are randomly fit to time length [70, 90] frames. Tests are done with window 80 / hop 40 and have shown equal error rate about 1%. Data used for test were selected from first 8 speakers of VoxCeleb1 test dataset, where 10 utterances per each speakers are randomly selected. Update: Evaluation on VoxCeleb1 selected pair showed 7.4% EER. The model can be downloaded at this GDrive link. Run After specifying traindir, testdir at config.yaml, run: This will create chkpt/name and logs/name at base directory(-b option, . in default) View tensorboardX Resuming from checkpoint Evaluate Possible improvments Try power-law compressed reconstruction error as loss function, instead of MSE. (See #14) Author Seungwon Park at MINDsLab (yyyyy@snu.ac.kr, swpark@mindslab.ai) License Apache License 2.0 This repository contains codes adapted/copied from the followings: utils/adabound.py from https://github.com/Luolc/AdaBound (Apache License 2.0) utils/audio.py from https://github.com/keithito/tacotron (MIT License) utils/hparams.py from https://github.com/HarryVolek/PyTorchSpeakerVerification (No License specified) utils/normalize-resample.sh from https://unix.stackexchange.com/a/216475

business-document-processing
github
LLM Vibe Score0.341
Human Vibe Score0.023080316664879252
SAPMar 21, 2025

business-document-processing

Python Client Library for the SAP AI Business Services: Document Classification and Document Information Extraction This repository contains the source code of a Python client library to facilitate the use of the SAP AI Business Services: Document Classification and Document Information Extraction. The client library provides two API Client classes that contain convenient methods to access these services and issue calls to the Document Classification REST API and Document Information Extraction REST API respectively. To use the library you need to have access to SAP Business Technology Platform. Check out the usage examples, they are very useful to get started with the services. Have a look at API documentation in order to use the library. Notes for users of the sap-document-classification-client library This library includes all the capabilities of the sap-document-classification-client, which will not be developed further. However, the code is still available here. If you want to switch to this library, you have to be aware of the following changes: The DCApiClient can now be imported directly from the top module via: The functions , , now return an iterator instead of a list. You can either analyze individual results using with within a try-catch block (e.g. to handle each failed document) or use to turn it to a list. The latter will raise an error if at least one document failed. The function now returns a list which is the "dataset" part of the API response json. (You just need to delete the \["dataset"\] from the response to work with it as until now) The function now returns a list which is the "results" part of the API response json. The function now returns a list which is the "models" part of the API response json. The function now returns a list which is the "deployments" part of the API response json. The library now raises the following custom exceptions: BDPApiException: Base exception for all exceptions of this library. Raise when no other exception is applicable. BDPClientException: Raised when an HTTP response with status code between 400 and 500 is returned. Usually means incorrect user input. (Replaces some HTTPErrors) BDPServerException: Raised when an HTTP response with status code between 500 and 600 is returned. Usually means that the server had some internal error. (Replaces some HTTPErrors) BDPUnauthorizedException: Raised when an HTTP response with status code 401 is returned. Usually means that a wrong OAuth credentials were provided. BDPFailedAsynchronousOperationException: Raised when an asynchronous job failed during processing. (Replaces FailedCallException) BDPPollingTimeoutException: Raised when an asynchronous job exceeds the set pollingmaxattempts. (Replaces PollingTimeoutException) The function now doesnt expect an 'url' and 'payload' parameters, but 'path' and 'json' parameters instead. Requirements This library requires properly setup Python 3.6 (or higher version) environment. Download and Installation This Python library should be consumed in the standard way by running or adding the library as a dependency of your code in requirements.txt` file. Demo usage Prerequisites: Get a Free Account on SAP BTP Trial Create Service Instance for Document Classification with Trial Account Create Service Instance for Document Information Extraction Document Classification To try out the Document classification service using the document classification client library you can also run the two demo links below: Try out classification using default model demo Try out training and classification using custom model demo (requires an enterprise account, trial account is not sufficient) Document Information Extraction Try out the Document Information Extraction service with this showcase Exercises Exercise 1 - Set up Document Information Extraction Service and UI Exercise 2 - Upload a document for extraction using UI application Exercise 3 - Visualize, correct extraction results and confirm document using UI application Exercise 4 - Get Auth token to use Document Information Extraction Rest API Exercise 5 - Get extraction results of document using Rest API Exercise 6 - Upload supplier Data for matching Exercise 7 - Upload document through Rest API to enrich the extraction Results with supplier data Known Issues Please see the issues section. How to obtain support In case you would like to contribute to this project, ask any questions or get support, please open an issue containing the description of your question or planned contribution in GitHub and we will get in touch. Licensing Please see our LICENSE for copyright and license information. Detailed information including third-party components and their licensing/copyright information is available via the REUSE tool.

Vibe Coding is Actually INSANE... (Vibe Coding Tutorial for Beginners)
youtube
LLM Vibe Score0.361
Human Vibe Score0.67
MemoryMar 21, 2025

Vibe Coding is Actually INSANE... (Vibe Coding Tutorial for Beginners)

🖼️ Infographic: https://memstechtips.gumroad.com/l/vibecoding Vibe Coding is Actually INSANE... (Vibe Coding Tutorial for Beginners) What is vibe coding? How to vibe code? Those are questions more and more people are asking these days due to the crazy rate at which agentic AI models like Claude 3.7 Sonnet are evolving every single day. In this vibe coding tutorial video, I give you a comprehensive overview and explanation of what vibe coding is, how you can get started with vibe coding, which tools to use and how to prompt these AI models to get the best results. I also show you step by step how you can install VS Code and configure the Cline coding extension with free API's from OpenRouter, so you can start coding apps for free ASAP! 📝 Website Article 🔗 https://memstechtips.com/vibe-coding-ai-powered-programming-guide/ 📺 RELATED VIDEOS 👉 https://www.youtube.com/playlist?list=PL8RYOts8u1Ut2PhX5z5FSwHaIDZrd0xHW 👉 https://www.youtube.com/playlist?list=PL8RYOts8u1Uu5xVLyE3r8TYjOR0I4chEZ 👉 https://www.youtube.com/playlist?list=PL8RYOts8u1UujBoTKVcz3HmybIWu86OZ7 🤝 WANNA SAY THANKS? 🔗 https://paypal.me/memstech 🔗 https://www.youtube.com/@memstechtips/join 👥 JOIN MY DISCORD COMMUNITY 🔗 https://discord.gg/zWGANV8QAX 🌐 CONNECT WITH ME 🔗 https://linktr.ee/memstechtips ⏱️ CHAPTERS: 00:00 - What is Vibe Coding? 02:28 - Key Tools and Technologies 04:00 - Setup Requirements and Benefits 05:14 - Quick Start Workflow and Common Pitfalls 08:31 - Step-by-Step Setup Guide (VS Code & Cline) 12:11 - Creating a CWPF Application Example 19:19 - Creating a Simple Website Example 27:22 - Comparing AI Models (DeepSeek vs Claude) 34:00 - Final Thoughts and Conclusion ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ DISCLAIMER: This video is for educational purposes only and demonstrates general troubleshooting techniques and procedures. I cannot be held responsible for any damage caused to your computer or software by following these steps. Use this information at your own risk. It is always advisable to seek professional assistance if you are not comfortable performing these procedures yourself. Additionally, some software and tools featured in this video may have specific licensing requirements or limitations. Please ensure you are using them in accordance with their respective terms of use. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ #vibecoding #cline #claudesonnet

What is Vibe Coding, should you Learn It?
youtube
LLM Vibe Score0.419
Human Vibe Score0.88
Stefan MischookMar 20, 2025

What is Vibe Coding, should you Learn It?

Vibe coding is coding with Ai as the Ai creates the boilerplate code for you. But does that mean you don't need to understand coding/development? #vibecoding #aidevelopment 🔥 STEF'S DEVELOPER BOOTCAMP AND MENTORING PROGRAM https://unclestef.com/ 📽️ Get your questions answered, sponsor a video: https://unclestef.com/blog/2025/03/04/sponsored-video-request/ 🎤 Listen to my Uncle Stef podcasts: https://unclestef.com/blog/2024/07/26/uncle-stef-podcast-all-episodes/ 🔥 JOIN STEF'S 'CODER'S CAREER PATHS' NEWSLETTER: https://newsletters.stefanmischook.com/coderscareerpaths_signup 🔥 FREE: LIZARD WIZARD KOMODO - TRANSFORMATIONAL MIND TRAINING: https://newsletters.stefanmischook.com/komodo Channel Discord Server: https://discord.gg/rn8za8aq2v WEB HOST PAYS FOR YOUR WEB DESIGN TRAINING IN 2023: https://www.killersites.com/blog/2020/web-hosting-company-pays-for-your-web-design-training/ POPULAR & EASY CODING COURSES: Full stack web developer course: https://school.studioweb.com/store/course/completewebdeveloper Python 3 Foundations & Certification: https://school.studioweb.com/store/course/python3foundations&certificationpackage Complete Freelancer: https://school.studioweb.com/store/course/complete_freelancer Complete Entrepreneur: https://school.studioweb.com/store/course/completewebentrepreneur 🦎 Lizard Wizard Course: https://school.studioweb.com/store/course/lizard_wizard 📚 BOOKS TO READ: My Beginners HTML5, CSS3: https://amzn.to/2wKsVTh … Complements Studioweb courses on HTML5, CSS3 and JavaScript. Refactoring: Improving the Design of Existing Code (2nd Edition) https://amzn.to/3o5cTbw HeadFirst Design Patterns: https://amzn.to/2LQ0Gdh Java Refactoring: Improving the Design of Existing Code (1st Edition) https://amzn.to/3a9nSsZ The Naked Ape: https://amzn.to/3fhS1Lj ✉️ STAY IN CONTACT: Stef's social links: Instagram: https://www.instagram.com/stefanmischook/?hl=en Twitter: https://twitter.com/killersites Stef's business channel: https://www.youtube.com/channel/UCZdr0ql_B240VBVINAX7Acg 👉 GOOGLE REVIEW: https://g.page/studioWebedu/review?mt Leave a Google review about Stef. MY MOUSE & KEYBOARD: Logitech Keyboard I use: https://amzn.to/38jYDqE Logitech mouse I use: https://amzn.to/2IeVvBj SUPPLEMENTS THAT WORK AMAZING FOR ME: Protein Essentials Beef Gelatine Powder: https://amzn.to/2Pf52vL ... Healed my very bad knee. If you have joint problems, this *could do miracles for you. Webber Naturals 88862 Glucosamine Chondroitin https://amzn.to/3ss9WEa MY CAMERA GEAR: Godox VL150 lights: https://amzn.to/3lhsYZP Sigma 18-35 lens: https://amzn.to/33sRh0T Canon EOS C70 Cinema Camera Thanks! Stef #mentoring #codecourses #unclestef #codingcoach

AI Created My Game in 6 Hours... Vibe Coding so you don't have to
youtube
LLM Vibe Score0.384
Human Vibe Score0.57
Mos NassarMar 20, 2025

AI Created My Game in 6 Hours... Vibe Coding so you don't have to

Vibe Coding with AI promised to revolutionize game development, so I tested it for 6 hours straight. I let AI build my entire Game from scratch. I'm sacrificing my time so you don't have to Steam Page: https://store.steampowered.com/app/3161930/DuckNorrisTales/ Duck Norris Tales demo : https://octomos.itch.io/duck-norris-cluckys-nightmare Live Da Life discord : https://discord.gg/RfSFtBHy3z Website : https://livedalife.io/ Twitter : https://twitter.com/livedalifeacad Main Twitter: https://twitter.com/MosNassar_ My Gear: Sony a7iii: https://amzn.to/3nJNALH Sony 24-105mm f/4.0 : https://amzn.to/38L7Lo2 Sony FE 28 mm f/2-22: https://amzn.to/35MBviF SONY FE 55mm F1.8 Z: https://amzn.to/3nNsOdW Aputure AL-MC : https://amzn.to/38PlREQ Learn how I edit my videos from this FREE ONLINE COURSE (14 days free trial link) : https://skl.sh/2W9aBMK Color grading Course : https://skl.sh/2QQijM6 Get 30 Days free trial of Epidemic Sound: https://epidemicsound.com/referral/xkr7og/ GET IN TOUCH : Instagram - https://www.instagram.com/mosnassar_ Twitter - https://twitter.com/MosNassar_ My Merch : https://teespring.com/stores/uncover If you enjoyed this video don't forget to LIKE, COMMENT, and smash that SUBSCRIBE button! DISCLAIMER: Links above include affiliate commission or referrals. I'm part of an affiliate network and I receive small compensation from partnering websites (at no extra cost to you!) and keep this channel going. Thanks for your support!

bubbln_network-automation
github
LLM Vibe Score0.421
Human Vibe Score0.004537250556463098
olasupoMar 14, 2025

bubbln_network-automation

Bubbln: An AI-driven Network Automation In the world of network engineering, automation has completely transformed the way things work. But, before automation, setting up and managing networks was a tedious job filled with challenges. Engineers had to manually type out configurations, often doing the same tasks repeatedly on different devices. This led to mistakes and wasted time. Then came automation tools like Ansible, Chef, and Puppet, which changed everything. They made network management much easier and allowed for scalability. But there was still a problem: creating automation scripts required a lot of technical know-how and was prone to errors because it relied on human input. And that's why we built Bubbln. It's a game-changer in network engineering, integrating AI into Ansible to take automation to the next level. With Bubbln, we can automatically generate and execute playbooks with incredible accuracy, thereby improving automation efficiency and increasing network engineer’s productivity. It was developed using Python programming language and acts as a bridge between ChatGPT and network systems, making interactions seamless and deployments effortless. Current Capabilities AI-Driven Playbook Generation for OSPF and EIGRP based networks: Bubbln has been rigorously tested to leverage ChatGPT for generation of playbooks for networks based on OSPF and EIGRP networks, with a very high accuracy rate. Auto-creation of Inventory files: Users do not need to prepare the hosts file. Bubbln will auto-generate this file from input provided by the user. Customizable Configurations: Users can input specific router protocols (OSPF or EIGRP), interface configurations, and other network details to tailor the generated playbooks. Documentation: Bubbln automatically creates a report that contains the network configurations, prompts, and generated playbooks for easy reference in future. No expertise required: By auto-generation of the playbooks and inventory file, Bubbln has been able to eliminate a major hurdle to network automation – need for users to learn the automation tools e.g Ansible, Chef. Improved Efficiency: With AI automation, Bubbln speeds up the deployment of network configurations, reducing the time required for manual playbook creation, thereby increasing the productivity of network engineers. Getting Started There are two main approaches to installing Bubbln on your local machine. Docker Container Bubbln has been packaged using docker containers for easy distribution and usage. The following steps can be followed to deploy the Bubbln container on your local machine. Ensure docker is installed on your local machine by entering the below command. This command works for windows and linux OS: The version of docker would be displayed if it is installed. Otherwise, please follow the link below to install docker on your machine: Windows: Docker Desktop for Windows Ubuntu: Docker Engine for Ubuntu CentOS: Docker Engine for CentOS Debian: Docker Engine for Debian Fedora: Docker Engine for Fedora Download the docker image: Create a directory for the project and download Bubbln image using the below command: Run the docker container using the below command: Install nano Update the sshipaddresses.txt file: Update the ssh_addresses.txt file with the SSH IP addresses of the routers you want to configure. Bubbln will utilize this information along with the login credentials (inputted at runtime) to automatically generate a hosts.yml file required by ansible for network configuration. To do this enter the below command to edit the file: Obtain an OpenAPI API Key: You may follow this guide to sign up and obtain an API key: Utilizing a Virtualization machine of choice, setup a network with the following basic configurations: Enable SSH on each of the routers. Configure IP addresses and enable only interfaces required for connectivity by Bubbln. Configure static routes to enable Bubbln reach the routers on the network. Ensure all the routers can be reached by ping and SSH from your host machine. Initialize Bubbln by entering the below command: Github Repository Clone You can clone Bubbln’s GitHub repository by following the below steps: Prerequisites Bubbln works well with Python 3.10. You need to ensure python3.10 is installed on your local machine. This can be confirmed by entering the below command: If it is not Installed, then the below command can be utilized to install python 3.10: Build and Prepare the Project Clone the Bubbln repository from GitHub: To clone the repository, first verify you have git installed on your machine by issuing the following commands: If git is installed, the version number would be displayed, otherwise, you can issue the following commands to have git installed on your machine: Navigate or create a directory for the project on your machine and issue the following commands to clone the Bubbln git repository: Create a Virtual Environment for the application Firstly, confirm virtualenv is installed on your machine by inputting the following command: If the output shows something similar to the below, then go to the next step to install virtualenv ` WARNING: Package(s) not found: env, virtual ` Issue the below command to install virtualenv: Create a virtual environment for the project: Activate the virtual environment: Install the dependencies You can then run the below command to install the necessary packages for the app. Update the sshipaddresses.txt file: Update the ssh_addresses.txt file with the SSH IP addresses of the routers you want to configure. Bubbln will utilize this information along with the login credentials (inputted at runtime) to automatically generate a hosts.yml file required by ansible for network configuration. Obtain an OpenAPI API Key: You may follow this guide to sign up and obtain an API key OpenAI Key: OpenAI Key Utilizing a Virtualization machine of choice, setup a network with the following basic configurations: Enable SSH on each of the routers. Configure IP addresses and enable only interfaces required for connectivity by Bubbln Configure static routes to enable Bubbln reach the routers on the network. Ensure all the routers can be reached by ping and SSH from your host machine. Initialize Bubbln While ensuring that python virtual environment is activated as stated in step 5, run the below command to initialize Bubbln How Bubbln Works Bubbln serves as an intermediary between ChatGPT and a network infrastructure, providing logic, control functions, and facilitating network automation. Its operation can be summarized as follows: !image Figure 1Bubbln architecture and interaction with a network of four routers. Initialization: When Bubbln is initialized, it checks the “userconfig.pkl” file to see if Bubbln has ever been initiated. This is indicated by the presence of a welcome message status in the file. If it exists, Bubbln jumps straight to request the user to input the OpenAI key. Otherwise, it displays a welcome message, and updates the userconfig.pkl file accordingly. Upon successful input of the API key, the user is prompted for the SSH credentials of the routers. These parameters are then encrypted and saved in the user_config.pkl file. The SSH credential is later decrypted and parsed as input to dynamically generate a hosts.yml file at runtime. Responsible Code Section: bubbln.py: welcomemessagefeature() !image Figure 2 Bubbln's welcome message. Parameter Input & Validation: In the parameter input stage, Bubbln first checks for the existence of a file called “router_configuration.pkl”. If it exists, the user is prompted to decide whether to load an existing configuration or input a new set of configurations. If the file is empty or non-existent, then users are prompted to input the configuration parameters for each router on the network. These parameters serve as variables that are combined with hardcoded instructions written in natural language to form the prompt sent to ChatGPT. Key parameters include: Router Configurations: OSPF Area OSPF Process ID Number of networks to advertise (OSPF/EIGRP) AS Number (EIGRP) Interface names IP Addresses (in CIDR format) This module also ensures that parameters are keyed in using the correct data type and format e.g. IP addresses are expected in CIDR format and OSPF Area should be of type integer. Upon completion of parameter input, all parameters are saved into a file called “router_configuration.pkl” upon validation of accuracy by the user. Responsible Code Section: parameter_input.py !image Figure 3 Bubbln receiving Network Parameters. Before generating the prompt, a summary of the inputted parameters is displayed for user validation. This step ensures accuracy and minimizes errors. Users are given the option to make corrections if any discrepancies are found. Responsible Code Section: parameterinput.py: validateinputs() !image Figure 4 Bubbln Awaiting Validation of Inputted Network Parameters. Auto-Generation of Prompt: After validation of inputted parameters, Bubbln composes the prompt by combining the inputted parameters with a set of well-engineered hardcoded instructions written in natural language. Responsible Code Section: prompt_generator.py ChatGPT Prompting: The auto-composed prompt is then sent to ChatGPT utilizing gpt-4 chatCompletions model with a temperature parameter of 0.2 and maximum tokens of 1500. The following functions were designed into this process stage Responsible Code Section: chatGPT_prompting.py !image Figure 5 ChatGPT prompting in progress Playbook Generation & Extraction: After ChatGPT processes the prompt from Bubbln, it provides a response which usually contains the generated playbook and explanatory notes. Bubbln then extracts the playbook from the explanatory notes by searching for “---” which usually connotes the start of playbooks and saves each generated playbook uniquely using the nomenclature RouteriPlaybook.yml. Responsible Code Section: playbook_extractor.py !image Figure 6 ChatGPT-generated playbook. Playbook Execution: Bubbln loads the saved “RouteriPlaybook.yml” playbook and dynamically generates the hosts.yml file and parses them to the python library ansiblerunner for further execution on the configured network. Bubbln generates the hosts.yml file at run time by using the pre-inputted SSH credentials in userconfig.pkl file - and decrypts them, as well as IP addresses from the sshipaddresses.txt file, as inputs Responsible Code Section: playbook_execution.py !image Figure 7 Playbook execution in progress Sample result of Executed Playbook Upon successful execution of all playbooks, a query of the routing table on router 4 indicates that router 4 could reach all the prefixes on the network. !image Figure 8 Output of 'sh ip route' executed on R1 File Management and Handling Throughout the execution process, Bubbln manages the creation, saving, and loading of various files to streamline the network automation process. user_config.pkl: This dictionary file dynamically created at run time is used to store encrypted API keys, SSH credentials and initial welcome message information. router_configuration.pkl: It is auto created by Bubbln and used to store network configuration parameters for easy loading during subsequent sessions. hosts.yml: This is a runtime autogenerated file that contains inventory of the network devices. It is auto deleted after the program runs. networkconfigurationreport.pdf: This auto-generated report by Bubbln is a documentation of all the routers configured their parameters, generated playbooks, and prompt for each execution of the Bubbln application. It is created after a successful execution of playbooks and network testing and is meant for auditing and documentation purposes. RouteriPlaybook.yml: After extraction of generated playbooks from ChatGPT’s raw response, Bubbln automatically saves a copy of the generated playbook using unique names for each playbook. !image Figure 9 File structure after successful deployment of a four-router network Providing Feedback We are glad to hear your thoughts and suggestions. Kindly do this through the discussion section of our GitHub - https://github.com/olasupo/bubbln_network-automation/discussions/1#discussion-6487475 We can also be reached on: Olasupo Okunaiya – olasupo.o@gmail.com

Vibe Coding: The Art of Ignorance
youtube
LLM Vibe Score0.29
Human Vibe Score0.38
Dylan CuriousMar 13, 2025

Vibe Coding: The Art of Ignorance

NEWSLETTER ✉️ https://dylancurious.beehiiv.com PATREON 💰 https://patreon.com/DylanCurious SOCIALS ⤵ ▶️ YouTube: https://www.youtube.com/@dylan_curious/videos 📸 Instagram: https://www.instagram.com/dylan_curious/reels/ 🐦 Twitter/X: https://x.com/dylan_curious 🧵 Threads: https://www.threads.net/@dylan_curious?hl=en 💼 LinkedIn: https://www.linkedin.com/in/dylancurious/recent-activity/all/ 👍 Facebook: https://www.facebook.com/DylanCurious/videos 📌 BlueSky: https://bsky.app/profile/dylancurious.bsky.social ☁️ TikTok: https://www.tiktok.com/@dylan_curious CHAPTERS ⤵ 00:00 - AI Social, News, & Research 02:32 - Support The Channel On Patreon! 02:56 - Vibe Coding Creates Full Blown Video Game 04:44 - Disney Rides Are Getting…Robotic 06:23 - Sony Is Creating AI-Powered Playstation Characters 07:23 - US Army Using AI To Purge DEI Training 09:17 - GPS Works…On the Moon! 10:06 - AI Simplifies Our Process To Achieve Quantum Entanglement 11:30 - Netflix’s “The Electric State” Looks Awesome 12:59 - Ex-Google CEO Issues Shocking Warning About WWIII 14:41 - Luma’s AI’s New Tool…Ray2 Flash 15:52 - New Feedback Framework For Training AI Robots 17:22 - AI Microplastic Detection Boosts Research 19:53 - Google Debuts New Gemini Text-Embedding 21:56 - OpenAI Might Be Changing Their Tune 24:18 - Julia McCoy Responds To World Chat Question 26:24 - AI Designed Church Service In Finland 27:51 - The Race For AGI…Who’s WInning? 30:35 - Catastrophe Theory and The Unseen Reality 32:55 - Like, Comment, Subscribe, & Support! SOURCES ⤵ @JuliaMcCoy https://www.youtube.com/@JuliaMcCoy https://www.youtube.com/watch?v=N4RnF-OPezI&t=1145s&ab_channel=FIVEFIRES https://youtu.be/TuK_v1J1BUo?si=UpeBx4vjutWC3Zl2 https://www.youtube.com/watch?v=QIw6ITiwgBU&ab_channel=Netflix https://www.youtube.com/watch?v=IhBuz-cnSNE&ab_channel=WesRoth https://www.nationalsecurity.ai/ https://www.youtube.com/watch?v=yUllcDzXFC8&ab_channel=LumaAI

Vibe Coding is Here - How AI is Changing How We Build Online
youtube
LLM Vibe Score0
Human Vibe Score0.28
a16zMar 13, 2025

Vibe Coding is Here - How AI is Changing How We Build Online

Vibe Coding: The Future of Software Development? (with Yoko Li & Justine Moore | a16z) What if you could build an app just by describing it? That’s the idea behind vibe coding — a new AI-driven approach that’s reshaping software development for engineers and non-technical users alike. Instead of writing detailed code, users guide an AI coding agent with simple prompts like “make this look cleaner” or “I want a button that does X.” In this episode, we sit down with Yoko Li and Justine Moore from a16z to break down the rise of vibe coding, its impact on software development, and why AI-powered text-to-web tools are taking off. We explore: How vibe coding works and why it’s gaining traction The emerging companies leading the space (Cursor, Lovable, Bolt, VZero, and more) Why engineers and total beginners are both using these tools The challenges of AI-driven development (when “vibes” go wrong!) Where this trend is heading—and what it means for the future of coding From software for one to enterprise-level applications, vibe coding is opening up new possibilities for creating on the web. Tune in to learn how it’s changing the way we build. Learn more and check out everything a16z is doing, including articles, projects, and more podcasts here – https://a16z.com/ai-web-app-builders/ Follow everyone on X: Yoko Li - https://x.com/stuffyokodraws Justine Moore - https://x.com/venturetwins Steph Smith - https://x.com/stephsmithio

Vibe Coding is the Future (?)
youtube
LLM Vibe Score0.365
Human Vibe Score0.69
Code MonkeyMar 13, 2025

Vibe Coding is the Future (?)

✅ FREE Game Dev Report Newsletter https://cmonkey.co/gamedevreportnewsletter ❤️ FREE Complete Courses https://cmonkey.co/freecourses ✅ Get my CComplete Course! https://cmonkey.co/csharpcourse 🎮 Play my Steam game! https://cmonkey.co/dinkyguardians ❤️ Watch my FREE Complete Courses https://www.youtube.com/watch?v=oZCbmB6opxY 🌍 Get my Complete Courses! ✅ https://unitycodemonkey.com/courses 👍 Learn to make awesome games step-by-step from start to finish. 🎮 Get my Steam Games https://unitycodemonkey.com/gamebundle Andrej Karpathy Twitter Post https://x.com/karpathy/status/1886192184808149383 Vibe Coding with AI in 2025 https://www.youtube.com/shorts/1_rSrkXovOk Vibe Coding is The Future https://www.youtube.com/watch?v=IACHfKmZMr8 🔴 RELATED VIDEOS 🔴 AI is creating illiterate programmers! (you?) https://www.youtube.com/watch?v=2H4ouL4bCUs AI Game Engine replacing Game Developers? https://www.youtube.com/watch?v=97C7xScuzTk Unity for NOT Game Dev? https://www.youtube.com/watch?v=yo7sFIahYQo How to SURVIVE as a Game Dev for a DECADE! (Over $1,000,000 Revenue!) https://www.youtube.com/watch?v=sfD4MMFcebE 💬 There is a new term popping up named Vibe Coding, this is apparently where you put your faith entirely in AI generated code and you never even look at it. You just prompt the AI, perhaps even with voice so you don't even use the keyboard, and you just blindly accept whatever answer the AI gives you. Is this really the future of coding? I definitely have some thoughts on this. 📝 Some Links are Affiliate links which means it costs the same to you and I get a nice commission. 🌍 Get Code Monkey on Steam! 👍 Interactive Tutorials, Complete Games and More! ✅ https://store.steampowered.com/app/1294220/ If you have any questions post them in the comments and I'll do my best to answer them. 🔔 Subscribe for more Unity Tutorials https://www.youtube.com/channel/UCFK6NCbuCIVzA6Yj1GZqCg?subconfirmation=1 See you next time! 📍 Support on Patreon https://www.patreon.com/unitycodemonkey 🎮 Grab the Game Bundle at https://unitycodemonkey.com/gameBundle.php 📝 Get the Code Monkey Utilities at https://unitycodemonkey.com/utils.php Hello and Welcome! I'm your Code Monkey and here you will learn everything about Game Development in Unity using C#. I've been developing games for several years with 8 published games on Steam and now I'm sharing my knowledge to help you on your own game development journey. I do Unity Tutorials on just about every topic, Unity Tutorials for Beginners and Unity Tutorials for Advanced users. Website: https://unitycodemonkey.com/ Twitter: https://twitter.com/UnityCodeMonkey Steam: https://store.steampowered.com/developer/EndlessLoopStudios

aion
github
LLM Vibe Score0.494
Human Vibe Score0.011340905117109681
aionnetworkFeb 28, 2025

aion

Aion Mainstream adoption of blockchains has been limited because of scalability, privacy, and interoperability challenges. Aion is a multi-tier blockchain network designed to address these challenges. Core to our hypothesis is the idea that many blockchains will be created to solve unique business challenges within unique industries. As such, the Aion network is designed to support custom blockchain architectures while providing a trustless mechanism for cross-chain interoperability. The Aion White Papers provides more details regarding our design and project roadmap. This repository contains the main (Java) kernel implementation and releases for the Aion Network. System Requirements Ubuntu 16.04 or a later version Getting Started Blockchain node concept To understand what is blockchain kernel: Node overview Developers If you're interested in building Open Applications, powered by Aion: Visit the Developer site of The Open Application Network : developer.theoan.com If you're interested in making improvements to the Java Implementation of Aion: Refer to the Build Aion kernel from source wiki for information on building this source code to a native binary or Docker image Refer to the Installation wiki for a guide on installing and configuring the kernel. The Owner's Manual wiki will include further instructions and details on working with the kernel. Please refer to the wiki pages for further documentation on mining/validating, using the Web3 API, command line options, etc. Miners/Validators If you're interested in being a validator on the Aion networks, refer to our Validator Docs Users If you're interested in interacting with dApps and using Aion, refer to our Aion Desktop Wallet Docs FAQ Where can I store my Aion? We recommend using the web-based Aion Wallet; more information can be found in “Docs”). Where can I stake my Aion? You can use the original staking interface which has support for staking pool operators, or the web-based Aion Wallet. Where can I check on a transaction on The Open Application Network? You can visit either the web-based Aion Wallet or the Aion Dashboard to view a transaction on the network. Where can I see the current network performance of The Open Application Network? You can visit the Aion Dashboard to see how the Open Application Network is performing. What should I do if the desktop wallet or the web based wallet are not functioning properly? First check in with the community on the community subreddit. If the community is not able to assist then you can submit a ticket through Github. The Open Application Network is currently providing support to help maintain the network; where can I see the funds that The Open Application Network has mined or received as a stake reward? All funds mined or rewarded for staking that the foundation receives are burned to this address: 0x0000000000000000000000000000000000000000000000000000000000000000 users can check the totals burned via the Aion Dashboard here. What is the total circulating supply of Aion? To view the current total circulating supply of Aion you can use the Aion Watch tool located here. Which networks are supported? The Mainnet network is supported. To view the dashboards for this networks use these links: Mainnet How can I export a list of my transactions? If you would like to download a copy of your transaction history you can use https://mainnet.theoan.com and search for your public address. In the bottom right of your screen is a “Download this Account” button which will allow you to select a date range and download a .csv file containing your transactions. Where can I access a copy of The OAN and Aion Brand Guidelines? The OAN and Aion Brand Guidelines can be located here they can be used by the community to create brand aligned content. My Ledger doesn’t seem to be recognized with applications in the Chrome Browser (Staking Interface or Wallet) When using your Ledger hardware wallet with Aion installed to access an account VIA the Chrome browser, users will need to enable the Aion contract on their Ledger device. This can be done by selecting: Aion > Setting > enable Contract. What happened to the Aiwa chrome extension wallet? Aiwa was owned and operated by a third-party organization called BlockX Labs, Aiwa was funded by a community grant during its lifespan. However, BlockX Labs is now reorganizing and will no longer support Aiwa. Usage of Aiwa has decreased significantly with other tools such as the web based wallet now available so the decision was made to deprecate it. I am unable to undelegate my staked Aion In order to undelegate your Aion: – You must have a sufficient Aion balance to perform the undelegation transaction (a minimum of 0.02 Aion is required for the transaction fee) – Your balance will be updated after a lock-up period of 8640 blocks (approximately 24 hours) – Ensure the amount follows this format: 999,999,999.999999999 – If you are using a ledger, please ensure that your firmware is up to date. – If you are using the desktop interface, ensure that you are using the latest version – For more information view this guide What happened to the swap process to convert ERC-20 Aion to the mainnet? As of January 31, 2022 swapping from ERC20 to Aion mainnet is no longer supported. The original Aion token swap from Ethereum to Aion was completed on December 10, 2018. However, in order to support the community members who missed the original swap deadline a manual process was available, this process has now been retired. Community Channels Newsfeed: @AionNewsfeed Info Bot: @AionTGbot Wiki: reddit.com/r/AionNetwork/Wiki Help Desk: https://helpdesk.theoan.com/ Contact To keep up to date and stay connected with current progress and development, reach out to us on the following channels: Aion Telegram Dispatch Alerts Aion on Twitter Aion Blog License Aion is released under the MIT license

Vibe Coding: Launch Your SaaS with AI (Cursor, Supabase, & Stripe)
youtube
LLM Vibe Score0.292
Human Vibe Score0.28
AI with MisbahFeb 28, 2025

Vibe Coding: Launch Your SaaS with AI (Cursor, Supabase, & Stripe)

In this video, I reveal how I built a fully functional SaaS application – complete with user authentication, course management, and Stripe payments – in just one day using AI coding tools like Cursor! You don't need to be a coding expert to achieve this. I'll walk you through the process, from using GitHub templates to leveraging Cursor's AI assistance for rapid development. Upcoming Course - Learn how to: Use AI coding tools to build complex web applications. Integrate Stripe for seamless payment processing. Implement user authentication and course management. Utilize GitHub templates for faster development. Understand the concept of "Vibe Coding" or "PromptBasedCoding". This course is perfect for aspiring entrepreneurs, developers looking to streamline their workflow, and anyone interested in the future of AI-powered development. 00:00:00 - Intro: Building a SaaS in a Day 00:00:10 - Overview of the SaaS Application (Features & Functionality) 00:00:40 - Course Preview & Payment Integration (Stripe) 00:00:50 - Introduction to VibeCoding/PromptBaseCoding & Templates 00:01:00 - Using GitHub Templates & Cursor Rules 00:01:20 - Setting Up Cursor & AI Prompting 00:01:50 - Iterative Development with Cursor 00:02:00 - Reviewing the Generated Files & "Vibe Coding" Explained 00:02:20 - New Course Announcement: Learn to Build Your Own SaaS 00:02:30 - Call to Action: Follow for Updates & Join the Builder Ecosystem Subscribe for more AI coding tutorials and SaaS development tips! #AICoding #SaaS #WebDevelopment #Cursor #Stripe #NoCode #LowCode #PromptEngineering #VibeCoding #PromptBaseCoding #SoftwareDevelopment"

Awesome-Ai-Tools
github
LLM Vibe Score0.385
Human Vibe Score0.0020930582944730723
aliammari1Feb 21, 2025

Awesome-Ai-Tools

Awesome-Ai-Tools This repo contains AI tools that will help you achieve your goals. The tools are categorized into different sections based on their functionality. Contents Awesome-Ai-Tools Contents Productivity Time Management Task Management Email Management Creativity Art Music Writing Communication Writing Personality Analysis Translation Data Science Machine Learning Data Analysis Data Visualization Natural Language Processing Text Classification Named Entity Recognition Computer Vision Image Classification Object Detection Robotics Robot Simulation Robot Control Miscellaneous Language Models Generative Models Productivity If you're looking to boost your productivity, there are a number of AI tools that can help. Time Management RescueTime - RescueTime is an AI-powered time tracking tool that helps you understand how you're spending your time on your computer. It can help you identify areas where you're wasting time and make adjustments to your workflow to be more productive. Focus@Will - Focus@Will is an AI-powered music service that helps you stay focused and productive while you work. It uses neuroscience to create music that is scientifically optimized to help you concentrate. Clockify - Clockify is an AI-powered time tracking tool that helps you track your time across different projects and tasks. It can help you identify areas where you're spending too much time and make adjustments to your workflow to be more productive. Trello - Trello is an AI-powered task management tool that helps you stay organized and on top of your to-do list. It can help you prioritize tasks, set deadlines, and even collaborate with others on projects. Motion - Motion is an AI-powered calendar and task management tool that automatically schedules your tasks and meetings for optimal productivity. Reclaim.ai - Reclaim is an intelligent calendar assistant that helps you protect your time by automatically scheduling meetings and tasks. Task Management Todoist - Todoist is an AI-powered task management tool that helps you stay organized and on top of your to-do list. It can help you prioritize tasks, set deadlines, and even suggest tasks based on your previous activity. Asana - Asana is an AI-powered task management tool that helps you stay organized and on top of your to-do list. It can help you prioritize tasks, set deadlines, and even collaborate with others on projects. Notion - Notion is an AI-powered productivity tool that can help you manage tasks, take notes, and collaborate with others on projects. It can also be used to create wikis, databases, and other types of content. Taskade - Taskade is an AI-powered productivity tool that can manage tasks and notes for individuals and teams. ClickUp - ClickUp is an AI-enhanced project management tool that helps teams organize work with automated task distributions and smart notifications. Monday.com - Monday.com uses AI to streamline workflow management and automate routine tasks. Email Management Boomerang - Boomerang is an AI-powered email management tool that helps you manage your inbox more efficiently. It can help you schedule emails to be sent later, remind you to follow up on emails, and even suggest responses to emails. SaneBox - SaneBox is an AI-powered email management tool that helps you manage your inbox more efficiently. It can help you prioritize emails, unsubscribe from unwanted emails, and even snooze emails to be dealt with later. Mailstrom - Mailstrom is an AI-powered email management tool that helps you clean up your inbox. It can help you quickly identify and delete unwanted emails, and even unsubscribe from newsletters and other types of email subscriptions. Creativity If you're looking to get more creative, there are a number of AI tools that can help. Art Artbreeder - Artbreeder is an AI-powered tool that allows you to create unique digital art by combining different images and styles. Runway ML - Runway is an AI-powered tool that allows users to edit and generate videos using natural language descriptions. Prisma - Prisma is an AI-powered tool that allows you to transform your photos into works of art using neural networks. Music AIVA - AIVA is an AI-powered music composition tool that can help you create original music for your projects. Writing monica - Monica is a chrome extension powered by ChatGPT API. It is designed to be your personal AI assistant for effortless chatting and copywriting. CopyAI - CopyAI is an AI-powered writing assistant that can help you generate high-quality marketing copy, product descriptions, and more. Grammarly - Grammarly is an AI-powered writing assistant that helps you catch grammar and spelling errors in your writing. It can also suggest improvements to your writing style to help you communicate more effectively. Jasper - Jasper is an AI writing assistant that helps create marketing copy, blog posts, and social media content. Rytr - Rytr is an AI writing tool that helps generate content in different tones and styles. Communication If you're looking to improve your communication skills, there are a number of AI tools that can help. Writing Linguix - Linguix is an AI-powered writing assistant that can help you improve your writing skills. It can catch grammar and spelling errors, suggest improvements to your writing style, and even help you avoid plagiarism. Hemingway Editor - Hemingway Editor is an AI-powered writing tool that helps you simplify your writing and make it more readable. It can help you identify complex sentences, passive voice, and other issues that can make your writing difficult to understand. Personality Analysis Crystal - Crystal is an AI-powered tool that helps you understand the personality of the people you're communicating with. It can provide insights into their communication style and suggest ways to communicate more effectively with them. IBM Watson Personality Insights - IBM Watson Personality Insights is a tool that uses natural language processing and machine learning algorithms to analyze text and provide insights into the personality traits of the author. Translation DeepL - DeepL is an AI-powered translation tool that provides high-quality translations in multiple languages. It uses neural network algorithms to provide more accurate translations than traditional translation tools. Google Translate - Google Translate is a free online translation tool that uses machine learning algorithms to provide translations in over 100 languages. Data Science If you're working with data, there are a number of AI tools that can help you analyze and make sense of it. Machine Learning DataRobot - DataRobot is an AI-powered platform that helps you build and deploy machine learning models. It can help you automate the process of building models and make predictions based on your data. TensorFlow - TensorFlow is an open-source machine learning framework developed by Google. It can help you build and train machine learning models for a variety of applications. PyTorch - PyTorch is another open-source machine learning framework that is popular among researchers and developers. It is known for its ease of use and flexibility. H2O.ai - H2O.ai is an open-source machine learning platform that allows you to build and deploy machine learning models at scale. PyTorch3d - Pytorch 3d is an open-source library for deep learning with 3d data. Auto-sklearn - Auto-sklearn is an automated machine learning toolkit that helps find the best machine learning pipeline for your dataset. Ludwig - Ludwig is a declarative machine learning framework that makes it easy to build and train models without writing code. Data Analysis Pandas - Pandas is an open-source data analysis library for Python. It can help you manipulate and analyze data in a variety of formats, including CSV, Excel, and SQL databases. RapidMiner - RapidMiner is an AI-powered data science platform that allows you to build and deploy predictive models without writing any code. Apache Spark - Apache Spark is an open-source big data processing framework that can help you analyze large datasets in a distributed computing environment. Data Visualization Tableau - Tableau is a data visualization tool that uses AI to help you explore and understand your data. It can help you identify patterns and trends in your data that might not be immediately obvious. Plotly - Plotly is an open-source data visualization library for Python. It can help you create interactive charts and graphs that can be embedded in web pages and other applications. D3.js - D3.js is a JavaScript library for data visualization that allows you to create dynamic and interactive visualizations using web standards like HTML, CSS, and SVG. Natural Language Processing If you're interested in natural language processing, there are a number of AI tools that can help you get started. Text Classification TextBlob - TextBlob is an open-source library for processing textual data in Python. It can help you perform tasks like sentiment analysis, part-of-speech tagging, and text classification. NLTK - NLTK (Natural Language Toolkit) is another open-source library for natural language processing in Python. It can help you perform tasks like tokenization, stemming, and named entity recognition. Amazon Comprehend - Amazon Comprehend is a natural language processing service that uses machine learning to analyze text and provide insights into the content and sentiment of the text. Named Entity Recognition spaCy - spaCy is an open-source library for advanced natural language processing in Python. It can help you build applications that can understand and analyze human language. One of its key features is named entity recognition, which can identify and classify entities like people, organizations, and locations. Google Cloud Natural Language API - Google Cloud Natural Language API is a natural language processing service that can analyze text and provide insights into the sentiment, entities, and syntax of the text. Computer Vision If you're interested in computer vision, there are a number of AI tools that can help you get started. Image Classification Clarifai - Clarifai is an AI-powered image recognition tool that can help you classify images based on their content. It can recognize objects, scenes, and even specific concepts like emotions and colors. Google Cloud Vision API - Google Cloud Vision API is a computer vision service that can analyze images and provide insights into the content of the images, including objects, faces, and text. Object Detection YOLO - YOLO (You Only Look Once) is an open-source object detection system that can detect objects in real-time video streams. It is known for its speed and accuracy. Amazon Rekognition - Amazon Rekognition is a computer vision service that can analyze images and videos and provide insights into the content of the media, including objects, faces, and text. Robotics If you're interested in robotics, there are a number of AI tools that can help you get started. Robot Simulation Gazebo - Gazebo is an open-source robot simulation tool that allows you to simulate robots in a virtual environment. It can help you test and debug your robot control algorithms before deploying them on a physical robot. Webots - Webots is another open-source robot simulation tool that allows you to simulate robots in a virtual environment. It supports a wide range of robots and sensors, and can be used for both research and education. Robot Control ROS - ROS (Robot Operating System) is an open-source framework for building robotics software. It can help you build and control robots using a variety of programming languages. Miscellaneous If you're looking for AI tools that don't fit into any of the above categories, here are a few to check out: Language Models GPT-3 - GPT-3 is an AI-powered language model developed by OpenAI. It can generate human-like text, answer questions, and even write code. BERT - BERT is a language model developed by Google AI. It is trained on a massive dataset of text and code, and can be used for a variety of tasks, including natural language understanding, question answering, and text classification. LLama 2 - LLama 2 models are a collection of pretrained and fine-tuned large language models developed and released by Meta AI . These models are built upon the success of LLama 1 and provide significant improvements, including a larger scale and more extensive context. Claude - Claude is an AI assistant developed by Anthropic that excels at analysis, writing, and coding tasks. PaLM 2 - PaLM 2 is Google's next-generation language model with improved multilingual, reasoning, and coding capabilities. Generative Models StyleGAN - StyleGAN is an AI-powered generative model that can create high-quality images of faces, animals, and other objects. It is known for its ability to create realistic and diverse images. Generative Pre-trained Transformer 3 (GPT-3) - GPT-3 is an AI-powered language model developed by OpenAI. It can generate human-like text, answer questions, and even write code.

kodyfire
github
LLM Vibe Score0.384
Human Vibe Score0.0032098142352129998
nooqtaFeb 2, 2025

kodyfire

Kody is a command-line tool for generating artifact files, powered by both classic and AI code generation techniques. It can be used by both technical and non-technical users to generate files across a wide range of technologies and programming languages. The code generation feature in Kody relies on OpenAI GPT, a language model that uses deep learning to generate human-like text, and ChatGPT to provide natural language processing capabilities. Table of Contents Installation Usage Getting Started Terminology Contributing License Installation Prerequisites Node.js (version 14 or later) To install kody, use npm with the following command: or You can check the documentation with Usage Options -v, --version: Output the current version -h, --help: Display help for command Commands prompt|ai [options] [prompt...]: AI powered prompt assistant to quickly generate an artifact batch [options]: Generate multiple digital artifact create [options] : Generate a new blank kody project generate|g [options] [kody] [concept]: Prompt assistant to quickly generate an artifact import|in [options] : Mass create artifacts from a source. init: Initialize a new kodyfire project install|i [kody]: Prompt user to choose to install list|ls [options] [kodyName]: List installed kodies within your current project. publish [template]: Publish the templates of the kody along with the assets.json and schema.ts files ride|↻: Prompt assistant to help build your kody.json file run [options]: Generate a digital artifact based on the selected technology run-script|rs: Run scripts search|s [keywords...]: Search kodyfire packages from npm registry watch|w [options]: Watch for file changes and run kody help [command]: Display help for command Getting Started Open the project you are willing to work on using vscode or your prefered editor. Generate artifacts using AI In case you want to exclusivly rely on AI to generate your artifacts. You don't need to install any additional kodies. Run the kody ai [prompt] command and follow the prompts. For example, to create a Laravel Controller named SampleController under API/V1 and add a comment on top saying Hello Kodyfire, run the following command You can use the experimental Speech-to-Text option to pass your prompt using your voice. The transcription relies on Whisper and requires SoX installed and available in your \$PATH. for the audio recording. For Linux For MacOS For Windows Download the binaries Generate your artifact using the classical method Search and install a kody Based on your project, search availables kodies and select the one that fits your need.. To search availables kodies by keyword runthe following command. if you don't specify a keyword all available kodies will be listed. Install your kody of choice. For example, if you want to install the react kody or Please note you can install as many kodies in the same project as you wish. Generate your artifact There are 2 methods you can generate your artifacts with: The generate command The run command Method 1: Generator mode kody generate The recommended way of using kody is using the generate command. The command will assist you creating your artifact based on the chosen concept. For example, a react component is considered a concept. In order to generate your artifacts, run the generate command. The syntax is kody g|generate [kody] [concept]. the assistant will prompt you to select the missing arguments. As an example, run the following command from your terminal: Method 2: Runner mode kody run The run command is similar to the generate command. The run requires a definition file which is simply a json file containing all the concept definitions you have created using the ride command. The generate command on the other hand creates one or more concept definition on the run and process them on one run. Every command has its use cases. Initialize kody In order to start using kody, you need to initialize your project. This will add the definition files required for kody runs. Important: Please run the command only once. The command will override existing definition files. We will disable overriding in a future version. Ride your kody In order to update your definition, use the kody ride command to assist you populate the required fields Launch a kody run Once you are satisified with your definition file, execute the run command to generate your artifacts. To run all kodies defined within your project, run the following command: Create your own kody In most cases you might need a custom kody to suit your needs Scaffold a new kody Create a basic kody using the scaffold command. Follow the prompts to setup your kody This will create a folder containing the basic structure for a kody. You can start using right away within your project. Setup your kody Install npm dependencies Build your kody Add your concepts and related templates //TODO This will build your kody and export the basic templates files. Add your kody as an NPM dependency to a test project In order to be able to use it within your test project run the following command Publish your kody Please remember that Kody is still in exploration phase and things will change frequently. Contribution is always highly requested. Prepare your kody Add the required kodyfire metadata to your package.json Publish to Github Intialize your project as a git repository and push to a public Github repo To do so, kindly follow these steps:- Intitialize a new Github repository and make it public. Open your project root folder locally from terminal and run the following commands:- Link your project to your Github repository. Publish to npm Once you are satisfied with your kody and you would to like to share it with the community. Run the following command. Note: You'll need an NPM account Share with community Congratulation publishing your first kody. Don't forget to share your kody repo link by opening an issue on Kody's github repository. Terminology Kody: Refers to the code generation command-line tool that generates digital artifacts. Artifacts: Refers to the various digital products generated by Kody based on the input provided. Note: Kody uses classical code generation techniques in addition to AI-powered code generation using OpenAI Codex and ChatGPT. Available kodies | Name | Description | | -------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------------------------------------- | | basic-kodyfire | A general purpose code generator that should handle most of the generation use cases | | typescript-kodyfire | Generate typescript related artifacts | | tsconfig-kodyfire | Generate tsconfig files for your typescript projects | | nextjs-kodyfire | Generate nextJs components and related artifacts | | react-kodyfire | Generate react components | | laravel-kodyfire | Laravel artifacts generation | | uml-kodyfire | Uml diagrams generation using plantuml | | readme-kodyfire | Readme file generation | | word-kodyfire | Generate ms word document based on a template | | pdf-kodyfire | Generate PDF document from HTML templates | | social-image-kodyfire | Generate dynamic images for social sharing based on HTML templates | | social-gif-kodyfire | Generate dynamic gif images for social sharing based on HTML templates | | linkedin-quizzes-kodyfire | Practice Linkedin skill assessement tests from your terminal | | chatgpt-kodyfire | Use chatgpt from the terminal. Allows you provide additional data from various sources (not implemented yet) and export to serveral outputs (markdown only now). | Contributing If you encounter any issues while using Kody or have suggestions for new features, feel free to open an issue or submit a pull request. Please read our contributing guidelines before making contributions. License Kody is MIT licensed.

Mastering-AI-for-Entrepreneurs-9-Free-Courses
github
LLM Vibe Score0.203
Human Vibe Score0
Softtechhub1Feb 1, 2025

Mastering-AI-for-Entrepreneurs-9-Free-Courses

Mastering-AI-for-Entrepreneurs-9-Free-Courses Introduction: The Entrepreneur's AI RevolutionArtificial Intelligence (AI) is changing the way we do business. It's not just for tech giants anymore. Small businesses and startups are using AI to work smarter, not harder. As an entrepreneur, you need to understand AI to stay ahead.Why AI is a must-have skill for entrepreneursAI is everywhere. It's in the apps we use, the products we buy, and the services we rely on. Businesses that use AI are seeing big improvements:They're making better decisions with data-driven insightsThey're automating routine tasks, freeing up time for creativityThey're personalizing customer experiences, boosting satisfaction and salesIf you're not using AI, you're falling behind. But here's the good news: you don't need to be a tech wizard to harness the power of AI.Breaking the barriers to AI learningThink AI is too complex? Think again. You don't need a computer science degree to understand and use AI in your business. Many AI tools are designed for non-technical users. They're intuitive and user-friendly.The best part? You can learn about AI for free. There are tons of high-quality courses available at no cost. These courses are designed for busy entrepreneurs like you. They cut through the jargon and focus on practical applications.What to expect from this articleWe've handpicked nine free courses that will turn you into an AI-savvy entrepreneur. Each course is unique, offering different perspectives and skills. We'll cover:What makes each course specialWhat you'll learnHow it applies to your businessWho it's best suited forReady to dive in? Let's explore these game-changing courses that will boost your AI knowledge and give your business an edge.1. Google AI Essentials: A Beginner's Guide to Practical AIWhy This Course Is EssentialGoogle AI Essentials is perfect if you're just starting out. It's designed for people who don't have a tech background. The course focuses on how AI can help you in your day-to-day work, not on complex theories.What You'll LearnThis course is all about making AI work for you. You'll discover how to:Use AI to boost your productivity. Generate ideas, create content, and manage tasks more efficiently.Streamline your workflows. Learn how AI can help with everyday tasks like drafting emails and organizing your schedule.Use AI responsibly. Understand the potential biases in AI and how to use it ethically.Key TakeawaysYou'll earn a certificate from Google. This looks great on your resume or LinkedIn profile.You'll learn how to work alongside AI tools to get better results in your business.You'll gain practical skills you can use right away to improve your work.Get StartedEnroll in Google AI Essentials2. Introduction to Generative AI: A Quick Start for EntrepreneursWhy This Course Works for Busy EntrepreneursThis course is short and sweet. In just 30 minutes, you'll get a solid grasp of generative AI. It's perfect if you're short on time but want to understand the basics.What You'll LearnThe fundamentals of generative AI: what it is, how it works, and its limitsHow generative AI differs from other types of AIReal-world applications of generative AI in businessHow It Helps Your BusinessAfter this course, you'll be able to:Make smarter decisions about using AI tools in your businessSpot opportunities where generative AI could solve problems or create valueUnderstand the potential and limitations of this technologyGet StartedEnroll in Introduction to Generative AI3. Generative AI with Large Language Models: Advanced Skills for EntrepreneursWhy This Course Stands OutThis course digs deeper into the technical side of AI. It's ideal if you have some coding experience and want to understand how AI models work under the hood.What You'll LearnYou'll gain key skills for working with Large Language Models (LLMs):How to gather and prepare data for AI modelsChoosing the right model for your needsEvaluating model performance and improving resultsYou'll also learn about:The architecture behind transformer models (the tech powering many AI tools)Techniques for fine-tuning models to your specific business needsWho Should Take This CourseThis course is best for entrepreneurs who:Have basic Python programming skillsUnderstand the fundamentals of machine learningWant to go beyond using AI tools to actually building and customizing themGet StartedEnroll in Generative AI with Large Language Models4. AI for Everyone by Andrew Ng: Simplifying AI for Business LeadersWhy It's Perfect for BeginnersAndrew Ng is a leading figure in AI education. He's known for making complex topics easy to understand. This course is designed for non-technical learners. You don't need any coding or math skills to benefit from it.What You'll LearnHow AI works at a high levelHow to spot problems in your business that AI can solveWays to assess how AI might impact your business processes and strategiesWhy Entrepreneurs Love This CourseIt explains AI concepts in plain English, without technical jargonYou can complete it in just 8 hours, fitting it into your busy scheduleIt focuses on the business value of AI, not just the technologyGet StartedStart with AI for Everyone on Coursera5. Generative AI: Introduction and ApplicationsWhy This Course Is Ideal for EntrepreneursThis course offers a broad view of generative AI applications. You'll learn about AI in text, image, audio, and more. It's packed with hands-on experience using popular AI tools.What You'll LearnThe basics and history of generative AI technologiesHow different industries are using AI, from marketing to creative projectsPractical skills through labs using tools like ChatGPT, DALL-E, and Stable DiffusionHow It Stands OutYou'll hear from real AI practitioners about their experiencesThe course teaches you how to use generative AI to innovate and improve efficiency in your businessGet StartedEnroll in Generative AI: Introduction and Applications6. Generative AI for Everyone by Andrew Ng: Unlocking ProductivityWhy This Course Is a Must-HaveThis course focuses on using generative AI tools for everyday business tasks. It's all about boosting your productivity and efficiency.What You'll LearnHands-on exercises to integrate AI tools into your daily workReal examples of how businesses are using generative AI to save time and moneyTechniques for prompt engineering to get better results from AI toolsHow It Helps EntrepreneursYou'll learn to automate repetitive tasks, freeing up time for strategic thinkingYou'll discover new ways to use AI tools in your business processesYou'll gain confidence in experimenting with AI to solve business challengesGet StartedGo deeper with DeepLearning.AI7. Generative AI for Business Leaders by LinkedIn LearningWhy This Course Focuses on Business ApplicationsThis course is tailored for leaders who want to integrate AI into their business operations. It provides practical insights for improving workflows and decision-making.What You'll LearnStrategies for using AI to optimize your business operationsHow to save time and resources with AI-powered toolsPractical methods for implementing AI in your company, regardless of sizeKey BenefitsThe course is designed for busy professionals, allowing you to learn at your own paceYou'll gain insights you can apply immediately to your businessIt covers both the potential and the limitations of AI in business settingsGet StartedLevel up on LinkedIn Learning8. AI for Beginners by Microsoft: A Structured Learning PathWhy This Course Builds a Strong AI FoundationMicrosoft's AI for Beginners is a comprehensive 12-week program. It covers core AI concepts in a structured, easy-to-follow format. The course combines theoretical knowledge with hands-on practice through quizzes and labs.What You'll LearnThe basics of AI, machine learning, and data scienceStep-by-step guidance to build a strong knowledge basePractical applications of AI in various business contextsHow to Approach This CourseDedicate 2-3 hours per week to complete the curriculumUse the structured format to gradually build your confidence in AI conceptsApply what you learn to real business scenarios as you progressGet StartedBuild foundations with Microsoft9. AI for Business Specialization by UPenn: Strategic Thinking with AIWhy This Course Is Perfect for Business LeadersThis specialization focuses on AI's transformative impact on core business functions. It covers how AI is changing marketing, finance, and operations.What You'll LearnHow to build an AI strategy tailored to your business needsWays to leverage AI to drive innovation across different departmentsTechniques for integrating AI into your business modelHow to Make the Most of This CourseTake detailed notes on how each module applies to your own business challengesUse the specialization to develop a long-term AI vision for your companyNetwork with other business leaders taking the course to share insights and experiencesGet StartedScale up with UPenn's business focusConclusion: Your Path to Becoming an AI-powered EntrepreneurWe've covered nine fantastic free courses that can transform you into an AI-savvy entrepreneur. Let's recap:Google AI Essentials: Perfect for beginners, focusing on practical AI applications.Introduction to Generative AI: A quick start to understand the basics of generative AI.Generative AI with Large Language Models: For those ready to dive into the technical side.AI for Everyone: A non-technical introduction to AI's business impact.Generative AI: Introduction and Applications: A broad look at generative AI across industries.Generative AI for Everyone: Focused on boosting productivity with AI tools.Generative AI for Business Leaders: Tailored for integrating AI into business operations.AI for Beginners: A structured path to build a strong AI foundation.AI for Business Specialization: Strategic thinking about AI in business functions.Remember, you don't need to tackle all these courses at once. Start small and build your knowledge gradually. Pick the course that aligns best with your current needs and business goals.Embracing AI is not just about staying competitive; it's about opening new doors for innovation and growth. These courses will help you see opportunities where AI can solve problems, improve efficiency, and create value for your business.The AI revolution is happening now. The sooner you start learning, the better positioned you'll be to lead in this new era. Each step you take in understanding AI is a step towards future-proofing your business.So, what are you waiting for? Choose a course, dive in, and start your journey to becoming an AI-powered entrepreneur today. The future of your business may depend on it.MORE ARTICLES FOR YOUHumanizzer Fastpass Bundle – OTO1 to OTO4: Get (Humanizzer + All OTOs) Fastpass for Massive 75% Discount Available Limited-Time OneHumanizzer Review: Build Lifelike Human AI Agents That Talk, Listen & Engage Face-To-Face!—In Your Voice, Just Like You!EasyListDetox App Review: A Windows tool with Giveaway Rights for effortlessly cleaning your email lists of duplicates, invalid, and disposable addresses. Simple, efficient, and time-savingAI Copy Kit Review: Google’s Latest AI Tech Tensorflow (Tf) Create Jaw-Dropping And Advanced Ultra HD Videos, Ultra Shorts, 4K Images, Voiceovers, and Any Other GPT 4-Powered Amazing Content In Minutes Without Any Complicated Tools!From Good to Great: 15 Books to Inspire Personal and Business TransformationFTC Affiliate Commission Disclaimer: Some links in this article may earn us a commission if you make a purchase. This doesn't affect our recommendations.

I built an AI Agent in 43 min to automate my workflows (Zero Coding)
youtube
LLM Vibe Score0.459
Human Vibe Score0.88
Greg IsenbergJan 31, 2025

I built an AI Agent in 43 min to automate my workflows (Zero Coding)

In this episode, Max Brodeur-Urbas, Gumloop's CEO, where we dive deep into how to build AI agents and how to automate any workflow. We cover various use cases, from automated sales outreach to content generation. Max shows us how Gumloop makes complex automations accessible to everyone by having user-friendly UI/UX, intuitive workflow buildouts, and easy custom integration creation. Timestamps: 00:00 - Intro 02:29 - Gumloop Workflow Overview 05:00 - Example: Lead Automation Workflow 10:23 - Templates for Workflows 12:21 - Example: YouTube to Blog Post Automation Workflow 21:03 - Gumloop Interfaces Demonstration 21:40 - Example: Media Ad Library Analyzer Automation Workflow 24:38 - Using Gumloop for SaaS Products 26:25 - Example: Analyze Daily Calendar Automation Workflow 27:47 - Output of Media Ad Library Analyzer Automation Workflow 28:43 - Cost of Running Gumloop 30:34 - Custom Node Builder Demonstration 34:18 - Gumloop Chrome Extension 37:06 - Final thoughts on business automation Gumloop Templates: https://www.gumloop.com/templates Key Points: • Demonstration of Gumloop's automation platform for building AI-powered workflows • Showcase of features including custom nodes, Chrome extension, and interface builder • Real-world examples of automated processes for sales, recruitment, and content generation • Discussion of practical business applications and cost-effectiveness of automation: Key Features Demonstrated: • Visual workflow builder • AI-powered content generation • Custom integration creation • Chrome extension functionality • Interface builder for non-technical users • Webhook integration capabilities 1) Gumloop is a visual workflow builder that lets you create powerful AI automations by connecting "nodes" - think Zapier meets ChatGPT, but WAY more powerful. Key features that stood out: 2) SUBFLOWS: Create reusable workflow components Build once, use everywhere Share with team members Perfect for complex operations Makes scaling easier 3) The YouTube Blog Post Generator is INSANE: Takes any YT video link Extracts transcript Generates TLDR summary Creates full blog post Adds video embed Posts to CMS Cost? About $1.62 per post 4) Competitor Ad Analysis automation: Scrapes competitor FB/IG ads Uses Gemini to analyze videos/images Generates strategy insights Sends beautiful email reports Runs on schedule Save 40+ hours/month 5) Custom Node Builder = game changer Create your own integrations No coding required AI helps write the code Share with your team Endless possibilities 6) Chrome Extension feature: Turn any workflow into a 1-click tool Works on any webpage Perfect for LinkedIn outreach Data enrichment Email automation 7) Why this matters: Most companies (even $1B+ ones) are still doing things manually that could be automated. The competitive advantage isn't just having AI - it's automating your workflows at scale. 8) Pricing & Getting Started: Free to try No CC required 1000 free credits with tutorial Build custom workflows Join their community Notable Quotes: "If you can list it as a list of steps, like for an intern, you would hand off a little sticky note being like, you do these 15 things in a row and that's the entire workflow, then you can 100% automate it." - Max "Being in business is a game of unfair advantages... And that means it's always about how do you save time as founders and executive teams." - Greg LCA helps Fortune 500s and fast-growing startups build their future - from Warner Music to Fortnite to Dropbox. We turn 'what if' into reality with AI, apps, and next-gen products https://latecheckout.agency/ BoringAds — ads agency that will build you profitable ad campaigns http://boringads.com/ BoringMarketing — SEO agency and tools to get your organic customers http://boringmarketing.com/ Startup Empire - a membership for builders who want to build cash-flowing businesses https://www.startupempire.co FIND ME ON SOCIAL X/Twitter: https://twitter.com/gregisenberg Instagram: https://instagram.com/gregisenberg/ LinkedIn: https://www.linkedin.com/in/gisenberg/ FIND MAX ON SOCIAL Gumloop: https://www.gumloop.com X/Twitter: https://x.com/maxbrodeururbas?lang=en LinkedIn: https://www.linkedin.com/in/max-brodeur-urbas-1a4b25172/

internet-tools-collection
github
LLM Vibe Score0.236
Human Vibe Score0.009333333333333334
bogdanmosicaJan 23, 2025

internet-tools-collection

Internet Tools Collection A collection of tools, website and AI for entrepreneurs, web designers, programmers and for everyone else. Content by category Artificial Intelligence Developers Design Entrepreneur Video Editing Stock videos Stock Photos Stock music Search Engine Optimization Blog Posts Resume Interviews No code website builder No code game builder Side Hustle Browser Extensions Other Students Artificial Intelligence Jasper - The Best AI Writing Assistant [](https://www.jasper.ai/) Create content 5x faster with artificial intelligence. Jasper is the highest quality AI copywriting tool with over 3,000 5-star reviews. Best for writing blog posts, social media content, and marketing copy. AutoDraw [](https://www.autodraw.com/) Fast drawing for everyone. AutoDraw pairs machine learning with drawings from talented artists to help you draw stuff fast. Rytr - Best AI Writer, Content Generator & Writing Assistant [](https://rytr.me/) Rytr is an AI writing assistant that helps you create high-quality content, in just a few seconds, at a fraction of the cost! Neevo - Neevo [](https://www.neevo.ai/) Kinetix Tech [](https://kinetix.tech/) Kinetix is a no-code 3D creation tool powered by Artificial Intelligence. The web-based platform leverages AI motion capture to convert a video into a 3D animation and lets you customize your avatars and environments. We make 3D animation accessible to every creator so they can create engaging stories. LALAL.AI: 100% AI-Powered Vocal and Instrumental Tracks Remover [](https://www.lalal.ai/) Split vocal and instrumental tracks quickly and accurately with LALAL.AI. Upload any audio file and receive high-quality extracted tracks in a few seconds. Copy.ai: Write better marketing copy and content with AI [](https://www.copy.ai/) Get great copy that sells. Copy.ai is an AI-powered copywriter that generates high-quality copy for your business. Get started for free, no credit card required! Marketing simplified! OpenAI [](https://openai.com/) OpenAI is an AI research and deployment company. Our mission is to ensure that artificial general intelligence benefits all of humanity. DALL·E 2 [](https://openai.com/dall-e-2/) DALL·E 2 is a new AI system that can create realistic images and art from a description in natural language. Steve.ai - World’s fastest way to create Videos [](https://www.steve.ai/) Steve.AI is an online Video making software that helps anyone to create Videos and animations in seconds. Octie.ai - Your A.I. ecommerce marketing assistant [](https://octie.ai/) Write emails, product descriptions, and more, with A.I. Created by Octane AI. hypnogram.xyz [](https://hypnogram.xyz/) Generate images from text descriptions using AI FakeYou. Deep Fake Text to Speech. [](https://fakeyou.com/) FakeYou is a text to speech wonderland where all of your dreams come true. Craiyon, formerly DALL-E mini [](https://www.craiyon.com/) Craiyon, formerly DALL-E mini, is an AI model that can draw images from any text prompt! Deck Rocks - Create Pictch Decks [](https://www.deck.rocks/) Writely | Using AI to Improve Your Writing [](https://www.writelyai.com/) Making the art of writing accessible to all Writesonic AI Writer - Best AI Writing Assistant [](https://writesonic.com/) Writesonic is an AI writer that's been trained on top-performing SEO content, high-performing ads, and converting sales copy to help you supercharge your writing and marketing efforts. Smart Copy - AI Copywriting Assistant | Unbounce [](https://unbounce.com/product/smart-copy/) Generate creative AI copy on-the-spot across your favourite tools Synthesia | #1 AI Video Generation Platform [](https://www.synthesia.io/) Create AI videos by simply typing in text. Easy to use, cheap and scalable. Make engaging videos with human presenters — directly from your browser. Free demo. NVIDIA Canvas: Turn Simple Brushstrokes into Realistic Images [](https://www.nvidia.com/en-us/studio/canvas/) Create backgrounds quickly, or speed up your concept exploration so you can spend more time visualizing ideas with the help of NVIDIA Canvas. Hotpot.ai - Hotpot.ai [](https://hotpot.ai/) Hotpot.ai makes graphic design and image editing easy. AI tools allow experts and non-designers to automate tedious tasks while attractive, easy-to-edit templates allow anyone to create device mockups, social media posts, marketing images, app icons, and other work graphics. Klaviyo: Marketing Automation Platform for Email & SMS [](https://www.klaviyo.com/) Klaviyo, an ecommerce marketing automation platform for email marketing and sms syncs your tech stack with your website store to scale your business. Search listening tool for market, customer & content research - AnswerThePublic [](https://answerthepublic.com/) Use our free tool to get instant, raw search insights, direct from the minds of your customers. Upgrade to a paid plan to monitor for new ways that people talk & ask questions about your brand, product or topic. Topic Mojo [](https://topicmojo.com/) Discover unique & newest queries around any topic and find what your customers are searching for. Pulling data from 50+ sources to enhance your topic research. AI Image Enlarger | Enlarge Image Without Losing Quality! [](https://imglarger.com/) AI Image Enlarger is a FREE online image enlarger that could upscale and enhance small images automatically. Make jpg/png pictures big without losing quality. Midjourney [](https://www.midjourney.com/app/) Kaedim - AI for turning 2D images to 3D models [](https://www.kaedim3d.com/webapp) AI for turning 2D images, sketches and photos to 3D models in seconds. Overdub: Ultra realistic text to speech voice cloning - Descript [](https://www.descript.com/overdub) Create a text to speech model of your voice. Try a live demo. Getting Started [](https://magenta.tensorflow.org/get-started) Resources to learn about Magenta Photosonic AI Art Generator | Create Unique Images with AI [](https://photosonic.writesonic.com/) Transform your imagination into stunning digital art with Photosonic - the AI art generator. With its creative suggestions, this Writesonic's AI image generator can help unleash your inner artist and share your creations with the world. Image Computer [](https://image.computer/) Most downloaded Instagram Captions App (+more creator tools) [](https://captionplus.app/) Join 3 Million+ Instagram Creators who use CaptionPlus to find Instagram Captions, Hashtags, Feed Planning, Reel Ideas, IG Story Design and more. Writecream - Best AI Writer & Content Generator - Writecream [](https://www.writecream.com/) Sentence Rewriter is a free tool to reword a sentence, paragraph and even entire essays in a short amount of time. Hypotenuse AI: AI Writing Assistant and Text Generator [](https://www.hypotenuse.ai/) Turn a few keywords into original, insightful articles, product descriptions and social media copy with AI copywriting—all in just minutes. Try it free today. Text to Speach Listnr: Generate realistic Text to Speech voiceovers in seconds [](https://www.listnr.tech/) AI Voiceover Generator with over 600+ voiceovers in 80+ languages, go from Text to Voice in seconds. Get started for Free! Free Text to Speech: Online, App, Software, Commercial license with Natural Sounding Voices. [](https://www.naturalreaders.com/) Free text to speech online app with natural voices, convert text to audio and mp3, for personal and commercial use Developers OverAPI.com | Collecting all the cheat sheets [](https://overapi.com/) OverAPI.com is a site collecting all the cheatsheets,all! Search Engine For Devs [](https://you.com/) Spline - Design tool for 3D web browser experiences [](https://spline.design/) Create web-based 3D browser experiences Image to HTML CSS converter. Convert image to HTML CSS with AI: Fronty [](https://fronty.com/) Fronty - Image to HTML CSS code converter. Convert image to HTML powered by AI. Sketchfab - The best 3D viewer on the web [](https://sketchfab.com/) With a community of over one million creators, we are the world’s largest platform to publish, share, and discover 3D content on web, mobile, AR, and VR. Railway [](https://railway.app/) Railway is an infrastructure platform where you can provision infrastructure, develop with that infrastructure locally, and then deploy to the cloud. JSON Crack - Crack your data into pieces [](https://jsoncrack.com/) Simple visualization tool for your JSON data. No forced structure, paste your JSON and view it instantly. Locofy.ai - ship your products 3-4x faster — with low code [](https://www.locofy.ai/) Turn your designs into production-ready frontend code for mobile apps and web. Ship products 3-4x faster with your existing design tools, tech stacks & workflows. Oh Shit, Git!?! [](https://ohshitgit.com/) Carbon | Create and share beautiful images of your source code [](https://carbon.now.sh/) Carbon is the easiest way to create and share beautiful images of your source code. GPRM : GitHub Profile ReadMe Maker [](https://gprm.itsvg.in/) Best Profile Generator, Create your perfect GitHub Profile ReadMe in the best possible way. Lots of features and tools included, all for free ! HubSpot | Software, Tools, and Resources to Help Your Business Grow Better [](https://www.hubspot.com/) HubSpot’s integrated CRM platform contains the marketing, sales, service, operations, and website-building software you need to grow your business. QuickRef.ME - Quick Reference Cheat Sheet [](https://quickref.me/) Share quick reference and cheat sheet for developers massCode | A free and open source code snippets manager for developers [](https://masscode.io/) Code snippets manager for developers, developed using web technologies. Snyk | Developer security | Develop fast. Stay secure. [](https://snyk.io/) Snyk helps software-driven businesses develop fast and stay secure. Continuously find and fix vulnerabilities for npm, Maven, NuGet, RubyGems, PyPI and more. Developer Roadmaps [](https://roadmap.sh/) Community driven roadmaps, articles, guides, quizzes, tips and resources for developers to learn from, identify their career paths, know what they don't know, find out the knowledge gaps, learn and improve. CSS Generators Get Waves – Create SVG waves for your next design [](https://getwaves.io/) A free SVG wave generator to make unique SVG waves for your next web design. Choose a curve, adjust complexity, randomize! Box Shadows [](https://box-shadow.dev/) Tridiv | CSS 3D Editor [](http://tridiv.com/) Tridiv is a web-based editor for creating 3D shapes in CSS Glassmorphism CSS Generator - Glass UI [](https://ui.glass/generator/) Generate CSS and HTML components using the glassmorphism design specifications based on the Glass UI library. Blobmaker - Make organic SVG shapes for your next design [](https://www.blobmaker.app/) Make organic SVG shapes for your next design. Modify the complexity, contrast, and color, to generate unique SVG blobs every time. Keyframes.app [](https://keyframes.app/) cssFilters.co - Custom and Instagram like photo filters for CSS [](https://www.cssfilters.co/) Visual playground for generating CSS for custom and Instagram like photo filters. Experiment with your own uploaded photo or select one from the Unsplash collection. CSS Animations Animista - CSS Animations on Demand [](https://animista.net/) Animista is a CSS animation library and a place where you can play with a collection of ready-made CSS animations and download only those you will use. Build Internal apps Superblocks | Save 100s of developer hours on internal tools [](https://www.superblocks.com/) Superblocks is the fast, easy and secure way for developers to build custom internal tools fast. Connect your databases & APIs. Drag and drop UI components. Extend with Python or Javascript. Deploy in 1-click. Secure and Monitor using your favorite tools Budibase | Build internal tools in minutes, the easy way [](https://budibase.com/) Budibase is a modern, open source low-code platform for building modern internal applications in minutes. Retool | Build internal tools, remarkably fast. [](https://retool.com/) Retool is the fast way to build internal tools. Drag-and-drop our building blocks and connect them to your databases and APIs to build your own tools, instantly. Connects with Postgres, REST APIs, GraphQL, Firebase, Google Sheets, and more. Built by developers, for developers. Trusted by startups and Fortune 500s. Sign up for free. GitHub Repositories GitHub - vasanthk/how-web-works: What happens behind the scenes when we type www.google.com in a browser? [](https://github.com/vasanthk/how-web-works) What happens behind the scenes when we type www.google.com in a browser? - GitHub - vasanthk/how-web-works: What happens behind the scenes when we type www.google.com in a browser? GitHub - kamranahmedse/developer-roadmap: Interactive roadmaps, guides and other educational content to help developers grow in their careers. [](https://github.com/kamranahmedse/developer-roadmap) Interactive roadmaps, guides and other educational content to help developers grow in their careers. - GitHub - kamranahmedse/developer-roadmap: Interactive roadmaps, guides and other educational content to help developers grow in their careers. GitHub - apptension/developer-handbook: An opinionated guide on how to become a professional Web/Mobile App Developer. [](https://github.com/apptension/developer-handbook) An opinionated guide on how to become a professional Web/Mobile App Developer. - GitHub - apptension/developer-handbook: An opinionated guide on how to become a professional Web/Mobile App Developer. ProfileMe.dev | Create an amazing GitHub profile in minutes [](https://www.profileme.dev/) ProfileMe.dev | Create an amazing GitHub profile in minutes GitHub - Kristories/awesome-guidelines: A curated list of high quality coding style conventions and standards. [](https://github.com/Kristories/awesome-guidelines) A curated list of high quality coding style conventions and standards. - GitHub - Kristories/awesome-guidelines: A curated list of high quality coding style conventions and standards. GitHub - tiimgreen/github-cheat-sheet: A list of cool features of Git and GitHub. [](https://github.com/tiimgreen/github-cheat-sheet) A list of cool features of Git and GitHub. Contribute to tiimgreen/github-cheat-sheet development by creating an account on GitHub. GitHub - andreasbm/web-skills: A visual overview of useful skills to learn as a web developer [](https://github.com/andreasbm/web-skills) A visual overview of useful skills to learn as a web developer - GitHub - andreasbm/web-skills: A visual overview of useful skills to learn as a web developer GitHub - Ebazhanov/linkedin-skill-assessments-quizzes: Full reference of LinkedIn answers 2022 for skill assessments (aws-lambda, rest-api, javascript, react, git, html, jquery, mongodb, java, Go, python, machine-learning, power-point) linkedin excel test lösungen, linkedin machine learning test LinkedIn test questions and answers [](https://github.com/Ebazhanov/linkedin-skill-assessments-quizzes) Full reference of LinkedIn answers 2022 for skill assessments (aws-lambda, rest-api, javascript, react, git, html, jquery, mongodb, java, Go, python, machine-learning, power-point) linkedin excel test lösungen, linkedin machine learning test LinkedIn test questions and answers - GitHub - Ebazhanov/linkedin-skill-assessments-quizzes: Full reference of LinkedIn answers 2022 for skill assessments (aws-lambda, rest-api, javascript, react, git, html, jquery, mongodb, java, Go, python, machine-learning, power-point) linkedin excel test lösungen, linkedin machine learning test LinkedIn test questions and answers Blockchain/Crypto Dashboards [](https://dune.com/) Blockchain ecosystem analytics by and for the community. Explore and share data from Ethereum, xDai, Polygon, Optimism, BSC and Solana for free. Introduction - The Anchor Book v0.24.0 [](https://book.anchor-lang.com/introduction/introduction.html) Crypto & Fiat Exchange Super App | Trade, Save & Spend | hi [](https://hi.com/) Buy, Trade, Send and Earn Crypto & Fiat. Deposit Bitcoin, ETH, USDT and other cryptos and start earning. Get the hi Debit Card and Multi-Currency IBAN Account. Moralis Web3 - Enterprise-Grade Web3 APIs [](https://moralis.io/) Bridge the development gap between Web2 and Web3 with Moralis’ powerful Web3 APIs. Mirror [](https://mirror.xyz/) Built on web3 for web3, Mirror’s robust publishing platform pushes the boundaries of writing online—whether it’s the next big white paper or a weekly community update. Makerdao [](https://blog.makerdao.com/) Sholi — software for Investors & Traders / Sholi MetriX [](https://sholi.io/) Sholi — software for Investors & Traders / Sholi MetriX Stock Trading Quiver Quantitative [](https://www.quiverquant.com/) Quiver Quantitative Chart Prime - The only tool you'll need for trading assets across all markets [](https://chartprime.com/) ChartPrime offers a toolkit that will take your trading game to the next level. Visit our site for a full rundown of features and helpful tutorials. Learning Hacker Rank [](https://www.hackerrank.com/) Coderbyte | Code Screening, Challenges, & Interview Prep [](https://coderbyte.com/) Improve your coding skills with our library of 300+ challenges and prepare for coding interviews with content from leading technology companies. Competitive Programming | Participate & Learn | CodeChef [](https://www.codechef.com/) Learn competitive programming with the help of CodeChef's coding competitions. Take part in these online coding contests to level up your skills Learn to Code - for Free | Codecademy [](https://www.codecademy.com/) Learn the technical skills to get the job you want. Join over 50 million people choosing Codecademy to start a new career (or advance in their current one). Free Code Camp [](https://www.freecodecamp.org/) Learn to Code — For Free Sololearn: Learn to Code [](https://www.sololearn.com/home) Join Now to learn the basics or advance your existing skills Mimo: The coding app you need to learn to code! Python, HTML, JavaScript [](https://getmimo.com/) Join more than 17 million learners worldwide. Learn to code for free. Learn Python, JavaScript, CSS, SQL, HTML, and more with our free code learning app. Free for developers [](https://free-for.dev/#/) Your Career in Web Development Starts Here | The Odin Project [](https://www.theodinproject.com/) The Odin Project empowers aspiring web developers to learn together for free Code Learning Games CheckiO - coding games and programming challenges for beginner and advanced [](https://checkio.org/) CheckiO - coding websites and programming games. Improve your coding skills by solving coding challenges and exercises online with your friends in a fun way. Exchanges experience with other users online through fun coding activities Coding for Kids | Game-Based Programming | CodeMonkey [](https://www.codemonkey.com/) CodeMonkey is a leading coding for kids program. Through its award-winning courses, millions of students learn how to code in real programming languages. Coding Games and Programming Challenges to Code Better [](https://www.codingame.com/) CodinGame is a challenge-based training platform for programmers where you can play with the hottest programming topics. Solve games, code AI bots, learn from your peers, have fun. Learn VIM while playing a game - VIM Adventures [](https://vim-adventures.com/) VIM Adventures is an online game based on VIM's keyboard shortcuts. It's the "Zelda meets text editing" game. So come have some fun and learn some VIM! CodeCombat - Coding games to learn Python and JavaScript [](https://codecombat.com/) Learn typed code through a programming game. Learn Python, JavaScript, and HTML as you solve puzzles and learn to make your own coding games and websites. Design Useberry - Codeless prototype analytics [](https://www.useberry.com/) User testing feedback & rich insights in minutes, not months! Figma: the collaborative interface design tool. [](https://www.figma.com/) Build better products as a team. Design, prototype, and gather feedback all in one place with Figma. Dribbble - Discover the World’s Top Designers & Creative Professionals [](https://dribbble.com/) Find Top Designers & Creative Professionals on Dribbble. We are where designers gain inspiration, feedback, community, and jobs. Your best resource to discover and connect with designers worldwide. Photopea | Online Photo Editor [](https://www.photopea.com/) Photopea Online Photo Editor lets you edit photos, apply effects, filters, add text, crop or resize pictures. Do Online Photo Editing in your browser for free! Toools.design – An archive of 1000+ Design Resources [](https://www.toools.design/) A growing archive of over a thousand design resources, weekly updated for the community. Discover highly useful design tools you never thought existed. All Online Tools in One Box | 10015 Tools [](https://10015.io/) All online tools you need in one box for free. Build anything online with “all-in-one toolbox”. All tools are easy-to-use, blazing fast & free. Phase - Digital Design Reinvented| Phase [](https://phase.com/) Design and prototype websites and apps visually and intuitively, in a new powerful product reworked for the digital age. Animated Backgrounds [](https://animatedbackgrounds.me/) A Collection of 30+ animated backgrounds for websites and blogs.With Animated Backgrounds, set a simple, elegant background animations on your websites and blogs. Trianglify.io · Low Poly Pattern Generator [](https://trianglify.io/) Trianglify.io is a tool for generating low poly triangle patterns that can be used as wallpapers and website assets. Cool Backgrounds [](https://coolbackgrounds.io/) Explore a beautifully curated selection of cool backgrounds that you can add to blogs, websites, or as desktop and phone wallpapers. SVG Repo - Free SVG Vectors and Icons [](https://www.svgrepo.com/) Free Vectors and Icons in SVG format. ✅ Download free mono or multi color vectors for commercial use. Search in 300.000+ Free SVG Vectors and Icons. Microcopy - Short copy text for your website. [](https://www.microcopy.me/) Search micro UX copy text: slogans, headlines, notifications, CTA, error messages, email, account preferences, and much more. 3D icons and icon paks - Free3Dicon [](https://free3dicon.com/) All 3D icons you need in one place. This is a collection of free, beautiful, trending 3D icons, that you can use in any project. Love 3D Icon [](https://free3dicons.com/) Downloads free 3D icons GIMP - GNU Image Manipulation Program [](https://www.gimp.org/) GIMP - The GNU Image Manipulation Program: The Free and Open Source Image Editor blender.org - Home of the Blender project - Free and Open 3D Creation Software [](https://www.blender.org/) The Freedom to Create 3D Design Software | 3D Modeling on the Web | SketchUp [](https://www.sketchup.com/) SketchUp is a premier 3D design software that truly makes 3D modeling for everyone, with a simple to learn yet robust toolset that empowers you to create whatever you can imagine. Free Logo Maker - Create a Logo in Seconds - Shopify [](https://www.shopify.com/tools/logo-maker) Free logo maker tool to generate custom design logos in seconds. This logo creator is built for entrepreneurs on the go with hundreds of templates, free vectors, fonts and icons to design your own logo. The easiest way to create business logos online. All your design tools in one place | Renderforest [](https://www.renderforest.com/) Time to get your brand noticed. Create professional videos, logos, mockups, websites, and graphics — all in one place. Get started now! Prompt Hero [](https://prompthero.com/) Type Scale - A Visual Calculator [](https://type-scale.com/) Preview and choose the right type scale for your project. Experiment with font size, scale and different webfonts. DreamFusion: Text-to-3D using 2D Diffusion [](https://dreamfusion3d.github.io/) DreamFusion: Text-to-3D using 2D Diffusion, 2022. The branding style guidelines documents archive [](https://brandingstyleguides.com/) Welcome to the brand design manual documents directory. Search over our worldwide style assets handpicked collection, access to PDF documents for inspiration. Super designer | Create beautiful designs with a few clicks [](https://superdesigner.co/) Create beautiful designs with a few clicks. Simple design tools to generate unique patterns, backgrounds, 3D shapes, colors & images for social media, websites and more Readymag—a design tool to create websites without coding [](https://readymag.com/) Meet the most elegant, simple and powerful web-tool for designing websites, presentations, portfolios and all kinds of digital publications. ffflux: Online SVG Fluid Gradient Background Generator | fffuel [](https://fffuel.co/ffflux/) SVG generator to make fluid gradient backgrounds that feel organic and motion-like. Perfect to add a feeling of motion and fluidity to your web designs. Generate unique SVG design assets | Haikei [](https://haikei.app/) A web-based design tool to generate unique SVG design assets for websites, social media, blog posts, desktop and mobile wallpapers, posters, and more! Our generators let you discover, customize, randomize, and export generative SVG design assets ready to use with your favorite design tools. UI/UX - Inspirational Free Website Builder Software | 10,000+ Free Templates [](https://nicepage.com/) Nicepage is your website builder software breaking limitations common for website builders with revolutionary freehand positioning. 7000+ Free Templates. Easy Drag-n-Drop. No coding. Mobile-friendly. Clean HTML. Super designer | Create beautiful designs with a few clicks [](https://superdesigner.co/) Create beautiful designs with a few clicks. Simple design tools to generate unique patterns, backgrounds, 3D shapes, colors & images for social media, websites and more Pika – Create beautiful mockups from screenshots [](https://pika.style/) Quickly create beautiful website and device mockup from screenshot. Pika lets you capture website screenshots form URL, add device and browser frames, customize background and more LiveTerm [](https://liveterm.vercel.app/) Minimal Gallery – Web design inspiration [](https://minimal.gallery/) For the love of beautiful, clean and functional websites. Awwwards - Website Awards - Best Web Design Trends [](https://www.awwwards.com/) Awwwards are the Website Awards that recognize and promote the talent and effort of the best developers, designers and web agencies in the world. Design Systems For Figma [](https://www.designsystemsforfigma.com/) A collection of Design Systems for Figma from all over the globe. Superside: Design At Scale For Ambitious Brands [](https://www.superside.com/) We are an always-on design company. Get a team of dedicated designers, speedy turnarounds, magical creative collaboration tech and the top 1% of global talent. UXArchive - Made by Waldo [](https://uxarchive.com/) UXArchive the world's largest library of mobile user flows. Be inspired to design the best user experiences. Search by Muzli [](https://search.muz.li/) Search, discover, test and create beautiful color palettes for your projects Siteinspire | Web Design Inspiration [](https://www.siteinspire.com/) SAVEE [](https://savee.it/) The best way to save and share inspiration. A little corner of the internet to find good landing page copywriting examples [](https://greatlandingpagecopy.com/) A little corner of the internet to find great landing page copywriting examples. The Best Landing Page Examples For Design Inspiration - SaaS Landing Page [](https://saaslandingpage.com/) SaaS Landing Page showcases the best landing page examples created by top-class SaaS companies. Get ideas and inspirations for your next design project. Websites Free templates Premium Bootstrap Themes and Templates: Download @ Creative Tim [](https://www.creative-tim.com/) UI Kits, Templates and Dashboards built on top of Bootstrap, Vue.js, React, Angular, Node.js and Laravel. Join over 2,014,387+ creatives to access all our products! Free Bootstrap Themes, Templates, Snippets, and Guides - Start Bootstrap [](https://startbootstrap.com/) Start Bootstrap develops free to download, open source Bootstrap 5 themes, templates, and snippets and creates guides and tutorials to help you learn more about designing and developing with Bootstrap. Free Website Templates [](https://freewebsitetemplates.com/) Get your free website templates here and use them on your website without needing to link back to us. One Page Love - One Page Website Inspiration and Templates [](https://onepagelove.com/) One Page Love is a One Page website design gallery showcasing the best Single Page websites, templates and resources. Free CSS | 3400 Free Website Templates, CSS Templates and Open Source Templates [](https://www.free-css.com/) Free CSS has 3400 free website templates, all templates are free CSS templates, open source templates or creative commons templates. Free Bootstrap Themes and Website Templates | BootstrapMade [](https://bootstrapmade.com/) At BootstrapMade, we create beautiful website templates and bootstrap themes using Bootstrap, the most popular HTML, CSS and JavaScript framework. Free and Premium Bootstrap Themes, Templates by Themesberg [](https://themesberg.com/) Free and Premium Bootstrap themes, templates, admin dashboards and UI kits used by over 38820 web developers and software companies HTML, Vue.js and React templates for startup landing pages - Cruip [](https://cruip.com/) Cruip is a gallery of premium and free HTML, Vue.js and React templates for startups and SaaS. Free Website Templates Download | WordPress Themes - W3Layouts [](https://w3layouts.com/) Want to download free website templates? W3Layouts WordPress themes and website templates are built with responsive web design techniques. Download now! Free HTML Landing Page Templates and UI Kits | UIdeck [](https://uideck.com/) Free HTML Landing Page Templates, Bootstrap Themes, React Templates, HTML Templates, Tailwind Templates, and UI Kits. Create Online Graphics Snappa - Quick & Easy Graphic Design Software [](https://snappa.com/) Snappa makes it easy to create any type of online graphic. Create & publish images for social media, blogs, ads, and more! Canva [](https://www.canva.com/) Polotno Studio - Make graphical designs [](https://studio.polotno.com) Free online design editor. Create images for social media, youtube previews, facebook covers Free Logo Maker: Design Custom Logos | Adobe Express [](https://www.adobe.com/express/create/logo) The Adobe Express logo maker is instant, intuitive, and intelligent. Use it to generate a wide range of possibilities for your own logo. Photo Editor: Fotor – Free Online Photo Editing & Image Editor [](https://www.fotor.com/) Fotor's online photo editor helps you edit photos with free online photo editing tools. Crop photos, resize images, and add effects/filters, text, and graphics in just a few clicks. Photoshop online has never been easier with Fotor's free online photo editor. VistaCreate – Free Graphic Design Software with 70,000+ Free Templates [](https://create.vista.com/) Looking for free graphic design software? Easily create professional designs with VistaCreate, a free design tool with powerful features and 50K+ ready-made templates Draw Freely | Inkscape [](https://inkscape.org/) Inkscape is professional quality vector graphics software which runs on Linux, Mac OS X and Windows desktop computers. Visual & Video Maker Trusted By 11 Million Users - Piktochart [](https://piktochart.com/) With Piktochart, you can create professional-looking infographics, flyers, posters, charts, videos, and more. No design experience needed. Start for free. The Web's Favorite Online Graphic Design Tool | Stencil [](https://getstencil.com/) Stencil is a fantastically easy-to-use online graphic design tool and image editor built for business owners, social media marketers, and bloggers. Pablo by Buffer - Design engaging images for your social media posts in under 30 seconds [](https://pablo.buffer.com/) Buffer makes it super easy to share any page you're reading. Keep your Buffer topped up and we automagically share them for you through the day. Free Online Graphic Design Software | Create stunning designs in seconds. [](https://desygner.com/) Easy drag and drop graphic design tool for anyone to use with 1000's of ready made templates. Create & print professional business cards, flyers, social posts and more. Color Pallet Color Palettes for Designers and Artists - Color Hunt [](https://colorhunt.co/) Discover the newest hand-picked color palettes of Color Hunt. Get color inspiration for your design and art projects. Coolors - The super fast color palettes generator! [](https://coolors.co/) Generate or browse beautiful color combinations for your designs. Get color palette inspiration from nature - colorpalettes.earth [](https://colorpalettes.earth/) Color palettes inspired by beautiful nature photos Color Palette Generator - Create Beautiful Color Schemes [](https://colors.muz.li/) Search, discover, test and create beautiful color palettes for your projects A Most Useful Color Picker | 0to255 [](https://0to255.com/) Find lighter and darker colors based on any color. Discover why over two million people have used 0to255 to choose colors for their website, logo, room interior, and print design projects. Colour Contrast Checker [](https://colourcontrast.cc/) Check the contrast between different colour combinations against WCAG standards Fonts Google Fonts [](https://fonts.google.com/) Making the web more beautiful, fast, and open through great typography Fonts In Use – Type at work in the real world. [](https://fontsinuse.com/) A searchable archive of typographic design, indexed by typeface, format, and topic. Wordmark - Helps you choose fonts! [](https://wordmark.it/) Wordmark helps you choose fonts by quickly displaying your text with your fonts. OH no Type Company [](https://ohnotype.co/) OH no Type Co. Retail and custom typefaces. Life’s a thrill, fonts are chill! Illustrations Illustrations | unDraw [](https://undraw.co/illustrations) The design project with open-source illustrations for any idea you can imagine and create. Create beautiful websites, products and applications with your color, for free. Design Junction [](https://designjunction.xyz/) Design Junction is a one-stop resource library for Designers and Creatives with curated list of best resources handpicked from around the web Humaaans: Mix-&-Match illustration library [](https://www.humaaans.com/) Mix-&-match illustrations of people with a design library for InVIsion Studio and Sketch. Stubborn - Free Illustrations Generator [](https://stubborn.fun/) Free illustrations generator for Figma and Sketch. Get the opportunity to design your characters using symbols and styles. Open Peeps, Hand-Drawn Illustration Library [](https://www.openpeeps.com/) Open Peeps is a hand-drawn illustration library to create scenes of people. You can use them in product illustration, marketing, comics, product states, user flows, personas, storyboarding, quinceañera invitations, or whatever you want! ⠀ Reshot | Free icons & illustrations [](https://www.reshot.com/) Design freely with instant downloads of curated SVG icons and vector illustrations. All free with commercial licensing. No attribution required. Blush: Illustrations for everyone [](https://blush.design/) Blush makes it easy to add free illustrations to your designs. Play with fully customizable graphics made by artists across the globe. Mockups Angle 4 - 5000+ Device Mockups for Figma, Sketch and XD [](https://angle.sh/) Vector mockups for iPhone, iPad, Android and Mac devices, including the new iPhone 13, Pro, Pro Max and Mini. Perfect for presenting your apps. Huge library of components, compositions, wallpapers and plugins made for Figma, Sketch and XD. Make Mockups, Logos, Videos and Designs in Seconds [](https://placeit.net/) Get unlimited downloads on all our 100K templates! You can make a logo, video, mockup, flyer, business card and social media image in seconds right from your browser. Free and premium tools for graphic designers | Lstore Graphics [](https://www.ls.graphics/) Free and premium mockups, UI/UX tools, scene creators for busy designers Logo Design & Brand Identity Platform for Entrepreneurs | Looka [](https://looka.com/) Logojoy is now Looka! Design a Logo, make a website, and create a Brand Identity you’ll love with the power of Artificial Intelligence. 100% free to use. Create stunning product mockups easily and online - Smartmockups [](https://smartmockups.com/) Smartmockups enables you to create stunning high-resolution mockups right inside your browser within one interface across multiple devices. Previewed - Free mockup generator for your app [](https://previewed.app/) Join Previewed to create stunning 3D image shots and animations for your app. Choose from hundreds of ready made mockups, or create your own. Free Design Software - Graphic Online Maker - Glorify [](https://www.glorify.com/) Create professional and high converting social media posts, ads, infographics, presentations, and more with Glorify, a free design software & graphic maker. Other BuiltWith Technology Lookup [](https://builtwith.com/) Web technology information profiler tool. Find out what a website is built with. Compress JPEG Images Online [](https://compressjpeg.com/) Compress JPEG images and photos for displaying on web pages, sharing on social networks or sending by email. PhotoRoom - Remove Background and Create Product Pictures [](https://www.photoroom.com/) Create product and portrait pictures using only your phone. Remove background, change background and showcase products. Magic Eraser - Remove unwanted things from images in seconds [](https://www.magiceraser.io/) Magic Eraser - Use AI to remove unwanted things from images in seconds. Upload an image, mark the bit you need removed, download the fixed up image. Compressor.io - optimize and compress JPEG photos and PNG images [](https://compressor.io/) Optimize and compress JPEG, PNG, SVG, GIF and WEBP images online. Compress, resize and rename your photos for free. Remove Video Background – Unscreen [](https://www.unscreen.com/) Remove the background of any video - 100% automatically, online & free! Goodbye Greenscreen. Hello Unscreen. Noun Project: Free Icons & Stock Photos for Everything [](https://thenounproject.com/) Noun Project features the most diverse collection of icons and stock photos ever. Download SVG and PNG. Browse over 5 million art-quality icons and photos. Design Principles [](https://principles.design/) An Open Source collection of Design Principles and methods Shapefest™ - A massive library of free 3D shapes [](https://www.shapefest.com/) A massive free library of beautifully rendered 3D shapes. 160,000+ high resolution PNG images in one cohesive library. Learning UX Degreeless.design - Everything I Learned in Design School [](https://degreeless.design/) This is a list of everything I've found useful in my journey of learning design, and an ongoing list of things I think you should read. For budding UX, UI, Interaction, or whatever other title designers. UX Tools | Practical UX skills and tools [](https://uxtools.co/) Lessons and resources from two full-time product designers. Built For Mars [](https://builtformars.com/) On a mission to help the world build better user experiences by demystifying UX. Thousands of hours of research packed into UX case studies. Case Study Club – Curated UX Case Study Gallery [](https://www.casestudy.club/) Case Study Club is the biggest curated gallery of the best UI/UX design case studies. Get inspired by industry-leading designers, openly sharing their UX process. The Guide to Design [](https://start.uxdesign.cc/) A self-guided class to help you get started in UX and answer key questions about craft, design, and career Uxcel - Where design careers are built [](https://app.uxcel.com/explore) Available on any device anywhere in the world, Uxcel is the best way to improve and learn UX design online in just 5 minutes per day. UI & UX Design Tips by Jim Raptis. [](https://www.uidesign.tips/) Learn UI & UX Design with practical byte-sized tips and in-depth articles from Jim Raptis. Entrepreneur Instant Username Search [](https://instantusername.com/#/) Instant Username Search checks out if your username is available on more than 100 social media sites. Results appear instantly as you type. Flourish | Data Visualization & Storytelling [](https://flourish.studio/) Beautiful, easy data visualization and storytelling PiPiADS - #1 TikTok Ads Spy Tool [](https://www.pipiads.com/) PiPiADS is the best tiktok ads spy tool .We provide tiktok advertising,advertising on tiktok,tiktok ads examples,tiktok ads library,tiktok ads best practices,so you can understand the tiktok ads cost and master the tiktok ads 2021 and tiktok ads manager. Minea - The best adspy for product search in ecommerce and dropshipping [](https://en.minea.com/) Minea is the ultimate e-commerce product search tool. Minea tracks all ads on all networks. Facebook Ads, influencer product placements, Snapspy, all networks are tracked. Stop paying adspy 149€ for one network and discover Minea. AdSpy [](https://adspy.com/) Google Trends [](https://trends.google.com/) ScoreApp: Advanced Quiz Funnel Marketing | Make a Quiz Today [](https://www.scoreapp.com/) ScoreApp makes quiz funnel marketing easy, so you can attract relevant warm leads, insightful data and increase your sales. Try for free today Mailmodo - Send Interactive Emails That Drive Conversions [](https://www.mailmodo.com/) Use Mailmodo to create and send interactive emails your customers love. Drive conversions and get better email ROI. Sign up for a free trial now. 185 Top E-Commerce Sites Ranked by User Experience Performance – Baymard Institute [](https://baymard.com/ux-benchmark) See the ranked UX performance of the 185 largest e-commerce sites in the US and Europe. The chart summarizes 50,000+ UX performance ratings. Metricool - Analyze, manage and measure your digital content [](https://metricool.com/) Social media scheduling, web analytics, link in bio and reporting. Metricool is free per live for one brand. START HERE Visualping: #1 Website change detection, monitoring and alerts [](https://visualping.io/) More than 1.5 millions users monitor changes in websites with Visualping, the No1 website change detection, website checker, webpage change monitoring and webpage change detection tool. Gumroad – Sell what you know and see what sticks [](https://gumroad.com/) Gumroad is a powerful, but simple, e-commerce platform. We make it easy to earn your first dollar online by selling digital products, memberships and more. Product Hunt – The best new products in tech. [](https://www.producthunt.com/) Product Hunt is a curation of the best new products, every day. Discover the latest mobile apps, websites, and technology products that everyone's talking about. 12ft Ladder [](https://12ft.io/) Show me a 10ft paywall, I’ll show you a 12ft ladder. namecheckr | Social and Domain Name Availability Search For Brand Professionals [](https://www.namecheckr.com/) Social and Domain Name Availability Search For Brand Professionals Excel AI Formula Generator - Excelformulabot.com [](https://excelformulabot.com/) Transform your text instructions into Excel formulas in seconds with the help of AI. Z-Library [](https://z-lib.org/) Global Print On Demand Platform | Gelato [](https://www.gelato.com/) Create and sell custom products online. With local production in 33 countries, easy integration, and 24/7 customer support, Gelato is an all-in-one platform. Freecycle: Front Door [](https://freecycle.org/) Free eBooks | Project Gutenberg [](https://www.gutenberg.org/) Project Gutenberg is a library of free eBooks. Convertio — File Converter [](https://convertio.co/) Convertio - Easy tool to convert files online. More than 309 different document, image, spreadsheet, ebook, archive, presentation, audio and video formats supported. Namechk [](https://namechk.com/) Crazy Egg Website — Optimization | Heatmaps, Recordings, Surveys & A/B Testing [](https://www.crazyegg.com/) Use Crazy Egg to see what's hot and what's not, and to know what your web visitors are doing with tools, such as heatmaps, recordings, surveys, A/B testing & more. Ifttt [](https://ifttt.com/) Also Asked [](https://alsoasked.com/) Business Name Generator - Easily create Brandable Business Names - Namelix [](https://namelix.com/) Namelix uses artificial intelligence to create a short, brandable business name. Search for domain availability, and instantly generate a logo for your new business Merch Informer [](https://merchinformer.com/) Headline Generator [](https://www.title-generator.com/) Title Generator: create 700 headlines with ONE CLICK: Content Ideas + Catchy Headlines + Ad Campaign E-mail Subject Lines + Emotional Titles. Simple - Efficient - One Click Make [](https://www.make.com/en) Create and add calculator widgets to your website | CALCONIC_ [](https://www.calconic.com/) Web calculator builder empowers you to choose from a pre-made templates or build your own calculator widgets from a scratch without any need of programming knowledge Boost Your Views And Subscribers On YouTube - vidIQ [](https://vidiq.com/) vidIQ helps you acquire the tools and knowledge needed to grow your audience faster on YouTube and beyond. Learn More Last Pass [](https://www.lastpass.com/) Starter Story: Learn How People Are Starting Successful Businesses [](https://www.starterstory.com/) Starter Story interviews successful entrepreneurs and shares the stories behind their businesses. In each interview, we ask how they got started, how they grew, and how they run their business today. How To Say No [](https://www.starterstory.com/how-to-say-no) Saying no is hard, but it's also essential for your sanity. Here are some templates for how to say no - so you can take back your life. Think with Google - Discover Marketing Research & Digital Trends [](https://www.thinkwithgoogle.com/) Uncover the latest marketing research and digital trends with data reports, guides, infographics, and articles from Think with Google. ClickUp™ | One app to replace them all [](https://clickup.com/) Our mission is to make the world more productive. To do this, we built one app to replace them all - Tasks, Docs, Goals, and Chat. The Manual [](https://manual.withcompound.com/) Wealth-planning resources for founders and startup employees Software for Amazon FBA Sellers & Walmart Sellers | Helium 10 [](https://www.helium10.com/) If you're looking for the best software for Amazon FBA & Walmart sellers on the market, check out Helium 10's capabilities online today! Buffer: All-you-need social media toolkit for small businesses [](https://buffer.com/) Use Buffer to manage your social media so that you have more time for your business. Join 160,000+ small businesses today. CPGD — The Consumer Packaged Goods Directory [](https://www.cpgd.xyz/) The Consumer Packaged Goods Directory is a platform to discover new brands and resources. We share weekly trends in our newsletter and partner with services to provide vetted, recommended platforms for our Directory brands. Jungle Scout [](https://www.junglescout.com/) BuzzSumo | The World's #1 Content Marketing Platform [](https://buzzsumo.com/) BuzzSumo powers the strategies of 500k+ marketers, with content marketing data on 8b articles, 42m websites, 300t engagements, 500k journalists & 492m questions. Login - Capital [](https://app.capital.xyz/) Raise, hold, spend, and send funds — all in one place. Marketing Pictory – Video Marketing Made Easy - Pictory.ai [](https://pictory.ai/) Pictory's powerful AI enables you to create and edit professional quality videos using text, no technical skills required or software to download. Tolstoy | Communicate with interactive videos [](https://www.gotolstoy.com/) Start having face-to-face conversations with your customers. Create Email Marketing Your Audience Will Love - MailerLite [](https://www.mailerlite.com/) Email marketing tools to grow your audience faster and drive revenue smarter. Get free access to premium features with a 30-day trial! Sign up now! Hypefury - Schedule & Automate Social Media Marketing [](https://hypefury.com/) Save time on social media while creating more value, and growing your audience faster. Schedule & automate your social media experience! Klaviyo: Marketing Automation Platform for Email & SMS [](https://www.klaviyo.com/) Klaviyo, an ecommerce marketing automation platform for email marketing and sms syncs your tech stack with your website store to scale your business. Online Email & Lead Scraper | Klean Leads [](https://www.kleanleads.com/) Klean Leads is an online email scraper & email address finder. Use it to book more appointments, get more replies, and close more sales. PhantomBuster [](https://phantombuster.com/) Call to Action Examples - 300+ CTA Phrases [](https://ctaexamples.com/) See the best CTA example in every situation covered by the library of 300+ CTA goals. Use the examples to create your own CTAs in minutes. Creative Center: one-stop creative solution for TikTok [](https://ads.tiktok.com/business/creativecenter/pc/en?from=001010) Come to get your next great idea for TikTok. Here you can find the best performing ads, viral videos, and trending hashtags across regions and verticals. Groove.cm GrooveFunnels, GrooveMail with CRM and Digital Marketing Automation Platform - Groove.cm with GrooveFunnels, GroovePages, GrooveKart [](https://groove.cm/) Groove is a website creator, page builder, sales funnel maker, membership site platform, email autoresponder, blog tool, shopping cart system, ecommerce store solution, affiliate manager, video marketing software and more apps to help build your online business. SurveyMonkey: The World’s Most Popular Free Online Survey Tool [](https://www.surveymonkey.com/) Use SurveyMonkey to drive your business forward by using our free online survey tool to capture the voices and opinions of the people who matter most to you. Video Maker | Create Videos Online | Promo.com [](https://promo.com/) Free customizable video maker to help boost your business. Video creator for ads, social media, product and explainer videos, and for anything else you need! beehiiv — The newsletter platform built for growth [](https://www.beehiiv.com/) Access the best tools available in email, helping your newsletter scale and monetize like never before. GetResponse | Professional Email Marketing for Everyone [](https://www.getresponse.com/) No matter your level of expertise, we have a solution for you. At GetResponse, it's email marketing done right. Start your free account today! Search Email Newsletter Archives : Email Tuna [](https://emailtuna.com/) Explore newsletters without subscribing. Get email design ideas, discount coupon codes and exclusive newsletters deals. Database of email newsletters archived from all over the internet. Other Tools Simplescraper — Scrape Websites and turn them into APIs [](https://simplescraper.io/) Web scraping made easy — a powerful and free Chrome extension for scraping websites in your browser, automated in the cloud, or via API. No code required. Exploding Topics - Discover the hottest new trends. [](https://explodingtopics.com/) See new market opportunities, trending topics, emerging technology, hot startups and more on Exploding Topics. Scribe | Visual step-by-step guides [](https://scribehow.com/) By capturing your process while you work, Scribe automatically generates a visual guide, ready to share with the click of a button. Get It Free – The internet's BEST place to find free stuff! [](https://getitfree.us/) The internet's BEST place to find free stuff! Inflact by Ingramer – Marketing toolkit for Instagram [](https://inflact.com/) Sell on Instagram, build your audience, curate content with the right set of tools. Free Online Form Builder & Form Creator | Jotform [](https://www.jotform.com/) We believe the right form makes all the difference. Go from busywork to less work with powerful forms that use conditional logic, accept payments, generate reports, and automate workflows. Manage Your Team’s Projects From Anywhere | Trello [](https://trello.com/en) Trello is the ultimate project management tool. Start up a board in seconds, automate tedious tasks, and collaborate anywhere, even on mobile. TikTok hashtag generator - tiktokhashtags.com [](https://tiktokhashtags.com/) Find out which are the best hashtags for your TikTok post. Create Infographics, Reports and Maps - Infogram [](https://infogram.com/) Infogram is an easy to use infographic and chart maker. Create and share beautiful infographics, online reports, and interactive maps. Make your own here. Confetto - Create Instagram content in minutes [](https://www.confet.to/) Confetto is an all-in-one social media marketing tool built for SMBs and Social Media Managers. Confetto helps you create high-quality content for your audience that maximizes your reach and engagement on social media. Design, copy-write, plan and schedule content all in one place. Find email addresses in seconds • Hunter (Email Hunter) [](https://hunter.io/) Hunter is the leading solution to find and verify professional email addresses. Start using Hunter and connect with the people that matter for your business. PlayPhrase.me: Site for cinema archaeologists. [](https://playphrase.me/) Travel and explore the world of cinema. Largest collection of video quotes from movies on the web. #1 Free SEO Tools → SEO Review Tools [](https://www.seoreviewtools.com/) SEO Review Tools: 42+ Free Online SEO Tools build with ❤! → Rank checker → Domain Authority Checker → Keyword Tool → Backlink Checker Podcastle: Seamless Podcast Recording & Editing [](https://podcastle.ai/) Podcastle is the simplest way to create professional-quality podcasts. Record, edit, transcribe, and export your content with the power of AI, in an intuitive web-based platform. Save Ads from TikTok & Facebook Ad Library - Foreplay [](https://www.foreplay.co/) The best way to save ads from TikTok Creative Center and Facebook Ad Library, Organize them into boards and share ad inspiration with your team. Supercharge your creative strategy. SiteRight - Automate Your Business [](https://www.siteright.co/) SiteRight combines the abilities of multiple online resources into a single dashboard allowing you to have full control over how you manage your business. Diffchecker - Compare text online to find the difference between two text files [](https://www.diffchecker.com/) Diffchecker will compare text to find the difference between two text files. Just paste your files and click Find Difference! Yout.com [](https://yout.com/) Yout.com allows you to record videos from YouTube, FaceBook, SoundCloud, VK and others too many formats with clipping. Intuitively easy to use, with Yout the Internet DVR, with a bit of extra. AI Content Generation | Competitor Analysis - Predis.ai [](https://predis.ai/) Predis helps brands and influencers communicate better on social media by providing AI-powered content strategy analysis, content and hashtag recommendations. Castr | #1 Live Video Streaming Solution With Video Hosting [](https://castr.io/) Castr is a live video streaming solution platform that delivers enterprise-grade live videos globally with CDN. Live event streaming, video hosting, pre-recorded live, multi stream – all in one place using Castr. Headliner - Promote your podcast, radio show or blog with video [](https://www.headliner.app/) Easily create videos to promote your podcast, radio show or blog. Share to Instagram, Facebook, Twitter, YouTube, Linkedin and anywhere video lives Create Presentations, Infographics, Design & Video | Visme [](https://www.visme.co/) Create professional presentations, interactive infographics, beautiful design and engaging videos, all in one place. Start using Visme today. Designrr - Create eBooks, Kindle books, Leadmagnets, Flipbooks and Blog posts from your content in 2 minutes [](https://designrr.io/) Upload any web page, MS Word, Video, Podcast or YouTube and it will create a stunning ebook and convert it to pdf, epub, Kindle or Flipbook. Quick and Easy to use. Full Training, 24x7 Support and Facebook Group Included. SwipeWell | Swipe File Software [](https://www.swipewell.app/) The only Chrome extension dedicated to helping you save, organize, and reference marketing examples (so you never feel stumped). Tango | Create how-to guides, in seconds [](https://www.tango.us/) Tango takes the pain out of documenting processes by automatically generating how-to guides while you work. Empower your team to do their best work. Ad Creative Bank [](https://www.theadcreativebank.com/) Get inspired by ads from across industries, learn new best practices, and start thinking creatively about your brand’s digital creative. Signature Hound • Free Email Signature and Template Generator [](https://signaturehound.com/) Our email signature generator is free and easy to use. Our customizable templates work with Gmail, Outlook, Office 365, Apple Mail and more. Organize All Of Your Marketing In One Place - CoSchedule [](https://coschedule.com/) Get more done in less time with the only work management software for marketers. B Ok - Books [](https://b-ok.xyz/categories) OmmWriter [](https://ommwriter.com/) Ommwriter Rebrandly | Custom URL Shortener, Branded Link Management, API [](https://www.rebrandly.com/) URL Shortener with custom domains. Shorten, brand and track URLs with the industry-leading link management platform. Free to try. API, Short URL, Custom Domains. Common Tools [](https://www.commontools.org/) Book Bolt [](https://bookbolt.io/) Zazzle [](https://www.zazzle.com/) InspiroBot [](https://inspirobot.me/) Download Free Cheat Sheets or Create Your Own! - Cheatography.com: Cheat Sheets For Every Occasion [](https://cheatography.com/) Find thousands of incredible, original programming cheat sheets, all free to download. No Code Chatbot Platform | Free Chatbot Platform | WotNot [](https://wotnot.io/) WotNot is the best no code chatbot platform to build AI bot easily without coding. Deploy bots and live chat on the Website, Messenger, WhatsApp, and more. SpyFu - Competitor Keyword Research Tools for Google Ads PPC & SEO [](https://www.spyfu.com/) Systeme.io - The only tool you need to launch your online business [](https://systeme.io/) Systeme.io has all the tools you need to grow your online business. Click here to create your FREE account! Productivity Temp Mail [](https://temp-mail.org/en/) The Visual Collaboration Platform for Every Team | Miro [](https://miro.com/) Scalable, secure, cross-device and enterprise-ready team collaboration whiteboard for distributed teams. Join 35M+ users from around the world. Grammarly: Free Online Writing Assistant [](https://www.grammarly.com/) Millions trust Grammarly’s free writing app to make their online writing clear and effective. Getting started is simple — download Grammarly’s extension today. Rize · Maximize Your Productivity [](https://rize.io/) Rize is a smart time tracker that improves your focus and helps you build better work habits. Motion | Manage calendars, meetings, projects & tasks in one app [](https://www.usemotion.com/) Automatically prioritize tasks, schedule meetings, and resolve calendar conflicts. Used by over 10k CEOs and professionals to improve focus, get more done, and streamline workday. Notion – One workspace. Every team. [](https://www.notion.so/) We’re more than a doc. Or a table. Customize Notion to work the way you do. Loom: Async Video Messaging for Work | Loom [](https://www.loom.com/) Record your screen, share your thoughts, and get things done faster with async video. Zapier | Automation that moves you forward [](https://zapier.com/) Workflow automation for everyone. Zapier automates your work across 5,000+ app integrations, so you can focus on what matters. Rows — The spreadsheet with superpowers [](https://rows.com/) Combine the power of a spreadsheet with built-in integrations from your business apps. Automate workflows and build tools that make work simpler. Free Online Form Builder | Tally [](https://tally.so/) Tally is the simplest way to create free forms & surveys. Create any type of form in seconds, without knowing how to code, and for free. Highbrow | Learn Something New Every Day. Join for Free! [](https://gohighbrow.com/) Highbrow helps you learn something new every day with 5-minute lessons delivered to your inbox every morning. Join over 400,000 lifelong learners today! Slick Write | Check your grammar. Proofread online. [](https://www.slickwrite.com/#!home) Slick Write is a powerful, FREE application that makes it easy to check your writing for grammar errors, potential stylistic mistakes, and other features of interest. Whether you're a blogger, novelist, SEO professional, or student writing an essay for school, Slick Write can help take your writing to the next level. Reverso [](https://www.reverso.net) Hemingway Editor [](https://hemingwayapp.com/) Web Apps by 123apps - Edit, Convert, Create [](https://123apps.com/) Splitbee – Your all-in-one analytics and conversion platform [](https://splitbee.io/) Track and optimize your online business with Splitbee. Analytics, Funnels, Automations, A/B Testing and more. PDF Tools Free PDF, Video, Image & Other Online Tools - TinyWow [](https://tinywow.com/) Smallpdf.com - A Free Solution to all your PDF Problems [](https://smallpdf.com/) Smallpdf - the platform that makes it super easy to convert and edit all your PDF files. Solving all your PDF problems in one place - and yes, free. Sejda helps with your PDF tasks [](https://www.sejda.com/) Sejda helps with your PDF tasks. Quick and simple online service, no installation required! Split, merge or convert PDF to images, alternate mix or split scans and many other. iLovePDF | Online PDF tools for PDF lovers [](https://www.ilovepdf.com/) iLovePDF is an online service to work with PDF files completely free and easy to use. Merge PDF, split PDF, compress PDF, office to PDF, PDF to JPG and more! Text rewrite QuillBot [](https://quillbot.com/) Pre Post SEO : Online SEO Tools [](https://www.prepostseo.com/) Free Online SEO Tools: plagiarism checker, grammar checker, image compressor, website seo checker, article rewriter, back link checker Wordtune | Your personal writing assistant & editor [](https://www.wordtune.com/) Wordtune is the ultimate AI writing tool that rewrites, rephrases, and rewords your writing! Trusted by over 1,000,000 users, Wordtune strengthens articles, academic papers, essays, emails and any other online content. Aliexpress alternatives CJdropshipping - Dropshipping from Worldwide to Worldwide! [](https://cjdropshipping.com/) China's reliable eCommerce dropshipping fulfillment supplier, helps small businesses ship worldwide, dropship and fulfillment services that are friendly to start-ups and small businesses, Shopify dropshipping. SaleHoo [](https://www.salehoo.com/) Alibaba.com: Manufacturers, Suppliers, Exporters & Importers from the world's largest online B2B marketplace [](https://www.alibaba.com/) Find quality Manufacturers, Suppliers, Exporters, Importers, Buyers, Wholesalers, Products and Trade Leads from our award-winning International Trade Site. Import & Export on alibaba.com Best Dropshipping Suppliers for US + EU Products | Spocket [](https://www.spocket.co/) Spocket allows you to easily start dropshipping top products from US and EU suppliers. Get started for free and see why Spocket consistently gets 5 stars. Best dropshipping supplier to the US [](https://www.usadrop.com/) THE ONLY AMERICAN-MADE FULFILLMENT CENTER IN CHINA. Our knowledge of the Worldwide dropshipping market and the Chinese Supply-Chain can't be beat! 阿里1688 [](https://www.1688.com/) 阿里巴巴(1688.com)是全球企业间(B2B)电子商务的著名品牌,为数千万网商提供海量商机信息和便捷安全的在线交易市场,也是商人们以商会友、真实互动的社区平台。目前1688.com已覆盖原材料、工业品、服装服饰、家居百货、小商品等12个行业大类,提供从原料--生产--加工--现货等一系列的供应产品和服务 Dropshipping Tools Oberlo | Where Self Made is Made [](https://www.oberlo.com/) Start selling online now with Shopify. All the videos, podcasts, ebooks, and dropshipping tools you'll need to build your online empire. Klaviyo: Marketing Automation Platform for Email & SMS [](https://www.klaviyo.com/) Klaviyo, an ecommerce marketing automation platform for email marketing and sms syncs your tech stack with your website store to scale your business. SMSBump | SMS Marketing E-Commerce App for Shopify [](https://smsbump.com/) SMSBump is an SMS marketing & automation app for Shopify. Segment customers, recover orders, send campaign text messages with a 35%+ click through rate. AfterShip: The #1 Shipment Tracking Platform [](https://www.aftership.com/) Order status lookup, branded tracking page, and multi-carrier tracking API for eCommerce. Supports USPS, FedEx, UPS, and 900+ carriers worldwide. #1 Dropshipping App | Zendrop [](https://zendrop.com/) Start and scale your own dropshipping business with Zendrop. Sell and easily fulfill your orders with the fastest shipping in the industry. Best Dropshipping Suppliers for US + EU Products | Spocket [](https://www.spocket.co/) Spocket allows you to easily start dropshipping top products from US and EU suppliers. Get started for free and see why Spocket consistently gets 5 stars. Video Editing Jitter • The simplest motion design tool on the web. [](https://jitter.video/) Animate your designs easily. Export your creations as videos or GIFs. All in your browser. DaVinci Resolve 18 | Blackmagic Design [](https://www.blackmagicdesign.com/products/davinciresolve) Professional video editing, color correction, visual effects and audio post production all in a single application. Free and paid versions for Mac, Windows and Linux. Online Video Editor | Video Creator | InVideo [](https://invideo.io/) InVideo's Online Video Editor Helps You Make Professional Videos From Premium Templates, Images, And Music. All your video needs in one place | Clipchamp [](https://clipchamp.com/) Fast-forward your creations with our video editing platform. Start with a video template or record your webcam or screen. Get the pro look with filters, transitions, text and more. Then, export in minutes and share in an instant. Descript | All-in-one audio/video editing, as easy as a doc. [](https://www.descript.com/) Record, transcribe, edit, mix, collaborate, and master your audio and video with Descript. Download for free →. Kapwing — Reach more people with your content [](https://www.kapwing.com/) Kapwing is a collaborative, online content creation platform that you can use to edit video and create content. Join over 10 million modern creators who trust Kapwing to create, edit, and grow their content on every channel. Panzoid [](https://panzoid.com/) Powerful, free online apps and community for creating beautiful custom content. Google Web Designer - Home [](https://webdesigner.withgoogle.com/) Kapwing — Reach more people with your content [](https://www.kapwing.com/) Kapwing is a collaborative, online content creation platform that you can use to edit video and create content. Join over 10 million modern creators who trust Kapwing to create, edit, and grow their content on every channel. ClipDrop [](https://clipdrop.co/) Create professional visuals without a photo studio CapCut [](https://www.capcut.com/) CapCut is an all-in-one online video editing software which makes creation, upload & share easier, with frame by frame track editor, cloud drive etc. VEED - Online Video Editor - Video Editing Made Simple [](https://www.veed.io/) Make stunning videos with a single click. Cut, trim, crop, add subtitles and more. Online, no account needed. Try it now, free. VEED Free Video Maker | Create & Edit Your Videos Easily - Animoto [](https://animoto.com/k/welcome) Create, edit, and share videos with our online video maker. Combine your photos, video clips, and music to make quality videos in minutes. Get started free! Runway - Online Video Editor | Everything you need to make content, fast. [](https://runwayml.com/) Discover advanced video editing capabilities to take your creations to the next level. CreatorKit - A.I. video creator for marketers [](https://creatorkit.com/) Create videos with just one click, using our A.I. video editor purpose built for marketers. Create scroll stopping videos, Instagram stories, Ads, Reels, and TikTok videos. Pixar in a Box | Computing | Khan Academy [](https://www.khanacademy.org/computing/pixar) 3D Video Motions Plask - AI Motion Capture and 3D Animation Tool [](https://plask.ai/) Plask is an all-in-one browser-based AI motion capture tool and animation editor that anybody can use, from motion designers to every day content creators. Captions Captions [](https://www.getcaptions.app/) Say hello to Captions, the only camera and editing app that automatically transcribes, captions and clips your talking videos for you. Stock videos Pexels [](https://www.pexels.com/) Pixabay [](https://pixabay.com/) Mixkit - Awesome free assets for your next video project [](https://mixkit.co/) Download Free Stock Video Footage, Stock Music & Premiere Pro Templates for your next video editing project. All assets can be downloaded for free! Free Stock Video Footage HD 4K Download Royalty-Free Clips [](https://www.videvo.net/) Download free stock video footage with over 300,000 video clips in 4K and HD. We also offer a wide selection of music and sound effect files with over 180,000 clips available. Click here to download royalty-free licensing videos, motion graphics, music and sound effects from Videvo today. Free Stock Video Footage HD Royalty-Free Videos Download [](https://mazwai.com/) Download free stock video footage with clips available in HD. Click here to download royalty-free licensing videos from Mazwai now. Royalty Free Stock Video Footage Clips | Vidsplay.com [](https://www.vidsplay.com/) Royalty Free Stock Video Footage Clips Free Stock Video Footage, Royalty Free Videos for Download [](https://coverr.co/) Download royalty free (for personal and commercial use), unique and beautiful video footage for your website or any project. No attribution required. Stock Photos Beautiful Free Images & Pictures | Unsplash [](https://unsplash.com/) Beautiful, free images and photos that you can download and use for any project. Better than any royalty free or stock photos. When we share, everyone wins - Creative Commons [](https://creativecommons.org/) Creative Commons licenses are 20! Honoring 20 years of open sharing using CC licenses, join us in 2022 to celebrate Better Sharing — advancing universal access to knowledge and culture, and fostering creativity, innovation, and collaboration. Help us reach our goal of raising $15 million for a future of Better Sharing.  20 Years of Better … Read More "When we share, everyone wins" Food Pictures • Foodiesfeed • Free Food Photos [](https://www.foodiesfeed.com/) Download 2000+ food pictures ⋆ The best free food photos for commercial use ⋆ CC0 license Free Stock Photos and Images for Websites & Commercial Use [](https://burst.shopify.com/) Browse thousands of beautiful copyright-free images. All our pictures are free to download for personal and commercial use, no attribution required. EyeEm | Authentic Stock Photography and Royalty-Free Images [](https://www.eyeem.com/) Explore high-quality, royalty-free stock photos for commercial use. License individual images or save money with our flexible subscription and image pack plans. picjumbo: Free Stock Photos [](https://picjumbo.com/) Free stock photos and images for your projects and websites.️ Beautiful 100% free high-resolution stock images with no watermark. Free Stock Photos, Images, and Vectors [](https://www.stockvault.net/) 139.738 free stock photos, textures, backgrounds and graphics for your next project. No attribution required. Free Stock Photos, PNGs, Templates & Mockups | rawpixel [](https://www.rawpixel.com/) Free images, PNGs, stickers, backgrounds, wallpapers, graphic templates and PSD mockups. All safe to use with commercial licenses. Free Commercial Stock Photos & Royalty Free Images | PikWizard [](https://pikwizard.com/) Free images, videos & free stock photos. Unlimited downloads ✓ Royalty-free Images ✓Copyright-free for commercial use ✓ No Attribution Required Design Bundles [](https://designbundles.net/) Stock music Royalty Free Music for video creators | Epidemic Sound [](https://www.epidemicsound.com/) Download premium Royalty free Music and SFX! Our free trial gives you access to over 35,000 tracks and 90,000 sound effects for video, streaming and more! Royalty-Free Music & SFX for Video Creators | Artlist [](https://artlist.io/) Explore the ultimate royalty-free music & sound effects catalogs for unlimited use in YouTube videos, social media & films created by inspiring indie artists worldwide. The go-to music licensing choice for all creators Royalty Free Audio Tracks - Envato Elements [](https://elements.envato.com/audio) Download Royalty Free Stock Audio Tracks for your next project from Envato Elements. Premium, High Quality handpicked Audio files ideal for any genre. License popular music for videos • Lickd [](https://lickd.co/) The only place you can license popular music for videos. Access 1M+ mainstream tracks, plus high-quality stock music for content creators NCS (NoCopyrightSounds) - free music for content creators [](https://ncs.io/) NCS is a Record Label dedicated to giving a platform to the next generation of Artists in electronic music, representing genres from house to dubstep via trap, drum & bass, electro pop and more. Search Engine Optimization Keyword Tool For Monthly Search Volume, CPC & Competition [](https://keywordseverywhere.com/) Keywords Everywhere is a browser add-on for Chrome & Firefox that shows search volume, CPC & competition on multiple websites. Semrush - Online Marketing Can Be Easy [](https://www.semrush.com/) Turn the algorithm into a friend. Make your business visible online with 55+ tools for SEO, PPC, content, social media, competitive research, and more. DuckDuckGo — Privacy, simplified. [](https://duckduckgo.com/) The Internet privacy company that empowers you to seamlessly take control of your personal information online, without any tradeoffs. SEO Software for 360° Analysis of Your Website [](https://seranking.com/) Leading SEO software for business owners, agencies, and SEO specialists. Track your rankings, monitor competitors, spot technical errors, and more. Skyrocket your organic traffic with Surfer [](https://surferseo.com/) Use Surfer to research, write, optimize, and audit! Everything you need to create a comprehensive content strategy that yields real results is right here. Ahrefs - SEO Tools & Resources To Grow Your Search Traffic [](https://ahrefs.com/) You don't have to be an SEO pro to rank higher and get more traffic. Join Ahrefs – we're a powerful but easy to learn SEO toolset with a passionate community. Neon Tools [](https://neontools.io/) Google Index Search [](https://lumpysoft.com/) Google Index Search SEO Backlink Checker & Link Building Toolset | Majestic.com [](https://majestic.com/) Develop backlink strategies with our Link Intelligence data, build the strongest SEO backlink campaigns to drive organic traffic and boost your rankings today. PageOptimizer Pro [](https://pageoptimizer.pro/) Plans Services SEO Consulting Learn SEO About Blog POP SEO Community Podcast Support POP On Page Workshops With Kyle Roof POP Chrome Extension Guide Tutorial Videos Frequently Asked Questions Best Practices Login Cancel Anytime Plans Services SEO Consulting Learn SEO About Blog POP SEO Community Podcast Support POP On Page… Keyword Chef - Keywords for Publishers [](https://keywordchef.com/) Rank Insanely Fast for Keywords Your Competition Can’t Find “Every long-tail keyword I find ends up ranking within a day” – Dane Eyerly, Owner at TextGoods.com Keyword Chef automatically finds and filters keywords for you. Real-time SERP analysis lets you find keywords nearly guaranteed to rank. Try for free → Let’s face it, most keyword tools ... Read more Notifier - Social Listening for Social Media and More! [](https://notifier.so/) Track keywords. Market your product for free. Drive the conversation. Easy. Free Trial. No obligation ever. Simple. Fast. Trusted by Top Companies. Free Keyword Research Tool from Wordtracker [](https://www.wordtracker.com/) The best FREE alternative to the Keyword Planner. Use Wordtracker to reveal 1000s of profitable longtail keywords with up to 10,000 results per search Blog Posts The 60 Hottest Front-end Tools of 2021 | CSS-Tricks - CSS-Tricks [](https://css-tricks.com/hottest-front-end-tools-in-2021/) A complete list of the most popular front-end tools in 2021, according to the Web Tools Weekly newsletter. See which resources made the list. Resume ResumeGlow - AI Powered Resume Builder [](https://resumeglow.com/) Get hired fast with a resume that grabs attention. Designed by a team of HR experts and typographers. Customizable templates with more than a million possible Create Your Job-winning Resume - (Free) Resume maker · Resume.io [](https://resume.io/) Free online resume maker, allows you to create a perfect Resume or Cover Letter in 5 minutes. See how easy it is to write a professional resume - apply for jobs today! Rezi - The Leading AI-Powered Free Resume Builder [](https://www.rezi.ai/) Rezi’s award-winning AI-powered resume builder is trusted by hundreds of thousands of job seekers. Create your perfect resume in minutes with Rezi. Create a Perfect Resume | Free Resume Builder | Resumaker.ai [](https://resumaker.ai/) Create your professional resume with this online resume maker. Choose a designer-made template and grab any employer attention in seconds. Trusted AI Resume Maker Helps You Get Hired Fast [](https://skillroads.com/) Reach a 96.4% success rate in the job hunt race with the best resume creator. Our innovative technologies and 24/7 support help you to become a perfect candidate for any job. Do not lose your chance to become the One. Kickresume | Best Online Resume & Cover Letter Builder [](https://www.kickresume.com/) Create your best resume yet. Online resume and cover letter builder used by 1,300,000 job seekers worldwide. Professional templates approved by recruiters. ResumeMaker.Online | Create a Professional Resume for Free [](https://www.resumemaker.online/) Save time with the easiest-to-use Resume Maker Online. Create an effective resume in just minutes and land your dream job. No Sign-up required, start now! Interviews Interview Warmup - Grow with Google [](https://grow.google/certificates/interview-warmup/) A quick way to prepare for your next interview. Practice key questions, get insights about your answers, and get more comfortable interviewing. No code website builder Carrd - Simple, free, fully responsive one-page sites for pretty much anything [](https://carrd.co/) A free platform for building simple, fully responsive one-page sites for pretty much anything. Webflow: Create a custom website | No-code website builder [](https://webflow.com/) Create professional, custom websites in a completely visual canvas with no code. Learn how to create a website by trying Webflow for free! Google Sites: Sign-in [](https://sites.google.com/) FlutterFlow - Build beautiful, modern apps incredibly fast! [](https://flutterflow.io/) FlutterFlow lets you build apps incredibly fast in your browser. Build fully functional apps with Firebase integration, API support, animations, and more. Export your code or even easier deploy directly to the app stores! Free Website Builder: Build a Free Website or Online Store | Weebly [](https://www.weebly.com/) Weebly’s free website builder makes it easy to create a website, blog, or online store. Find customizable templates, domains, and easy-to-use tools for any type of business website. Glide • No Code App Builder • Nocode Application Development [](https://www.glideapps.com/) Create the apps your business needs, without coding, waiting or overpaying. Get started for free and build an app today Adalo - Build Your Own No Code App [](https://www.adalo.com/) Adalo makes creating apps as easy as putting together a slide deck. Turn your idea into a real native app — no code needed! Siter.io - The collaborative web design tool, no-code website builder [](https://siter.io/) Siter.io is a visual website builder for designers. Prototype, design, and create responsive websites in the browser. Work together with your team in one place. Elementor: #1 Free WordPress Website Builder | Elementor.com [](https://elementor.com/) Elementor is the platform web creators choose to build professional WordPress websites, grow their skills, and build their business. Start for free today! No code app builder | Bravo Studio [](https://www.bravostudio.app/) Your no-code mobile app builder for iOS and Android. Create MVP’s, validate ideas and publish on App Store and Google Play Store. Home [](https://typedream.com/) The simplest way to build a website with no-code, as easy as writing on Notion. Try Typedream for free and upgrade for custom domains, collaborators, and unlimited pages. Free Website Builder | Create a Free Website | Wix.com [](https://www.wix.com/) Create a website with Wix’s robust website builder. With 900+ strategically designed templates and advanced SEO and marketing tools, build your brand online today. Free responsive Emails & Landing Pages drag-and-drop Editor | BEE [](https://beefree.io/) Free responsive emails and landing pages editor. With BEE drag-and-drop builders embedded in many software applications you can start designing now! Home [](https://typedream.com/) The simplest way to build a website with no-code, as easy as writing on Notion. Try Typedream for free and upgrade for custom domains, collaborators, and unlimited pages. Ownit Connected Checkout [](https://www.ownit.co/) Ownit Connected Checkout Bookmark.com | No-code Website Builder to Start Your Business [](https://www.bookmark.com/) Our AI powered platform ensures your business is future proof. Try Bookmark for free. The best way to build web apps without code | Bubble [](https://bubble.io/) Bubble introduces a new way to build software. It’s a no-code tool that lets you build SaaS platforms, marketplaces and CRMs without code. Bubble hosts all web apps on its cloud platform. Responsive Web Design | Website Creation | Editor X [](https://www.editorx.com/) Experience the future of website design with responsive layouts, CSS precision and smooth drag and drop. Create a Website for Free. Tilda Website Builder [](https://tilda.cc/) Create a website, online store, landing page with Tilda intuitive website builder. Build your site from hundreds of pre-designed templates and publish it today. No code required. No-code headless commerce and websites | Unstack Inc. [](https://www.unstack.com/) Deploy high performance eCommerce storefronts and websites without the engineering overhead using Unstack's no-code CMS Best Drag-and-Drop Website Builder | Jemi [](https://jemi.so/) The modern website builder for creatives, entrepreneurs, and dreamers. Build a beautiful link in bio site, portfolio, or landing page in minutes. No-code website builder that works like Notion [](https://popsy.co/) Create a beautiful no-code website in minutes. Popsy works just like Notion but is built from the ground up for building websites. Choose a free template. Edit content just like in Notion. Customize styles without code. Free Notion icons and illustrations. Unbounce - The Landing Page Builder & Platform [](https://unbounce.com/) Grow your relevance, leads, and sales with Unbounce. Use Unbounce to easily create and optimize landing pages for your small business and boost conversions with AI insights. Low-code Front-end Design & Development Platform | TeleportHQ [](https://teleporthq.io/) Front-end development platform, with a visual builder and headless content modelling capabilities. Static website creation, and UI development tools. Other tools used in no code website MemberSpace - Turn any part of your website into members-only with just a few clicks [](https://www.memberspace.com/) Create memberships on your website for anything you want like courses, video tutorials, member directories, and more while having 100% control over look & feel. Triggre | The number one true no-code platform to run your business [](https://www.triggre.com/) The best no-code platform to create highly advanced business applications in hours, without programming. Try it now for free! No code game builder Welcome to Buildbox [](https://signup.buildbox.com/) Welcome to Buildbox Flowlab Game Creator - Make games online [](https://flowlab.io/) Flowlab is an online game creator. Make your own games to share with friends. Make 2D Games With GameMaker | Free Video Game Maker [](https://gamemaker.io/) Make a game with GameMaker, the best free video game engine. Perfect for beginners and professionals. Learn to build your own 2D games with our simple tutorials. Side Hustle Side Hustle Stack [](https://sidehustlestack.co/) Side Hustle Stack is a resource for finding platform-based work, ranging from gig work and side hustles to platforms that help you start a small business that can grow. Fiverr [](https://www.fiverr.com/) Remotasks: Work From Home, Online Bootcamp Training [](https://www.remotasks.com/en) Make money doing tasks. Start earning today! Free bootcamp training offered online. Sign up for a free Remotasks account and work from home. Earn up to $200/month. Transcribe Speech to Text | Rev [](https://www.rev.com/) Transcribe Speech to Text with Rev. Reach your audience with clear and accurate captions, transcripts, and subtitles. AI Training Data and other Data Management Services [](https://www.clickworker.com/) AI training data, SEO texts, web research, tagging, surveys and more - Use the crowdsourcing principle with the power of >4.5M Clickworkers. Automate your Busy Work - Byron People-Powered Assistants [](https://www.hibyron.com/) Byron is an on demand US based virtual assistant platform that gives individuals and teams the ability to quickly outsource their non-essential tasks. Jobs Websites - Remote Latest Crypto Jobs, Web3 Jobs and Blockchain Jobs in the leading tech companies. [](https://cryptojobslist.com/) New Cryptocurrency Jobs, Web3 Jobs and Blockchain Jobs on CryptoJobsList — the leading site to find and post jobs. Connect with companies hiring in a few clicks and begin your next experience in the industry. Updated daily. Remote Jobs: Design, Marketing, Programming, Writing & More [](https://justremote.co/) Discover Remote Jobs from around the world. Give up the commute, work remotely and do what you love, daily, from anywhere. Find your perfect remote development, design, sales or marketing job today. Remote Ok [](https://remoteok.com/) Hire Freelancers & Remote Workers For Free [](https://talent.hubstaff.com/) Find and hire the highest quality freelancers from around the world - for free. Choose from thousands of developers, digital marketers, creatives and more. We Work Remotely: Remote jobs in design, programming, marketing and more [](https://weworkremotely.com/) Find the most qualified people in the most unexpected places: Hire remote! We Work Remotely is the best place to find and list remote jobs that aren't restricted by commutes or a particular geographic area. Browse thousands of remote work jobs today. Angel [](https://angel.co/) Remote Work: Jobs, Companies & Virtual Teams - Remote.co [](https://remote.co/) Remote.co is the definitive remote work job board for online job seekers and companies hiring. Start your remote job search here! FlexJobs: Best Remote Jobs, Work from Home Jobs, Online Jobs & More [](https://www.flexjobs.com/) The #1 job search site for hand-screened flexible and remote jobs (work from home jobs) since 2007. Plus get resume, coaching and career help. Join today! Remote jobs remotefront.io [](https://remotefront.io/) All remote jobs at remotefront.io Daily Virtual Events Helping You Grow Professionally [](https://powertofly.com/) PowerToFly is where you receive expert career advice, free video training, coaching and exclusive access to jobs and events at top companies. Best Remote and Work from Home Jobs - Virtual Vocations [](https://www.virtualvocations.com/) Best work from home jobs and remote jobs in over 50 categories for professionals, digital nomads, telecommuting workers and entry level jobseekers. Education, healthcare, medical, customer support and tech job openings. Remote Jobs | Working Nomads [](https://www.workingnomads.com/jobs) Remote jobs for digital working nomads. Start your telecommuting career and work remotely from home or places around the world. Job Search, Companies Hiring Near Me, and Advice | The Muse [](https://www.themuse.com/) Find jobs at the best companies hiring near you and get free career advice. Startupers [](https://www.startupers.com/) NoDesk - Where Everyone Works Remote [](https://nodesk.co/) Browse and apply to the best new remote jobs at leading remote companies and startups for free. Join hundreds of companies that use NoDesk to build their remote teams. Browser Extensions Blackbox - Select. Copy. Paste & Search - Magazinul web Chrome [](https://chrome.google.com/webstore/detail/blackbox-select-copy-past/mcgbeeipkmelnpldkobichboakdfaeon) Fastest Way to Copy Text from Videos & Images Octotree - GitHub code tree - Magazinul web Chrome [](https://chrome.google.com/webstore/detail/octotree-github-code-tree/bkhaagjahfmjljalopjnoealnfndnagc) GitHub on steroids WhatFont - Chrome Web Store [](https://chrome.google.com/webstore/detail/whatfont/jabopobgcpjmedljpbcaablpmlmfcogm?hl=en) The easiest way to identify fonts on web pages. Window Resizer - Chrome Web Store [](https://chrome.google.com/webstore/detail/window-resizer/kkelicaakdanhinjdeammmilcgefonfh?hl=en) Resize the browser window to emulate various screen resolutions. Amino: CSS Editor - Magazinul web Chrome [](https://chrome.google.com/webstore/detail/amino-css-editor/pbcpfbcibpcbfbmddogfhcijfpboeaaf) Live CSS Editor. Write custom CSS for any website and see your changes in real time. Checkbot: SEO, Web Speed & Security Tester 🚀 - Chrome Web Store [](https://chrome.google.com/webstore/detail/checkbot-seo-web-speed-se/dagohlmlhagincbfilmkadjgmdnkjinl?hl=en) Test SEO/speed/security of 100s of pages in a click! Check broken links, HTML/JavaScript/CSS, URL redirects, duplicate titles... Honey: Automatic Coupons & Rewards - Magazinul web Chrome [](https://chrome.google.com/webstore/detail/honey-automatic-coupons-r/bmnlcjabgnpnenekpadlanbbkooimhnj) Save money and earn rewards when you shop online. Tango: screenshots, training, & documentation - Magazinul web Chrome [](https://chrome.google.com/webstore/detail/tango-screenshots-trainin/lggdbpblkekjjbobadliahffoaobaknh) Automatically create beautiful step-by-step guides with screenshots, in seconds. No code browser automation | axiom.ai [](https://axiom.ai/) Build browser bots quickly, without code. Automate website actions and repetitive tasks using just your browser, on any website or web app. No Code Browser extensions builder Bildr - Visual Web Development in your Browser [](https://www.bildr.com/) Visually build SaaS products, Chrome extensions, and web3 dApps Other Repurposing content for social media the easy way » Repurpose.io [](https://repurpose.io/) Repurposing content for social media made easy. Automatically repurpose YouTube, TikTok, Lives, Podcasts, and Zoom calls. Try it for FREE. Smart Serials: Your serial numbers database [](https://smartserials.com/) This is your main source of free serial numbers, unlock keys in a clean environment safe to browse by all ages. Old versions of Windows, Mac and Linux Software, Apps & Abandonware Games - Download at OldVersion.com [](http://www.oldversion.com/) Online Room Planner - Design Your Room [](http://www.planyourroom.com/) Planyourroom.com is a wonderful website to redesign each room in your house by picking out perfect furniture options to fit your unique space. BoredHumans.com - Fun AI Programs You Can Use Online [](https://boredhumans.com/) Fun AI programs you can use online. AI games, fake people, computer generated art, machine learning demos, and more. BNProject | Home [](https://buynothingproject.org/) Open Source Alternatives to Proprietary Software [](https://www.opensourcealternative.to/) Discover 400+ popular open source alternatives to proprietary SaaS. URL Shortener - Short URLs & Custom Free Link Shortener | Bitly [](https://bitly.com/) Bitly’s Connections Platform is more than a free URL shortener, with robust link management software, advanced QR Code features, and a Link-in-bio solution. TinEye Reverse Image Search [](https://tineye.com/) Good Books | Books recommended by successful people [](https://www.goodbooks.io/) Looking for the best books to read in 2022? Discover the best book recommendations from the world's most successful, influential and interesting people. Directory - Website Recommendations [](https://tokapps.com/directory/) 0 TRIED & TESTED WEBSITES LISTED Insanely Useful Websites A combination of useful websites for businesses, freelancers, DIYers, and individuals in a centralised area.All websites have been tried and tested. Filter Websites Audio Business Tools Copywriting Design Entertainment Graphics Guides Health Marketing PC Resources Savings SEO Software Travel Video Apply filter Watch Anime Online, Free Anime Streaming Online on Zoro.to Anime Website [](https://zoro.to/) Zoro is a Free anime streaming website which you can watch English Subbed and Dubbed Anime online with No Account and Daily update. WATCH NOW! Animated Drawings [](https://sketch.metademolab.com/) Bring children's drawings to life, by animating characters to move around! Alternativeto [](https://alternativeto.net/) Chatroulette [](https://chatroulette.com/) Random meetings around the world Tiktok Downloader - Download Video tiktok Without Watermark - SnapTik [](https://snaptik.app/en) TikTok Video Downloader - SnapTik.App is one of the best free Download video Tiktok No Watermark tool available online. You can download TikTok video from any device you have. Imgflip - Create and Share Awesome Images [](https://imgflip.com/) Flip through memes, gifs, and other funny images. Make your own images with our Meme Generator or Animated GIF Maker. Fake Text Message | Make Fake Text Conversation [](https://ifaketextmessage.com/) Fake Text Message is a tool to create a Fake Text Conversation and a Fake iMessage. ✂Templatemaker ︎ [](https://www.templatemaker.nl/en/) Omni Calculator [](https://www.omnicalculator.com/) Omni Calculator solves 2960 problems anywhere from finance and business to health. It’s so fast and easy you won’t want to do the math again! Watch Movies Online Free | Watch Series HD Free [](https://hdtoday.tv/) Free Access to the Biggest library of HD Movies and HD Series online - NO ADS - No Account Required - Fast Free Streaming Students Answers - The Most Trusted Place for Answering Life's Questions [](https://www.answers.com/) Answers is the place to go to get the answers you need and to ask the questions you want Wolfram|Alpha: Computational Intelligence [](https://www.wolframalpha.com/) Compute answers using Wolfram's breakthrough technology & knowledgebase, relied on by millions of students & professionals. For math, science, nutrition, history, geography, engineering, mathematics, linguistics, sports, finance, music… Online Math Tools - Simple, free and easy to use math utilities [](https://onlinemathtools.com/) World's simplest collection of useful mathematics utilities. Generate number sequences, draw fractals, do quick matrix and numerical calculations and more! edX | Free Online Courses by Harvard, MIT, & more | edX [](https://www.edx.org/) Access 2000 free online courses from 140 leading institutions worldwide. Gain new skills and earn a certificate of completion. Join today. Sci-Hub [](https://sci-hub.hkvisa.net/) Sci-Hub,mg.scihub.ltd,sci-hub.tw,The project is supported by user donations. Imagine the world with free access to knowledge for everyone ‐ a world without any paywalls. DigitalDefynd - Find the Best + Free Courses Online [](https://digitaldefynd.com/) 4 Million+ Learners | 96,000+ Courses | 45,000+ Free Courses | 1200+ Free Certificates Learn Anything [](https://learn-anything.xyz/) Search Interactive Mind Maps to learn anything HubSpot Academy - Homepage [](https://academy.hubspot.com/) HubSpot Academy is the worldwide leader in inbound marketing, sales, and customer service/support training.

Stop Learning Excel—Meet the AI Spreadsheet
youtube
LLM Vibe Score0.335
Human Vibe Score0.41
Kevin StratvertDec 13, 2024

Stop Learning Excel—Meet the AI Spreadsheet

Mastering Excel used to mean memorizing complex formulas like VLOOKUP, creating pivot tables, and manually sorting data. But now, AI spreadsheets are here to change the game! In this video, I showcase 7 ways AI makes spreadsheets effortless, even for beginners. With Bricks, an AI-powered and free spreadsheet tool, I’ll demonstrate how you can: Automate table joins without formulas Sort data with simple prompts Apply conditional formatting in seconds Filter data dynamically Summarize or group data effortlessly Create charts automatically Remove duplicates with ease Whether you're a spreadsheet pro or just getting started, this video will show you how AI can handle all the hard work for you. I’ve even included a sample Excel workbook so you can follow along and try these features for yourself. Are you ready to embrace the future of spreadsheets? Watch now and see why it might be time to stop learning Excel and start using AI! Host: Kevin Stratvert 📚 RESOURCES Download the sample workbook: https://1drv.ms/x/s!AmxrofZZlZ-whfhLV1BgrO5mxYgTsg?e=nEousp Sign up for Bricks: https://bit.ly/newaispreadsheet ⌚ TIMESTAMPS 00:00 - Introduction 00:28 - Get Bricks 01:02 - Effortless Table Joins with AI 02:54 - Simplified Sorting with AI 03:58 - Conditional Formatting with AI 05:03 - Filtering Made Smarter with AI 06:20 - AI Pivot Tables for Instant Insights 07:09 - AI Charts 07:59 - Removing Duplicates with AI 09:14 - Bonus: Data Types 11:51 - Export to Excel 12:12 - Wrap Up 📺 RELATED VIDEOS Playlist with all my videos on Bricks: https://www.youtube.com/playlist?list=PLlKpQrBME6xLZLJCmqdM4i5GQhXscRvTS 📩 NEWSLETTER Get the latest high-quality tutorial and tips and tricks videos emailed to your inbox each week: https://kevinstratvert.com/newsletter/ 🔽 CONNECT WITH ME Official website: http://www.kevinstratvert.com LinkedIn: https://www.linkedin.com/in/kevinstratvert/ Discord: https://bit.ly/KevinStratvertDiscord Twitter: https://twitter.com/kevstrat Facebook: https://www.facebook.com/Kevin-Stratvert-101912218227818 TikTok: https://www.tiktok.com/@kevinstratvert Instagram: https://www.instagram.com/kevinstratvert/ 🎁 TOOLS AND DISCOUNTS ✅ 🎙️ Voicemod AI Voice Changer | 5% off | https://link.xsolla.com/KZBi89AY ✅ 🌐 Squarespace Websites | https://squarespace.syuh.net/XYaqYM ✅ 🔍 Grammarly | https://grammarly.go2cloud.org/SH3nL ✅ 📹 CapCut | https://bit.ly/installcapcut ✅ 🛍️ Shopify | https://shopify.pxf.io/XY9rPa ✅ 📋 Notion | https://affiliate.notion.so/rffva4tr71ax ✅ 🖼️ Figma | https://psxid.figma.com/lqjg97licpry ✅ 🤖 ElevenLabs Text-to-Speech | https://try.elevenlabs.io/taqepq60mptr ✅ 💵 Quickbooks Online | https://bit.ly/intuitquickbooksonline ✅ 👥 Hubspot | https://hubspot.sjv.io/DKo6jb ✅ 📈 Semrush | https://bit.ly/semrush14dayfreetrial ✅ 🎥 Descript | https://get.descript.com/sf22jb63w2tx ✅ 🏓 Smartsheet | https://bit.ly/trysmartsheet 🎒 MY COURSES Go from Excel novice to data analysis ninja in just 2 hours: https://kevinstratvert.thinkific.com/ 🙏 REQUEST VIDEOS https://forms.gle/BDrTNUoxheEoMLGt5 🔔 SUBSCRIBE ON YOUTUBE https://www.youtube.com/user/kevlers?sub_confirmation=1 🙌 SUPPORT THE CHANNEL Hit the THANKS button in any video! Amazon affiliate link: https://amzn.to/3kCP2yz ⚖ DISCLOSURE Some links are affiliate links. Purchasing through these links gives me a small commission to support videos on this channel. The price to you is the same. #stratvert #bricks

ai automation agency: making $200,000 a month from building automated marketing workflows
youtube
LLM Vibe Score0.355
Human Vibe Score0.41
Cody SchneiderDec 4, 2024

ai automation agency: making $200,000 a month from building automated marketing workflows

Sub to my newsletter for growth tactics and startup ideas - https://investorupdate.beehiiv.com/subscribe In the Pit Podcast with Cody Schneider Talent Fiber: Hire marketing specialists 80% less than US equivalents - https://talentfiber.com/ Swell AI: Content marketing powered by AI - https://www.swellai.com/ Drafthorse AI: Write and publish hundres of SEO for blog posts in minutes - https://www.drafthorseai.com/ Landing Cat: Build thousands of ecommerce collection pages in minutes - https://www.landingcat.com/ Summary In this episode, I chat with Michael Greenberg about AI automation in marketing services. We discuss building AI automation agencies, opportunities in productized services, and specific AI-powered marketing workflows. Michael shares insights on content creation strategies, including social media posts, podcasts, and virtual influencers. We also explore the technical aspects of implementing AI systems and the business considerations for entrepreneurs in this space. Michael provides perspectives on the challenges of running an AI automation agency and balancing experimentation with focus in entrepreneurship. Timestamps: 0:00 - Process Automation in Marketing 10:20 - Process Automation in Marketing 18:41 - AI-Powered Ghostwriting System 23:32 - Generating Content at Scale with AI 28:23 - AI Avatars and Virtual Influencers 35:13 - Creating Artificial Controversy with AI 47:35 - Balancing Experimentation and Focus in Business Host Links Personal email newsletter - https://investorupdate.beehiiv.com/subscribe https://twitter.com/codyschneiderxx https://www.linkedin.com/in/codyxschneider/ https://codyschneider.com/ https://inthepitpodcast.com/ Guest Links https://x.com/gentoftech https://www.linkedin.com/in/gentoftech/ https://www.3rdbrain.co/

ai-learning-roadmap
github
LLM Vibe Score0.442
Human Vibe Score0.035708035270567436
gopala-krNov 30, 2024

ai-learning-roadmap

Lists of all AI related learning materials and practical tools to get started with AI apps Design Thinking – An Introduction Stanford's virtual Crash Course in Design Thinking Amazon Web Services Learning Material AWS AI Session– The session provides an overview of all Amazon AI technology offerings (Lex, Polly, Rekognition, ML, and Deep Learning AMI) Self-Paced Labs AWS self-paced labs provide hands-on practice in a live AWS environment with AWS services and real-world cloud scenarios. Follow step-by-step instructions to learn a service, practice a use case, or prepare for AWS Certification. Introductory Lab Introduction to AWS Lambda Lex Introduction to Amazon Lex Amazon Lex Webinar Amazon Lex: AWS conversational interface (chat bot) Documentation Polly Introduction to Amazon Polly Amazon Polly Webinar - Amazon Polly – AWS Text To Speech (TTS) service Documentation What is Amazon Polly? Developer Resources Rekognition Introduction to Amazon Rekognition Amazon Rekognition - Deep Learning-Based Image Analysis Webinar Amazon Rekognition – AWS image recognition service Documentation – What is Amazon Rekognition? Machine Learning Machine Learning Session 1 – Empowering Developers to Build Smart Applications Session 2 - Predicting Customer Churn with Amazon Machine Learning AWS Machine Learning – End to end, managed service for creating and testing ML models and then deploying those models into production Documentation What is Amazon Machine Learning? Developer Resources AWS Deep Learning AMI – Amazon Machine Image (AMI) optimized for deep learning efforts Recommended Additional Resources Take your skills to the next level with fundamental, advanced, and expert level labs. Creating Amazon EC2 Instances with Microsoft Windows Building Your First Amazon Virtual Private Cloud (VPC) Working with AWS CodeCommit on Windows Working with Amazon DynamoDB Google Cloud - Learning Material Below is the learning material that will help you learn about Google Cloud. Network Networking 101 – 43 mins The codelab provides common cloud developer experience as follows: Set up your lab environment and learn how to work with your GCP environment. Use of common open source tools to explore your network around the world. Deploy a common use case: use of HTTP Load Balancing and Managed Instance Groups to host a scalable, multi-region web server. Testing and monitoring your network and instances. Cleanup. Developing Solutions for Google Cloud Platform – 8 hours Infrastructure Build a Slack Bot with Node.js on Kubernotes – 43 mins Creating a Virtual Machine – 10 mins Getting Started with App Engine (Python) – 13 mins Data Introduction to Google Cloud Data Prep – 7 mins Create a Managed MySQL database with Cloud SQL – 19 mins Upload Objects to Cloud Storage – 11 mins AI, Big Data & Machine Learning Introduction to Google Cloud Machine Learning – 1 hour Machine Learning APIs by Example – 30 min Google Cloud Platform Big Data and Machine Learning Fundamentals Additional AI Materials Auto-awesome: Advanced Data Science on Google Cloud Platform – 45 min Run a Big Data Text Processing Pipeline in Cloud Dataflow – 21 min Image Classification Using Cloud ML Engine & Datalab – 58 min Structured Data Regression Using Cloud ML Engine & Datalab – 58 min (Optional) Deep Learning & Tensorflow Tensorflow and Deep Learning Tutorial – 2:35 hours Deep Learning Course – advanced users only Additional Reference Material Big Data & Machine Learning @ Google Cloud Next '17 - A collection of 49 videos IBM Watson Learning Material (Contributions are welcome in this space) [IBM Watson Overview]() [IBM Watson Cognitive APIs]() [IBM Watson Knowledge Studio]() Visual Studio UCI datasets Microsoft Chat Bots Learning Material Skills Prerequisite Git and Github NodeJS VS Code IDE Training Paths If you have the above Prerequisite skills, then take Advanced Training Path else take Novice Training Path. Prerequisite Tutorials Git and Github Node.js Node.js Tutorials for Beginners Node.js Tutorial in VS Code Introduction To Visual Studio Code Novice Training Path Environment Set Up Download and Install Git Set up GitHub Account_ Download and Install NodeJS Download and Install IDE - Visual Studio Code Download and Install the Bot Framework Emulator Git clone the Bot Education project - git clone Set Up Azure Free Trial Account Cognitive Services (Defining Intelligence) Read Cognitive Services ADS Education Deck – git clone Review the guide for Understanding Natural language with LUIS Complete the NLP (LUIS) Training Lab from the installed Bot Education project – \bot-education\Student-Resources\Labs\CognitiveServices\Lab_SetupLanguageModel.md Bot Framework (Building Chat Bots) Read Bot Framework ADS Education Deck from downloaded - (Your Path)\bot-extras Review Bot Framework documentation (Core Concepts, Bot Builder for NodeJS, and Bot Intelligence) - Setup local environment and run emulator from the installed Bot Education project – \bot-education\Student-Resources\Labs\Node\Lab1_SetupCheckModel.md Review and test in the emulator the “bot-hello” from \bot-education\Student-Resources\BOTs\Node\bot-hello Advanced Training Path Environment Set Up Download and Install Git Set up GitHub Account_ Download and Install NodeJS Download and Install IDE - Visual Studio Code Download and Install the Bot Framework Emulator Git clone the Bot Education project - git clone Set Up Azure Free Trial Account Git clone the Bot Builder Samples – git clone Cognitive Services (Defining Intelligence) Read Cognitive Services ADS Education Deck – git clone Review the guide for Understanding Natural language with LUIS Bot Framework (Building Chat Bots) Read Bot Framework ADS Education Deck from downloaded - (Your Path)\bot-extras Review Bot Framework documentation (Core Concepts, Bot Builder for NodeJS, and Bot Intelligence) - Setup local environment and run emulator from the installed Bot Education project – \bot-education\Student-Resources\Labs\Node\Lab1_SetupCheckModel.md Cognitive Services (Defining Intelligence) - Labs Complete the NLP (LUIS) Training Lab from the installed BOT Education project \bot-education\Student-Resources\Labs\CognitiveServices\Lab_SetupLanguageModel.md Review, Deploy and run the LUIS BOT sample Bot Framework (Building Chat Bots) – Labs Setup local environment and run emulator from the installed Bot Education project \bot-education\Student-Resources\Labs\Node\Lab1_SetupCheckModel.md Review and test in the emulator the “bot-hello” from \bot-education\Student-Resources\BOTs\Node\bot-hello Review and test in the emulator the “bot-recognizers” from \bot-education\Student-Resources\BOTs\Node\bot-recognizers Lecture Videos Source Berkeley Lecture TitleLecturerSemester Lecture 1 Introduction Dan Klein Fall 2012 Lecture 2 Uninformed Search Dan Klein Fall 2012 Lecture 3 Informed Search Dan Klein Fall 2012 Lecture 4 Constraint Satisfaction Problems I Dan Klein Fall 2012 Lecture 5 Constraint Satisfaction Problems II Dan Klein Fall 2012 Lecture 6 Adversarial Search Dan Klein Fall 2012 Lecture 7 Expectimax and Utilities Dan Klein Fall 2012 Lecture 8 Markov Decision Processes I Dan Klein Fall 2012 Lecture 9 Markov Decision Processes II Dan Klein Fall 2012 Lecture 10 Reinforcement Learning I Dan Klein Fall 2012 Lecture 11 Reinforcement Learning II Dan Klein Fall 2012 Lecture 12 Probability Pieter Abbeel Spring 2014 Lecture 13 Markov Models Pieter Abbeel Spring 2014 Lecture 14 Hidden Markov Models Dan Klein Fall 2013 Lecture 15 Applications of HMMs / Speech Pieter Abbeel Spring 2014 Lecture 16 Bayes' Nets: Representation Pieter Abbeel Spring 2014 Lecture 17 Bayes' Nets: Independence Pieter Abbeel Spring 2014 Lecture 18 Bayes' Nets: Inference Pieter Abbeel Spring 2014 Lecture 19 Bayes' Nets: Sampling Pieter Abbeel Fall 2013 Lecture 20 Decision Diagrams / Value of Perfect Information Pieter Abbeel Spring 2014 Lecture 21 Machine Learning: Naive Bayes Nicholas Hay Spring 2014 Lecture 22 Machine Learning: Perceptrons Pieter Abbeel Spring 2014 Lecture 23 Machine Learning: Kernels and Clustering Pieter Abbeel Spring 2014 Lecture 24 Advanced Applications: NLP, Games, and Robotic Cars Pieter Abbeel Spring 2014 Lecture 25 Advanced Applications: Computer Vision and Robotics Pieter Abbeel Spring 2014 Additionally, there are additional Step-By-Step videos which supplement the lecture's materials. These videos are listed below: Lecture TitleLecturerNotes SBS-1 DFS and BFS Pieter Abbeel Lec: Uninformed Search SBS-2 A* Search Pieter Abbeel Lec: Informed Search SBS-3 Alpha-Beta Pruning Pieter Abbeel Lec: Adversarial Search SBS-4 D-Separation Pieter Abbeel Lec: Bayes' Nets: Independence SBS-5 Elimination of One Variable Pieter Abbeel Lec: Bayes' Nets: Inference SBS-6 Variable Elimination Pieter Abbeel Lec: Bayes' Nets: Inference SBS-7 Sampling Pieter Abbeel Lec: Bayes' Nets: Sampling SBS-8 Gibbs' Sampling Michael Liang Lec: Bayes' Nets: Sampling --> SBS-8 Maximum Likelihood Pieter Abbeel Lec: Machine Learning: Naive Bayes SBS-9 Laplace Smoothing Pieter Abbeel Lec: Machine Learning: Naive Bayes SBS-10 Perceptrons Pieter Abbeel Lec: Machine Learning: Perceptrons Per-Semester Video Archive(Berkeley) The lecture videos from the most recent offerings are posted below. Spring 2014 Lecture Videos Fall 2013 Lecture Videos Spring 2013 Lecture Videos Fall 2012 Lecture Videos Spring 2014 Lecture TitleLecturerNotes Lecture 1 Introduction Pieter Abbeel Lecture 2 Uninformed Search Pieter Abbeel Lecture 3 Informed Search Pieter Abbeel Lecture 4 Constraint Satisfaction Problems I Pieter Abbeel Recording is a bit flaky, see Fall 2013 Lecture 4 for alternative Lecture 5 Constraint Satisfaction Problems II Pieter Abbeel Lecture 6 Adversarial Search Pieter Abbeel Lecture 7 Expectimax and Utilities Pieter Abbeel Lecture 8 Markov Decision Processes I Pieter Abbeel Lecture 9 Markov Decision Processes II Pieter Abbeel Lecture 10 Reinforcement Learning I Pieter Abbeel Lecture 11 Reinforcement Learning II Pieter Abbeel Lecture 12 Probability Pieter Abbeel Lecture 13 Markov Models Pieter Abbeel Lecture 14 Hidden Markov Models Pieter Abbeel Recording is a bit flaky, see Fall 2013 Lecture 18 for alternative Lecture 15 Applications of HMMs / Speech Pieter Abbeel Lecture 16 Bayes' Nets: Representation Pieter Abbeel Lecture 17 Bayes' Nets: Independence Pieter Abbeel Lecture 18 Bayes' Nets: Inference Pieter Abbeel Lecture 19 Bayes' Nets: Sampling Pieter Abbeel Unrecorded, see Fall 2013 Lecture 16 Lecture 20 Decision Diagrams / Value of Perfect Information Pieter Abbeel Lecture 21 Machine Learning: Naive Bayes Nicholas Hay Lecture 22 Machine Learning: Perceptrons Pieter Abbeel Lecture 23 Machine Learning: Kernels and Clustering Pieter Abbeel Lecture 24 Advanced Applications: NLP, Games, and Robotic Cars Pieter Abbeel Lecture 25 Advanced Applications: Computer Vision and Robotics Pieter Abbeel Lecture 26 Conclusion Pieter Abbeel Unrecorded Fall 2013 Lecture TitleLecturerNotes Lecture 1 Introduction Dan Klein Lecture 2 Uninformed Search Dan Klein Lecture 3 Informed Search Dan Klein Lecture 4 Constraint Satisfaction Problems I Dan Klein Lecture 5 Constraint Satisfaction Problems II Dan Klein Lecture 6 Adversarial Search Dan Klein Lecture 7 Expectimax and Utilities Dan Klein Lecture 8 Markov Decision Processes I Dan Klein Lecture 9 Markov Decision Processes II Dan Klein Lecture 10 Reinforcement Learning I Dan Klein Lecture 11 Reinforcement Learning II Dan Klein Lecture 12 Probability Pieter Abbeel Lecture 13 Bayes' Nets: Representation Pieter Abbeel Lecture 14 Bayes' Nets: Independence Dan Klein Lecture 15 Bayes' Nets: Inference Pieter Abbeel Lecture 16 Bayes' Nets: Sampling Pieter Abbeel Lecture 17 Decision Diagrams / Value of Perfect Information Pieter Abbeel Lecture 18 Hidden Markov Models Dan Klein Lecture 19 Applications of HMMs / Speech Dan Klein Lecture 20 Machine Learning: Naive Bayes Dan Klein Lecture 21 Machine Learning: Perceptrons Dan Klein Lecture 22 Machine Learning: Kernels and Clustering Pieter Abbeel Lecture 23 Machine Learning: Decision Trees and Neural Nets Pieter Abbeel Lecture 24 Advanced Applications: NLP and Robotic Cars Dan Klein Unrecorded, see Spring 2013 Lecture 24 Lecture 25 Advanced Applications: Computer Vision and Robotics Pieter Abbeel Lecture 26 Conclusion Dan Klein,Pieter Abbeel Unrecorded Spring 2013 Lecture TitleLecturerNotes Lecture 1 Introduction Pieter Abbeel Video Down Lecture 2 Uninformed Search Pieter Abbeel Lecture 3 Informed Search Pieter Abbeel Lecture 4 Constraint Satisfaction Problems I Pieter Abbeel Lecture 5 Constraint Satisfaction Problems II Pieter Abbeel Unrecorded, see Fall 2012 Lecture 5 Lecture 6 Adversarial Search Pieter Abbeel Lecture 7 Expectimax and Utilities Pieter Abbeel Lecture 8 Markov Decision Processes I Pieter Abbeel Lecture 9 Markov Decision Processes II Pieter Abbeel Lecture 10 Reinforcement Learning I Pieter Abbeel Lecture 11 Reinforcement Learning II Pieter Abbeel Lecture 12 Probability Pieter Abbeel Lecture 13 Bayes' Nets: Representation Pieter Abbeel Lecture 14 Bayes' Nets: Independence Pieter Abbeel Lecture 15 Bayes' Nets: Inference Pieter Abbeel Lecture 16 Bayes' Nets: Sampling Pieter Abbeel Lecture 17 Decision Diagrams / Value of Perfect Information Pieter Abbeel Lecture 18 Hidden Markov Models Pieter Abbeel Lecture 19 Applications of HMMs / Speech Pieter Abbeel Lecture 20 Machine Learning: Naive Bayes Pieter Abbeel Lecture 21 Machine Learning: Perceptrons I Nicholas Hay Lecture 22 Machine Learning: Perceptrons II Pieter Abbeel Lecture 23 Machine Learning: Kernels and Clustering Pieter Abbeel Lecture 24 Advanced Applications: NLP and Robotic Cars Pieter Abbeel Lecture 25 Advanced Applications: Computer Vision and Robotics Pieter Abbeel Lecture 26 Conclusion Pieter Abbeel Unrecorded Fall 2012 Lecture TitleLecturerNotes Lecture 1 Introduction Dan Klein Lecture 2 Uninformed Search Dan Klein Lecture 3 Informed Search Dan Klein Lecture 4 Constraint Satisfaction Problems I Dan Klein Lecture 5 Constraint Satisfaction Problems II Dan Klein Lecture 6 Adversarial Search Dan Klein Lecture 7 Expectimax and Utilities Dan Klein Lecture 8 Markov Decision Processes I Dan Klein Lecture 9 Markov Decision Processes II Dan Klein Lecture 10 Reinforcement Learning I Dan Klein Lecture 11 Reinforcement Learning II Dan Klein Lecture 12 Probability Pieter Abbeel Lecture 13 Bayes' Nets: Representation Pieter Abbeel Lecture 14 Bayes' Nets: Independence Pieter Abbeel Lecture 15 Bayes' Nets: Inference Pieter Abbeel Lecture 16 Bayes' Nets: Sampling Pieter Abbeel Lecture 17 Decision Diagrams / Value of Perfect Information Pieter Abbeel Lecture 18 Hidden Markov Models Pieter Abbeel Lecture 19 Applications of HMMs / Speech Dan Klein Lecture 20 Machine Learning: Naive Bayes Dan Klein Lecture 21 Machine Learning: Perceptrons Dan Klein Lecture 22 Machine Learning: Kernels and Clustering Dan Klein Lecture 23 Machine Learning: Decision Trees and Neural Nets Pieter Abbeel Lecture 24 Advanced Applications: Computer Vision and Robotics Pieter Abbeel Lecture 25 Advanced Applications: NLP and Robotic Cars Dan Klein,Pieter Abbeel Unrecorded Lecture 26 Conclusion Dan Klein,Pieter Abbeel Unrecorded Lecture Slides Here is the complete set of lecture slides, including videos, and videos of demos run in lecture: Slides [~3 GB]. The list below contains all the lecture powerpoint slides: Lecture 1: Introduction Lecture 2: Uninformed Search Lecture 3: Informed Search Lecture 4: CSPs I Lecture 5: CSPs II Lecture 6: Adversarial Search Lecture 7: Expectimax Search and Utilities Lecture 8: MDPs I Lecture 9: MDPs II Lecture 10: Reinforcement Learning I Lecture 11: Reinforcement Learning II Lecture 12: Probability Lecture 13: Markov Models Lecture 14: Hidden Markov Models Lecture 15: Particle Filters and Applications of HMMs Lecture 16: Bayes Nets I: Representation Lecture 17: Bayes Nets II: Independence Lecture 18: Bayes Nets III: Inference Lecture 19: Bayes Nets IV: Sampling Lecture 20: Decision Diagrams and VPI Lecture 21: Naive Bayes Lecture 22: Perceptron Lecture 23: Kernels and Clustering Lecture 24: Advanced Applications (NLP, Games, Cars) Lecture 25: Advanced Applications (Computer Vision and Robotics) Lecture 26: Conclusion The source files for all live in-lecture demos are being prepared from Berkeley AI for release Selected Research Papers Latest arxiv paper submissionson AI Peter Norvig-Teach Yourself Programming in Ten Years How to do Research At the MIT AI Lab A Roadmap towards Machine Intelligence Collaborative Filtering with Recurrent Neural Networks (2016) Wide & Deep Learning for Recommender Systems (2016) Deep Collaborative Filtering via Marginalized Denoising Auto-encoder (2015) Nonparametric bayesian multitask collaborative filtering (2013) Tensorflow: Large-scale machine learning on heterogeneous distributed systems https://infoscience.epfl.ch/record/82802/files/rr02-46.pdf Theano: A CPU and GPU math expression compiler. Caffe: Convolutional architecture for fast feature embedding Chainer: A powerful, flexible and intuitive framework of neural networks Large Scale Distributed Deep Networks Large-scale video classification with convolutional neural networks Efficient Estimation of Word Representations in Vector Space Grammar as a Foreign Language Going Deeper with Convolutions ON RECTIFIED LINEAR UNITS FOR SPEECH PROCESSING Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups. Multi-digit Number Recognition from Street View Imagery using Deep Convolutional Neural Networks google turning its lucrative web search over to AI machines Stanford Syllabus CS 20SI: Tensorflow for Deep Learning Research Crowd-Based Personalized Natural Language Explanations for Recommendations Comparative Study of Deep Learning Software Frameworks RedditML- What Are You Reading AI-Powered Social Bots(16 Jun 2017) The Many Tribes of Artificial Intelligence Source:https://medium.com/intuitionmachine/infographic-best-practices-in-training-deep-learning-networks-b8a3df1db53 The Deep Learning Roadmap Source:https://medium.com/intuitionmachine/the-deep-learning-roadmap-f0b4cac7009a Best Practices for Training Deep Learning Networks Source: https://medium.com/intuitionmachine/infographic-best-practices-in-training-deep-learning-networks-b8a3df1db53 ML/DL Cheatsheets Neural Network Architectures Source: http://www.asimovinstitute.org/neural-network-zoo/ Microsoft Azure Algorithm Flowchart Source: https://docs.microsoft.com/en-us/azure/machine-learning/machine-learning-algorithm-cheat-sheet SAS Algorithm Flowchart Source: http://blogs.sas.com/content/subconsciousmusings/2017/04/12/machine-learning-algorithm-use/ Algorithm Summary Source: http://machinelearningmastery.com/a-tour-of-machine-learning-algorithms/ Source: http://thinkbigdata.in/best-known-machine-learning-algorithms-infographic/ Algorithm Pro/Con Source: https://blog.dataiku.com/machine-learning-explained-algorithms-are-your-friend Python Algorithms Source: https://www.analyticsvidhya.com/blog/2015/09/full-cheatsheet-machine-learning-algorithms/ Python Basics Source: http://datasciencefree.com/python.pdf Source: https://www.datacamp.com/community/tutorials/python-data-science-cheat-sheet-basics#gs.0x1rxEA Numpy Source: https://www.dataquest.io/blog/numpy-cheat-sheet/ Source: http://datasciencefree.com/numpy.pdf Source: https://www.datacamp.com/community/blog/python-numpy-cheat-sheet#gs.Nw3V6CE Source: https://github.com/donnemartin/data-science-ipython-notebooks/blob/master/numpy/numpy.ipynb Pandas Source: http://datasciencefree.com/pandas.pdf Source: https://www.datacamp.com/community/blog/python-pandas-cheat-sheet#gs.S4P4T=U Source: https://github.com/donnemartin/data-science-ipython-notebooks/blob/master/pandas/pandas.ipynb Matplotlib Source: https://www.datacamp.com/community/blog/python-matplotlib-cheat-sheet Source: https://github.com/donnemartin/data-science-ipython-notebooks/blob/master/matplotlib/matplotlib.ipynb Scikit Learn Source: https://www.datacamp.com/community/blog/scikit-learn-cheat-sheet#gs.fZ2A1Jk Source: http://peekaboo-vision.blogspot.de/2013/01/machine-learning-cheat-sheet-for-scikit.html Source: https://github.com/rcompton/mlcheatsheet/blob/master/supervised_learning.ipynb Tensorflow Source: https://github.com/aymericdamien/TensorFlow-Examples/blob/master/notebooks/1Introduction/basicoperations.ipynb Pytorch Source: https://github.com/bfortuner/pytorch-cheatsheet Math Probability Source: http://www.wzchen.com/s/probability_cheatsheet.pdf Linear Algebra Source: https://minireference.com/static/tutorials/linearalgebrain4pages.pdf Statistics Source: http://web.mit.edu/~csvoss/Public/usabo/stats_handout.pdf Calculus Source: http://tutorial.math.lamar.edu/getfile.aspx?file=B,41,N

10 Must-Try AI Tools For Your Business (2025)
youtube
LLM Vibe Score0.368
Human Vibe Score0.48
Hostinger AcademyNov 7, 2024

10 Must-Try AI Tools For Your Business (2025)

Unlock the power of AI with these 10 must-try tools that can transform your business in 2024! 👉 https://bit.ly/4ffsvUV 💥 Use the discount code WB10 for 10% OFF! Whether you're looking to boost productivity, automate tasks, or improve decision-making, this video covers the top AI tools that will give your business a competitive edge. Watch to find out how each tool works and which one suits your business needs the best! 📌 Handy Links 10 Powerful Prompts to Elevate Your ChatGPT Experience https://www.youtube.com/watch?v=W7aPrXlVRO8 BEST AI Content Generation Tools for Content Creators (2024) https://www.youtube.com/watch?v=bUB5JEJEvI8&t=347s Join the Hostinger Referral Program: https://www.hostinger.com/referral-program Join the Hostinger Affiliate Program: https://www.hostinger.com/affiliates Join our team at Hostinger: https://www.hostinger.com/career ⭐Follow Us⭐ TikTok: https://www.tiktok.com/@hostingeracademy Instagram: https://www.instagram.com/hostingeracademy/ Twitter: https://twitter.com/Hostinger Facebook: https://www.facebook.com/Hostinger/ Reddit: https://www.reddit.com/r/Hostinger Hostinger Tutorials: https://www.hostinger.com/tutorials Subscribe to our channel: https://www.youtube.com/c/HostingerAcademy/?sub_confirmation=1 🕒 Timestamps 00:00 - Introduction 00:50 - Presentations.ai 01:51 - CoralAI 02:37 - Hostinger Website Builder 04:28 - Zapier 05:22 - Do not pay (AI lawyer) 06:08 - Adobe Firefly 07:05 - Twain 07:46 - ChatGPT 09:05 - ocean.io 09:58 - Autopod _ 🚀10 Must-Try AI Tools For Your Business (2024) Discover the top AI tools that can transform the way you work, boost productivity, and streamline your business operations! 📌 Why These Tools Matter These tools can help you: 👉 Automate time-consuming tasks 👉 Enhance your creative output 👉 Streamline communication and operations 👉 Improve business decision-making 📌 Featured AI Tools Here’s a quick look at the must-try AI tools for 2024: 👉 Gamma – Create stunning presentations with ease. 👉 ChatPDF – Ask questions and get insights from PDFs. 👉 Hostinger Website Builder – Build websites with AI-powered ease. 👉 Zapier – Automate your workflows seamlessly. 👉 Do Not Pay (AI Lawyer) – Get legal advice with AI support. 👉 Adobe Firefly – Create incredible visuals using generative AI. 👉 Twain – AI-driven insights for business decisions. 👉 ChatGPT – Revolutionize how you interact with AI chat. 👉 Ocean.io – Discover leads and grow your business. 👉 Autopod – Automate your podcast editing effortlessly. Watch the full video to dive deeper into how these AI tools can elevate your business in 2024! _ ▶ Want to see more awesome tutorials like this in the future? Consider subscribing 😁 https://www.youtube.com/c/HostingerAcademy/?sub_confirmation=1 Thank you for watching! Let us know in the comments below if you have any questions. Good luck on your online journey. 🚀 #AItoolsforbusiness #AIwebsiteBuilder #HostingerAcademy

AI
github
LLM Vibe Score0.358
Human Vibe Score0.006489749001329033
MatousMarikOct 31, 2024

AI

AI I This repository contains practical tasks for the Artificial Intelligence 1 course, that is based on book by Russel and Norvig Artificial Intellignece: A Modern Approach, 4th Edition. Tasks are designed to review AI algorithms and use them to play games. Requirements All assignments will be written in python. Task were created for python 3.9 however there should not be any problems with backward compatibility. You can solve all assignments while working exclusively with python standard library, however for game visualizations you will need to install modul pygame. For installation you can use pip: python3 -m pip install -U pygame --user If you need more detailed, platform-specific instructions you can visit pygame-GettingStarted. Assignments In total there will be 5 programming assignments whose solutions will be submitted via ReCodEx. In each of them you will write an AI agent that plays suitable games for corresponding lecture topic. Moreover there will by partial assignments, in which you will need to implement algorithms, that will allow you to implement suitable agent functions, however your agent implementation can use any approach you like. | Game | Suggested Approach | | ---- | ------ | | Dino | rule-based agent | | Pac-Man | uniform-cost search | | Sokoban | A* with custom heuristics | | Cell Wars | minimax or Monte Carlo tree search | | Minesweeper | backtracking search for CSPs | Note that information provided in the early assignments is omitted in later ones.

n8n Masterclass: Build AI Agents & Automate Workflows (Beginner to Pro)
youtube
LLM Vibe Score0.396
Human Vibe Score0.64
Nate Herk | AI AutomationOct 20, 2024

n8n Masterclass: Build AI Agents & Automate Workflows (Beginner to Pro)

JOIN THE FREE SKOOL COMMUNITY👇 https://www.skool.com/ai-automation-society-3440/about 🌟 Join my paid Skool community if you want to go deeper with n8n and AI Automations👇 https://www.skool.com/ai-automation-society-plus/about 🚧 Start Building with n8n! (I get kickback if you sign up here - thank you!) https://n8n.partnerlinks.io/22crlu8afq5r 💻 Book A Call If You're Interested in Implementing AI Agents Into Your Business: https://truehorizon.ai/ Welcome to the ultimate n8n masterclass! Whether you're a complete beginner or have little coding experience, this video will guide you step-by-step through everything you need to know to start automating workflows and building powerful AI agents with n8n. In this video, you'll learn: ⚙️ The basics of n8n, building your first workflow, and connecting with 300+ integrations. 🌐 How to use APIs and HTTP requests in n8n. 🧠 Harnessing the power of RAG (Retrieval-Augmented Generation) and vector databases for AI-powered automation. 🛠️ Creating custom tools and integrating them into workflows to build smarter AI agents. 🔗 Advanced concepts like webhooks, error handling, and scaling workflows for real-world automation. 📈 Best practices to keep your workflows optimized, scalable, and resilient. By the end, you’ll have the confidence to create your own AI agent automations, trigger workflows with webhooks, use APIs, and more! 💡 If you found this video helpful, don’t forget to like, comment, and subscribe for more content on n8n, AI agents, and automation. Let me know in the comments what you plan to automate next! Business Inquiries: 📧 nateherk@uppitai.com WATCH NEXT: https://youtu.be/JUx2ZfNfD64 TIMESTAMPS 00:00 What is n8n? 02:50 Why Should You Learn n8n? 04:53 Part 1: Getting Started 05:09 Self-Hosted vs Cloud 08:25 Workflows, Nodes, Executions 09:45 n8n Interface 16:05 Part 2: Core Concepts 16:28 Types of Nodes 19:00 Building Example Workflow 36:28 Part 3: RAG and Vector Databases 36:55 What is RAG? 38:23 What are Vector Databases? 44:07 Building RAG AI Agent 1:01:56 Part 4: Expanding Agents 1:02:31 n8n Workflows as Tools 1:05:23 Showcasing Agent Examples 1:10:20 Part 5: APIs & HTTP Requests 1:11:33 What is an API? 1:12:49 What is an HTTP Request? 1:13:14 How They Work Together 1:15:04 HTTP Request Examples in n8n 1:21:42 Part 6: The Final Part 1:22:24 Error Workflows 1:26:20 Best Practices 1:28:30 Next Steps Gear I Used: Camera: Razer Kiyo Pro Microphone: HyperX SoloCast Background Music: https://www.youtube.com/watch?v=Q7HjxOAU5Kc&t=0s Don't forget to like, subscribe, and hit the notification bell to stay updated with my latest videos on AI agents and automations!

FORGET ChatGPT, This AI TOOL is a GAMECHANGER 🔥
youtube
LLM Vibe Score0.299
Human Vibe Score0.32
Ishan SharmaOct 19, 2024

FORGET ChatGPT, This AI TOOL is a GAMECHANGER 🔥

Ishan Sharma: FORGET ChatGPT, This AI TOOL is a GAMECHANGER 🔥 Google just dropped NotebookLM and it is changing everything. I was using ChatGPT so far for research and learning, but NotebookLM has stolen the charm. NotebookLM lets you convert PDFs, YouTube videos, or Websites into Audio Podcasts This is a 2 person conversational podcast about the topic And Trust me, it sounds too good to be AI generated. It's powered by Google's latest Gemini 1.5 model. But that’s not all! You can add multiple sources in a notebook And also get a summary, table of contents, study guide And MORE in seconds! THIS is a game changer for learners, researchers, and creators! Helping you skyrocket your productivity. It's also great for school students and college students to learn anything faster. Try it out on notebooklm.google.com 📸 Instagram: https://bit.ly/ishansharma7390ig Join MarkitUpX Discord Server: https://discord.gg/fwSpTje4rh Timestamps 😁 About Me: https://bit.ly/aboutishansharma 📱 Twitter: https://bit.ly/ishansharma7390twt 📝 LinkedIn: https://bit.ly/ishansharma7390li 🌟 Please leave a LIKE ❤️ and SUBSCRIBE for more AMAZING content! 🌟 3 Books You Should Read 📈Psychology of Money: https://amzn.to/30wx4bW 👀Subtle Art of Not Giving a F: https://amzn.to/30zwWbP 💼Rework: https://amzn.to/3ALsAuz Tech I use every day 💻MacBook Air M1: https://amzn.to/2YWKPjG 📺LG 29' Ultrawide Monitor: https://amzn.to/3aG0p5p 🎥Sony ZV1: https://amzn.to/3ANqgDb 🎙Blue Yeti Mic: https://amzn.to/2YYbiNN ⽴Tripod Stand: https://amzn.to/3mVUiQc 🔅Ring Light: https://amzn.to/2YQlzLJ 🎧Marshall Major II Headphone: https://amzn.to/3lLhTDQ 🖱Logitech mouse: https://amzn.to/3p8edOC 💺Green Soul Chair: https://amzn.to/3mWIxZP ✨ Tags ✨ ishan sharma,FORGET ChatGPT This AI TOOL is a GAMECHANGER,chatgpt,gpt-4o,chatgpt 4o,gpt4o,openai,gpt 4,openai sora,microsoft openai,artificial intelligence,ai,chatbot,gpt-4,chatgpt-4,new gpt ai model,chatgpt vision,chatgpt chatbot,chatgpt4o,new ai,chat gpt,chatgpt 4,gpt update,chat gpt 4o,google notebooklm,google notebook,google notebook app,google notebooklm tool,google keep,google ai,google app ai,google notes app,google notes ai ✨ Hashtags ✨ #google #chatgpt #ai

Coding a FULL App with AI (You Won't Believe This)
youtube
LLM Vibe Score0.38
Human Vibe Score0.89
Creator MagicSep 30, 2024

Coding a FULL App with AI (You Won't Believe This)

Want to build your own apps but don't know how to code? In this video, I show you how I built a fully functional AI powered YouTube comments app using only AI tools in just 3 days! This is a step by step guide that covers everything from brainstorming app ideas and creating a roadmap, to generating code and designing a beautiful user interface. 🧞 Sign up for the Comment Genie Beta: https://mrc.fm/cgbeta ✨ Weekly AI Newsletter: https://mrc.fm/creatormagic You can get $100 free credit for Linode to host your no code app. Use the link here for Linode here: https://mrc.fm/linode We'll be using these awesome AI tools: ● ChatGPT: https://mrc.fm/chatgpt For brainstorming ideas, creating a development roadmap and generating code. ● Cursor: https://mrc.fm/cursor This AI powered code generator will do the heavy lifting and write most of the code for us. ● Replit: https://mrc.fm/replit We'll use Replit to host our code in the cloud and quickly test our app online. ● v0: https://mrc.fm/v0 This AI powered design tool helps you create beautiful and responsive user interfaces without any coding. ● Midjourney: https://mrc.fm/midjourney We'll use Midjourney (or your favourite AI art generator) to quickly create a stunning logo for our app. I also share some bonus tips and tricks to help you get the most out of AI powered app development. Let me know in the comments what you're building with AI! Here are the time-stamped chapters in the requested format: 0:00 Introduction 0:25 Brainstorming with AI using ChatGPT 1:49 OpenAI ChatGPT o1 Preview for tech stack 2:55 Using Replit for cloud based coding 3:18 Introducing Cursor Composer for AI assisted coding 5:50 Testing out our AI developed app 6:48 Using v0 for frontend graphic design 8:35 Creating a logo with Midjourney 9:14 List of no code AI tools for developing apps 9:58 Tips for optimal AI assisted coding 11:49 Deploying the app with Linode 12:46 Demo of the Comment Genie app 13:12 Responding to feedback from beta testers 14:12 Conclusion 12:35 Demonstrating the Comment Genie app 13:24 Implementing user feedback 14:44 Conclusion and call for viewer feedback

How To Start A Business Using Only AI
youtube
LLM Vibe Score0.362
Human Vibe Score0.56
Learn With ShopifySep 2, 2024

How To Start A Business Using Only AI

How to Use AI to Start a Business in 2024. ► Shopify Free Trial https://utm.io/uhpKC ► YouTube takes on TikTok Shop with expanded Shopify partnership https://youtube.com/shorts/XdzbDOak9BI?si=eNUZL8AgZK6f0XJg Unlock Your Entrepreneurial Potential with AI! Ever dreamed of starting a business but felt overwhelmed by the complexity? AI is here to revolutionize the way we work! In this video, we'll guide you through the exciting process of launching your own venture using artificial intelligence. Discover how to: Identify profitable niche ideas using AI-powered market research tools Create compelling content with AI-driven writing assistants Design stunning visuals effortlessly using AI design platforms Build and manage your online store without technical expertise Expand your reach by easily adding your products to social media networks like Instagram, YouTube, and TikTok Whether you're a seasoned entrepreneur or just starting out, this video will equip you with the knowledge and tools to turn your business dreams into reality. Get ready to harness the power of AI and embark on a successful entrepreneurial journey! –––––––––––––––––––––––––––––––––––––––––––– Watch More Learn with Shopify Video Tutorials: ► How to Connect Your Shopify Store To Your YouTube Channel https://youtu.be/ymD5M8w-drk?si=tLt52iNd0VKrL5eW ► YouTube Shopping Tutorial: The Best Way To Sell Your Shopify Products on YouTube LIVE https://youtu.be/AUtEP7LTNeg?si=imvS2pUTsLvhcZmT ► How To Create Beautiful Presentations With AI https://youtu.be/BZ_ObFC7NVA ► What is Shopify Magic and Shopify Sidekick? (And How To Use It) https://youtu.be/Y7Rlr5gxPp4 ► Prompt Engineering Tutorial Part 1: An Introduction to AI Prompting https://youtu.be/zBaa8Ct2C-k?si=ZshSj72IdgpGrAN5 ► Prompt Engineering Tutorial Part 2: Text-to-Text https://youtu.be/ZlQHPt86h6s ► Prompt Engineering Tutorial Part 3: Text-to-Image https://youtu.be/6RAStep_3OI ► Prompt Engineering Tutorial Part 4: Text to Video https://youtu.be/QgjL0fNTwHc ► How to Sell on Instagram https://youtu.be/cqmUWuA2w2U –––––––––––––––––––––––––––––––––––––––––––– 🔔 Subscribe to @learnwithshopify for more productivity tutorials and tips for entrepreneurs of all stages. Here's what we'll cover in this video: 0:00 Intro 0:48 Idea generation using AI 2:20 How to market research using AI 3:14 Naming your business using AI 4:44 AI Logo Generator 6:10 AI Product Creation 9:48 How to upload products to your website 12:15 How to list your items on Instagram 13:06 How to list your items on YouTube 13:40 How to list your items on TikTok 14:04 Marketing using AI 15:30 Legalization –––––––––––––––––––––––––––––––––––––––––––– 📈 Related Videos: 20 Mobile AI Apps https://youtu.be/OSAFKU8FL44 TikTok Marketing Tutorial (Organic Strategy) https://youtu.be/SeWNUUEtZOY TikTok Marketing Tutorial (Paid Ads) https://youtu.be/RIy9ZN3B5CA Reddit for Business Tutorial https://youtu.be/FcYtZg1uGMA LinkedIn Ads Tutorial https://youtu.be/WMKldiJ8mEw 🔗 Useful Resources: ► 64 Best Small Business Ideas To Start in 2024 https://utm.io/uhpKB ► Free Shopify Business Course https://utm.io/uhpKE ► Join our Shopify community https://utm.io/uhpKC –––––––––––––––––––––––––––––––––––––––––––– -- TOOLS & RESOURCES – ► Sign Up To Shopify Today https://utm.io/uhpKC ► Shopify Masters Podcast @shopifymasters ► Shopify Podcast https://utm.io/uhlvZ ► 10 Amazing AI Tools For Your Business In 2024 https://youtu.be/TKAO1ykK994 ► 10 ChatGPT Tips & Tricks https://youtu.be/88tVeKj0-7k ► How to make money with Instagram Reels https://youtu.be/U831lmASZRY ► The OFFICIAL Shopify Tutorial - The COMPLETE GUIDE https://youtu.be/ferhOYx1NMo –––––––––––––––––––––––––––––––––––––––––––– #Shopify #aitools #businesscoaching #businessideas

Airtable builds with Amazon Bedrock to transform workflows with generative AI | Amazon Web Services
youtube
LLM Vibe Score0.273
Human Vibe Score0.17
Amazon Web ServicesMar 20, 2024

Airtable builds with Amazon Bedrock to transform workflows with generative AI | Amazon Web Services

Airtable, a cloud based low-code platform, enables non-programmers to build next-gen business applications. To democratize AI adoption for non-technical users across organizations, Airtable launched Airtable AI, powered by Amazon Bedrock. Through this partnership, Airtable AI seamlessly incorporates powerful foundation models like Anthropic's Claude and Amazon's Titan on Amazon Bedrock, allowing customers to choose models that best suits their use cases and workflows. Key benefits include a unified API for integrating AWS services, secure hosting of foundation models and data, access to cutting-edge technologies, fostering bottoms-up AI adoption among non-technical teams, and generative AI use cases including content generation, automation actions, and intelligent data Q&A. All this is unified within Airtable's intuitive low-code environment. Learn more at: https://go.aws/3Ta68X4 Subscribe: More AWS videos: https://go.aws/3m5yEMW More AWS events videos: https://go.aws/3ZHq4BK Do you have technical AWS questions? Ask the community of experts on AWS re:Post: https://go.aws/3lPaoPb ABOUT AWS Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform, offering over 200 fully featured services from data centers globally. Millions of customers — including the fastest-growing startups, largest enterprises, and leading government agencies — are using AWS to lower costs, become more agile, and innovate faster. #AmazonBedrock #FoundationModels #generativeAI #AnthropicClaude #AmazonTitan #Airtable #AWS #AmazonWebServices #CloudComputing

10 Best AI Business Ideas 2024
youtube
LLM Vibe Score0.408
Human Vibe Score0.48
AI UncoveredMar 3, 2024

10 Best AI Business Ideas 2024

10 Best AI Business Ideas 2024 🔒 Keep Your Digital Life Private and Be Safe Online: https://nordvpn.com/safetyfirst Are you curious about the future of business in the exciting realm of Artificial Intelligence (AI)? Look no further! In this captivating video, we unveil the top 10 AI business ideas that are set to revolutionize the entrepreneurial landscape in 2024. From cutting-edge technology to innovative solutions, we delve into the most promising ventures that harness the power of AI to drive success and growth. Discover how AI is reshaping traditional business models and opening up endless possibilities for aspiring entrepreneurs. Whether you're a seasoned professional or a budding visionary, these handpicked AI business ideas offer a gateway to prosperity in the ever-evolving digital age. Join us as we explore groundbreaking concepts that blend creativity with computational intelligence, paving the way for unprecedented innovation and profitability. From automated customer service to personalized marketing strategies, AI is poised to transform every aspect of modern business operations. Dive deep into the realm of AI-powered startups and witness firsthand how these groundbreaking ideas are shaping the future of commerce. With our expert insights and comprehensive analysis, you'll gain invaluable knowledge to embark on your own AI-driven entrepreneurial journey. Don't miss out on the opportunity to stay ahead of the curve and capitalize on the transformative potential of AI in business. Join us as we unveil the 10 best AI business ideas for 2024 and embark on a journey towards success in the dynamic world of artificial intelligence. Subscribe now and stay tuned for more cutting-edge content that empowers you to thrive in the digital economy of tomorrow! #ai #artificialintelligence #aibusiness Subscribe for more! Welcome to AI Uncovered, your ultimate destination for exploring the fascinating world of artificial intelligence! Our channel delves deep into the latest AI trends and technology, providing insights into cutting-edge AI tools, AI news, and breakthroughs in artificial general intelligence (AGI). We simplify complex concepts, making AI explained in a way that is accessible to everyone. At AI Uncovered, we're passionate about uncovering the most captivating stories in AI, including the marvels of ChatGPT and advancements by organizations like OpenAI. Our content spans a wide range of topics, from science news and AI innovations to in-depth discussions on the ethical implications of artificial intelligence. Our mission is to enlighten, inspire, and inform our audience about the rapidly evolving technology landscape. Whether you're a tech enthusiast, a professional seeking to stay ahead of AI trends, or someone curious about the future of artificial intelligence, AI Uncovered is the perfect place to expand your knowledge. Join us as we uncover the secrets behind AI tools and their potential to revolutionize our world. Subscribe to AI Uncovered and stay tuned for enlightening content that bridges the gap between AI novices and experts, covering AI news, AGI, ChatGPT, OpenAI, artificial intelligence, and more. Together, let's explore the limitless possibilities of technology and AI. Disclaimer: Some links included in this description might be affiliate links. If you purchase a product or service through the links that we provide, we may receive a small commission. There is no additional charge for you. Thank you for supporting AI Uncovered so we can continue to provide you with free, high-quality content. _ 🌟 Contact: ai.uncovered.ai@gmail.com

BEST FIGMA AI TOOLS for UI/UX Designers 2024⚡️| Saptarshi Prakash #shorts
youtube
LLM Vibe Score0.303
Human Vibe Score0.34
Saptarshi PrakashNov 1, 2023

BEST FIGMA AI TOOLS for UI/UX Designers 2024⚡️| Saptarshi Prakash #shorts

AI will definitely replace UI/UX Designers who are not using these Free Figma AI Plugins in their Designs: Magestic | AI-powered icons and illustrations: https://www.figma.com/community/plugin/1148175024770495469/magestic-ai-generated-icon-sets Wireframe Designer | AI-Powered Wireframes: https://www.figma.com/community/plugin/1228969298040149016/wireframe-designer FigGPT | ChatGPT-powered plugin for website copies: https://www.figma.com/community/plugin/1207913933994957698/figgpt Make sure to leave a LIKE, and SUBSCRIBE for more Figma Plugins & UI/UX Design Tips! Share your thoughts in the comments below! 📩 Join my community: https://nas.io/sapta Join my Instagram broadcast channel to never miss an update: https://ig.me/j/AbadG67M--mvwepf/ Get on a call with me: https://topmate.io/sapta Buy me a coffee: https://www.buymeacoffee.com/saptarshipr 😃 ABOUT ME This is Saptarshi (a.k.a. Sapta), an engineer turned self-taught Product Designer based out of Bangalore, India. I have worked with some of the very well known startups of India and learned anything and everything that is needed to create amazing experiences for the users. I'm also an active speaker, teacher and community builder, and have delivered over 60 talks, workshops and webinars on design. In this channel, I post videos with tips, strategies, tutorials and general gyaan to scale your career in Design. If you are into it, you may want to subscribe and hit the bell icon to that you don't miss out :) 💻 📷 🎤 MY GEAR My Desk: https://bengaluru.featherlitestore.com/product/motorized-height-adjustable-table/ Sony A7iv: https://amzn.to/3KQZ0LM (Primary camera) Samyang 24-70mm F2.8 lens: https://amzn.to/3qDYHx0 Sony a6300: https://amzn.to/3gIx0v1 (Secondary Camera) Sigma 16mm F1.4 lens: https://amzn.to/38DFPRR Sony 50mm F1.8 lens: https://amzn.to/3rufcaB Samson G-Track Pro condenser mic: https://amzn.to/37Rixsw Rode Wireless Go 2 : https://amzn.to/3KQXBU0 Boya Lavalier Mic: https://amzn.to/2M0MZI7 Godox SL60w light : https://amzn.to/3HgSU3O Godox SB-UE 80cm softbox : https://amzn.to/3GdNq8h DIGITEK DTR 500 BH (60 Inch) Tripod: https://amzn.to/39d1m48 📲 SOCIALS Instagram: https://www.instagram.com/saptarshiux/ Twitter: https://twitter.com/saptarshipr Dribbble: https://dribbble.com/saptarshipr LinkedIn: https://www.linkedin.com/in/saptarshipr/ Medium: https://medium.com/@saptarshipr 🎶 MUSIC The jingles and the background score is composed by Sargam Prakash, an awesome designer and musician. Do check out his channel. Sargam Prakash: https://www.youtube.com/user/sargampr 🌟 TAGS BEST FIGMA AI TOOLS for UI/UX Designers,figma plugins,figma ai plugins,figma tools,figma ai tools,figma ui design,figma design plugins,figma update,figma ai,figma,ui ux design,figma design,ui ux designer 2023,ui/ux design,ux design,user experience design,ui/ux design india,figma tutorial,figma tutorial for beginners,ux,ui,design,ui design,ui ux,uiux,ai tools,chatgpt,openai,ui ux design tutorial for beginners,sapta,saptarshipr,saptarshi,prakash,swiggy 🌟 HASHTAGS #uiux #design #graphicdesign

Workflow Automation with AI and Zapier | CXOTalk #808
youtube
LLM Vibe Score0.388
Human Vibe Score0.37
CXOTalkOct 23, 2023

Workflow Automation with AI and Zapier | CXOTalk #808

#zapier #workflowautomation #workflow #aiautomation The rising significance of enterprise AI presents a unique hurdle: seamlessly integrating AI-based business workflows into operational systems, especially for non-programmers. On CXOTalk episode 808, we explore these issues with Mike Knoop, co-founder of Zapier and the company's AI lead. The conversation with Mike covers the rationale behind integrating AI, the technological advancements AI brings to workflow automation solutions, and its broader impact on business agility. Join the CXOTalk community: www.cxotalk.com/subscribe Read the full transcript: https://www.cxotalk.com/episode/ai-workflows-in-business-a-practical-guide Key points in the discussion include: ► The potential of AI-powered automation to empower more business users with customized workflows. But governance, accuracy, and security are key challenges to consider when implementing AI workflows. ► Initial use cases include generating creative ideas, summarizing unstructured data, and making powerful business process automations easier to build for non-technical users. ► Customer service and marketing are excellent starting points for AI automation. Watch this conversation to gain practical advice on using low-code, no-code tools to automate AI in the enterprise. Mike Knoop is the co-founder and Head of Zapier AI at Zapier. Mike has a B.S. in mechanical engineering from the University of Missouri, where his research topic was focused on finite element modeling and optimization. Michael Krigsman is an industry analyst and publisher of CXOTalk. For three decades, he has advised enterprise technology companies on market messaging and positioning strategy. He has written over 1,000 blogs on leadership and digital transformation and created almost 1,000 video interviews with the world’s top business leaders on these topics. His work has been referenced in the media over 1,000 times and in over 50 books. He has presented and moderated panels at numerous industry events around the world.

The Massive Opportunity in Building AI Businesses | Alex Hormozi
youtube
LLM Vibe Score0.44
Human Vibe Score0.89
Liam OttleySep 24, 2023

The Massive Opportunity in Building AI Businesses | Alex Hormozi

📚 Join the #1 community for AI entrepreneurs and connect with 100,000+ members: https://bit.ly/3uRIRB3 📈 We help industry experts, entrepreneurs & developers build and scale their AI Agency: https://bit.ly/skoolmain 🤝 Need AI Solutions Built? Work with me: https://b.link/qv62vqy6 ⚒️ Build AI Agents Without Coding: https://agentivehub.com/ 🚀 Apply to Join My Team at Morningside AI: https://tally.so/r/wbYr52 Alex Hormozi and I sat down for a chat about how to start an AI business in 2023. Alex shared his advice for people wanting to start an AI Automation Agency, including how to sell emerging technology like AI, the importance of a good development team as an AI entrepreneur and his thoughts on AI businesses and startups that he's seeing at Acquisition.com. Alex Hormozi also gave his thoughts on the huge opportunity for entrepreneurs to help businesses integrate AI into their businesses or "AI-ify" them, including exactly how he'd go about it himself. This call was an exclusive interview for my AAA Accelerator members who were able to watch this conversation LIVE. Timestamps: 0:00 - Intro 0:58 - AI Automation Agencies Explained 5:13 - How To Sell Emerging Tech 6:23 - The Opportunity Of AI Automation 10:00 - Choosing The Right Business Partner 11:14 - What Problems To Solve In An Industry 12:06 - Where AAA Might Falter or Excel 14:32 - The Importance Of Development Resources 19:17 - How Important Is Downtime As An Entrepreneur?

Best AI Tools for Accountants
youtube
LLM Vibe Score0.36
Human Vibe Score0.21
Miles EducationJun 28, 2023

Best AI Tools for Accountants

We’re pretty sure, these great AI tools will help you in the long run. Let’s have a look at their importance: VIC.AI: AI-powered Accounting Made Effortless! Their advanced algorithms are trained on vast invoice data, eliminating the need for templates or memorization. Accurate from day one, their Autopilot technology seamlessly integrates AI for streamlined invoicing.😇 Indy: AI-Powered Accounting Made Fast and Affordable! Freelancers, businesses, and entrepreneurs can tackle accounting tasks up to 20x faster than traditional software. Create income statements and financial statements in a fraction of the time, all at a lower cost than traditional accountants.🤔 Docyt: AI-Powered Accounting Automation for Faster Decision Making. Digitize financial data, automate workflows, and make faster decisions. Reduce costs and simplify bookkeeping and back-office tasks.😲 Blue Dot is an innovative market leader with a cutting-edge financial platform. Their all-in-one Tax Compliance Platform combines digitization, tax compliance, and automation to analyze employee spend data for VAT, Taxable Employee Benefits, and Corporate Income Tax.🥹 Comment below the names of the AI tools that you use for accounting.👇👇 #aitoolsforaccountants #aitools #ai #technology #upskill #cpa #uscpa #CPAexam #cpajobs #CPAscope #MilesEducation #workintheus #talent #fyp #explore #accounting #accountants #accountancy #cpacoursereview #jobopportunities #CPAplacement #CPAsalary #success #career

The future of AI
youtube
LLM Vibe Score0.471
Human Vibe Score0.61
GaryVeeMay 9, 2023

The future of AI

When voice and ai hit scale … shits gonna get interesting… — Thanks for watching! Join My Discord!: https://www.garyvee.com/discord Check out another series on my channel: Keynotes: https://www.youtube.com/watch?v=6vCDlmhRmBo&list=PLfA33-E9P7FCEF1izpctGGoak841XYzrJ NFTs: https://www.youtube.com/watch?v=AwMJ6bScB2s&list=PLfA33-E9P7FAcvsVSFqzSuJhHu3SkW2Ma Business Meetings: https://www.youtube.com/watch?v=wILI_VV6z4Y&list=PLfA33-E9P7FCTIY62wkqZ-E1cwpc2hxBJ Gary Vaynerchuk Original Films: https://youtube.com/playlist?list=PLfA33-E9P7FAvnrOcgy4MvIcCXxoyjuku Trash Talk: https://youtube.com/playlist?list=PLfA33-E9P7FDelN4bXFgtJuczC9HHmm2- WeeklyVee: https://youtube.com/playlist?list=PLfA33-E9P7FBPjdQcF6uedz9fdk8XKn-b — Gary Vaynerchuk is a serial entrepreneur, and serves as the Chairman of VaynerX, the CEO of VaynerMedia and the Creator & CEO of VeeFriends. Gary is considered one of the leading global minds on what’s next in culture, relevance and the internet. Known as “GaryVee” he is described as one of the most forward thinkers in business – he acutely recognizes trends and patterns early to help others understand how these shifts impact markets and consumer behavior. Whether its emerging artists, esports, NFT investing or digital communications, Gary understands how to bring brand relevance to the forefront. He is a prolific angel investor with early investments in companies such as Facebook, Twitter, Tumblr, Venmo, Snapchat, Coinbase and Uber. Gary is an entrepreneur at heart — he builds businesses. Today, he helps Fortune 1000 brands leverage consumer attention through his full service advertising agency, VaynerMedia which has offices in NY, LA, London, Mexico City, LATAM and Singapore. VaynerMedia is part of the VaynerX holding company which also includes VaynerProductions, VaynerNFT, Gallery Media Group, The Sasha Group, Tracer, VaynerSpeakers, VaynerTalent, and VaynerCommerce. Gary is also the Co-Founder of VaynerSports, Resy and Empathy Wines. Gary guided both Resy and Empathy to successful exits — both were sold respectively to American Express and Constellation Brands. He’s also a Board Member at Candy Digital, Co-Founder of VCR Group, Co-Founder of ArtOfficial, and Creator & CEO of VeeFriends. Gary was recently named to the Fortune list of the Top 50 Influential people in the NFT industry. In addition to running multiple businesses, Gary documents his life daily as a CEO through his social media channels which has more than 34 million followers and garnishes over 272 million monthly impressions/views across all platforms. His podcast ‘The GaryVee Audio Experience’ ranks among the top podcasts globally. He is a five-time New York Times Best-Selling Author and one of the most highly sought after public speakers. Gary serves on the board of MikMak, Bojangles Restaurants, and Pencils of Promise. He is also a longtime Well Member of Charity:Water.