VibeBuilders.ai Logo
VibeBuilders.ai

Major

Explore resources related to major to help implement AI solutions for your business.

Technical founders - is "bulling" your way through learning right for a startup? [I will not promote]
reddit
LLM Vibe Score0
Human Vibe Score0
JustZed32This week

Technical founders - is "bulling" your way through learning right for a startup? [I will not promote]

Sup, This is a question for technical founders. \--a little backstory-- I am starting a company in AI field that creates something nobody has ever done before. 7 months in. \--- How most software companies are created - you have an improvement idea, then you have a thousand or so problems to solve to make that improvement happen, and for each one that you don't know, you go to Stackoverflow or ChatGPT to look for solutions for that problem. Which involves next-to-no upfront preparation because for vast majority of traditional software you can solve it on-the-go - "traditional" software is very easy compared to, say, mechanical, pharma or AI engineering. However, for more advanced disciplines - can you just "Google" it on-the-go? I'm a solo founder, and 8 months in, creating a foundational model, BECAUSE I did not know things upfront, I've wasted at least 3 months doing something which was mostly technically unviable in the first place. Out of 14000 lines of code that I've done (including tests), I had to scrap 10000 recently. Imagine the scale of it. Obviously I didn't even know how ML works when I've started. Major fuck-up. How do you operate in industries which you've done before? How do you determine that it's time to start creating you big technological leaps instead of continuing to learn? Cheers. Edit: No need to push me on business topics. I know how to create value very well. It's only a tech question, and I'm only asking because - well - to deliver my value, I need to do a lot of novel tech.

Anyone finding that they just don't NEED to add more Employees anymore? (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score0.6
wilschroterThis week

Anyone finding that they just don't NEED to add more Employees anymore? (I will not promote)

A friend of mine who was looking for work asked me if we were hiring and I responded "You know, it's weird but all of our growth goals don't seem to map back to hiring people anymore." This isn't about the economy or growth goals. It's a really fascinating shift in focus and costs for startups. My gut reaction is that I HATE the idea of not creating more jobs. In my career I've hired thousands of people, and I've always prided myself on job creation. We just sold a company that employed 200 people last year, and I'm proud of the work we were able to create. What's interesting is that I simply don't feel like we NEED to like we used to. As we're looking at all of our growth goals, for the first time I'm not assigning FTEs to them. Nearly everything we're doing is actually reducing the need for more humans, not adding them - and we're not even trying to reduce the need. Obviously the timing of AI has had a major impact. Product - Our team is shipping more code than ever before, and even our designers who have never touched code are shipping final code. If we doubled the size of the team, it would make no difference (this is a big deal considering the historical cost here). Marketing - So many aspects of our marketing are getting automated and streamlined, to the point where even a single FTE can create a massive amount of reach across channels. Support - Our Success team is able to effectively respond to tickets in a fraction of the time, which essentially doubles their capacity without adding any more staff. Management - With less staff we need less managers, which are a big expense, but it also means reporting and decisions are more streamlined, which is a positive. But it also means those positions simply don't get created like they used to. I think this is a big deal for the younger startups because it translates into needing less capital (or none!) which provides for more ownership and agency. Clearly we still need some folks to build out the core team, but that's very different than a massive staffing line item. Anyone else here finding the same trend? Opposite? I don't have a strong opinion either way, but I'd love to hear how other Founders are processing this.

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies) (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies) (I will not promote)

AI Palette is an AI-driven platform that helps food and beverage companies predict emerging product trends. I had the opportunity recently to sit down with the founder to get his advice on building an AI-first startup, which he'll be going through in this post. (I will not promote) About AI Palette: Co-founders: >!2 (Somsubhra GanChoudhuri, Himanshu Upreti)!!100+!!$12.7M USD!!AI-powered predictive analytics for the CPG (Consumer Packaged Goods) industry!!Signed first paying customer in the first year!!65+ global brands, including Cargill, Diageo, Ajinomoto, Symrise, Mondelez, and L’Oréal, use AI Palette!!Every new product launched has secured a paying client within months!!Expanded into Beauty & Personal Care (BPC), onboarding one of India’s largest BPC companies within weeks!!Launched multiple new product lines in the last two years, creating a unified suite for brand innovation!Identify the pain points in your industry for ideas* When I was working in the flavour and fragrance industry, I noticed a major issue CPG companies faced: launching a product took at least one to two years. For instance, if a company decided today to launch a new juice, it wouldn’t hit the market until 2027. This long timeline made it difficult to stay relevant and on top of trends. Another big problem I noticed was that companies relied heavily on market research to determine what products to launch. While this might work for current consumer preferences, it was highly inefficient since the product wouldn’t actually reach the market for several years. By the time the product launched, the consumer trends had already shifted, making that research outdated. That’s where AI can play a crucial role. Instead of looking at what consumers like today, we realised that companies should use AI to predict what they will want next. This allows businesses to create products that are ahead of the curve. Right now, the failure rate for new product launches is alarmingly high, with 8 out of 10 products failing. By leveraging AI, companies can avoid wasting resources on products that won’t succeed, leading to better, more successful launches. Start by talking to as many industry experts as possible to identify the real problems When we first had the idea for AI Palette, it was just a hunch, a gut feeling—we had no idea whether people would actually pay for it. To validate the idea, we reached out to as many people as we could within the industry. Since our focus area was all about consumer insights, we spoke to professionals in the CPG sector, particularly those in the insights departments of CPG companies. Through these early conversations, we began to see a common pattern emerge and identified the exact problem we wanted to solve. Don’t tell people what you’re building—listen to their frustrations and challenges first. Going into these early customer conversations, our goal was to listen and understand their challenges without telling them what we were trying to build. This is crucial as it ensures that you can gather as much data about the problem to truly understand it and that you aren't biasing their answers by showing your solution. This process helped us in two key ways: First, it validated that there was a real problem in the industry through the number of people who spoke about experiencing the same problem. Second, it allowed us to understand the exact scale and depth of the problem—e.g., how much money companies were spending on consumer research, what kind of tools they were currently using, etc. Narrow down your focus to a small, actionable area to solve initially. Once we were certain that there was a clear problem worth solving, we didn’t try to tackle everything at once. As a small team of two people, we started by focusing on a specific area of the problem—something big enough to matter but small enough for us to handle. Then, we approached customers with a potential solution and asked them for feedback. We learnt that our solution seemed promising, but we wanted to validate it further. If customers are willing to pay you for the solution, it’s a strong validation signal for market demand. One of our early customer interviewees even asked us to deliver the solution, which we did manually at first. We used machine learning models to analyse the data and presented the results in a slide deck. They paid us for the work, which was a critical moment. It meant we had something with real potential, and we had customers willing to pay us before we had even built the full product. This was the key validation that we needed. By the time we were ready to build the product, we had already gathered crucial insights from our early customers. We understood the specific information they wanted and how they wanted the results to be presented. This input was invaluable in shaping the development of our final product. Building & Product Development Start with a simple concept/design to validate with customers before building When we realised the problem and solution, we began by designing the product, but not by jumping straight into coding. Instead, we created wireframes and user interfaces using tools like InVision and Figma. This allowed us to visually represent the product without the need for backend or frontend development at first. The goal was to showcase how the product would look and feel, helping potential customers understand its value before we even started building. We showed these designs to potential customers and asked for feedback. Would they want to buy this product? Would they pay for it? We didn’t dive into actual development until we found a customer willing to pay a significant amount for the solution. This approach helped us ensure we were on the right track and didn’t waste time or resources building something customers didn’t actually want. Deliver your solution using a manual consulting approach before developing an automated product Initially, we solved problems for customers in a more "consulting" manner, delivering insights manually. Recall how I mentioned that when one of our early customer interviewees asked us to deliver the solution, we initially did it manually by using machine learning models to analyse the data and presenting the results to them in a slide deck. This works for the initial stages of validating your solution, as you don't want to invest too much time into building a full-blown MVP before understanding the exact features and functionalities that your users want. However, after confirming that customers were willing to pay for what we provided, we moved forward with actual product development. This shift from a manual service to product development was key to scaling in a sustainable manner, as our building was guided by real-world feedback and insights rather than intuition. Let ongoing customer feedback drive iteration and the product roadmap Once we built the first version of the product, it was basic, solving only one problem. But as we worked closely with customers, they requested additional features and functionalities to make it more useful. As a result, we continued to evolve the product to handle more complex use cases, gradually developing new modules based on customer feedback. Product development is a continuous process. Our early customers pushed us to expand features and modules, from solving just 20% of their problems to tackling 50–60% of their needs. These demands shaped our product roadmap and guided the development of new features, ultimately resulting in a more complete solution. Revenue and user numbers are key metrics for assessing product-market fit. However, critical mass varies across industries Product-market fit (PMF) can often be gauged by looking at the size of your revenue and the number of customers you're serving. Once you've reached a certain critical mass of customers, you can usually tell that you're starting to hit product-market fit. However, this critical mass varies by industry and the type of customers you're targeting. For example, if you're building an app for a broad consumer market, you may need thousands of users. But for enterprise software, product-market fit may be reached with just a few dozen key customers. Compare customer engagement and retention with other available solutions on the market for product-market fit Revenue and the number of customers alone isn't always enough to determine if you're reaching product-market fit. The type of customer and the use case for your product also matter. The level of engagement with your product—how much time users are spending on the platform—is also an important metric to track. The more time they spend, the more likely it is that your product is meeting a crucial need. Another way to evaluate product-market fit is by assessing retention, i.e whether users are returning to your platform and relying on it consistently, as compared to other solutions available. That's another key indication that your solution is gaining traction in the market. Business Model & Monetisation Prioritise scalability Initially, we started with a consulting-type model where we tailor-made specific solutions for each customer use-case we encountered and delivered the CPG insights manually, but we soon realized that this wasn't scalable. The problem with consulting is that you need to do the same work repeatedly for every new project, which requires a large team to handle the workload. That is not how you sustain a high-growth startup. To solve this, we focused on building a product that would address the most common problems faced by our customers. Once built, this product could be sold to thousands of customers without significant overheads, making the business scalable. With this in mind, we decided on a SaaS (Software as a Service) business model. The benefit of SaaS is that once you create the software, you can sell it to many customers without adding extra overhead. This results in a business with higher margins, where the same product can serve many customers simultaneously, making it much more efficient than the consulting model. Adopt a predictable, simplistic business model for efficiency. Look to industry practices for guidance When it came to monetisation, we considered the needs of our CPG customers, who I knew from experience were already accustomed to paying annual subscriptions for sales databases and other software services. We decided to adopt the same model and charge our customers an annual upfront fee. This model worked well for our target market, aligning with industry standards and ensuring stable, recurring revenue. Moreover, our target CPG customers were already used to this business model and didn't have to choose from a huge variety of payment options, making closing sales a straightforward and efficient process. Marketing & Sales Educate the market to position yourself as a thought leader When we started, AI was not widely understood, especially in the CPG industry. We had to create awareness around both AI and its potential value. Our strategy focused on educating potential users and customers about AI, its relevance, and why they should invest in it. This education was crucial to the success of our marketing efforts. To establish credibility, we adopted a thought leadership approach. We wrote blogs on the importance of AI and how it could solve problems for CPG companies. We also participated in events and conferences to demonstrate our expertise in applying AI to the industry. This helped us build our brand and reputation as leaders in the AI space for CPG, and word-of-mouth spread as customers recognized us as the go-to company for AI solutions. It’s tempting for startups to offer products for free in the hopes of gaining early traction with customers, but this approach doesn't work in the long run. Free offerings don’t establish the value of your product, and customers may not take them seriously. You should always charge for pilots, even if the fee is minimal, to ensure that the customer is serious about potentially working with you, and that they are committed and engaged with the product. Pilots/POCs/Demos should aim to give a "flavour" of what you can deliver A paid pilot/POC trial also gives you the opportunity to provide a “flavour” of what your product can deliver, helping to build confidence and trust with the client. It allows customers to experience a detailed preview of what your product can do, which builds anticipation and desire for the full functionality. During this phase, ensure your product is built to give them a taste of the value you can provide, which sets the stage for a broader, more impactful adoption down the line. Fundraising & Financial Management Leverage PR to generate inbound interest from VCs When it comes to fundraising, our approach was fairly traditional—we reached out to VCs and used connections from existing investors to make introductions. However, looking back, one thing that really helped us build momentum during our fundraising process was getting featured in Tech in Asia. This wasn’t planned; it just so happened that Tech in Asia was doing a series on AI startups in Southeast Asia and they reached out to us for an article. During the interview, they asked if we were fundraising, and we mentioned that we were. As a result, several VCs we hadn’t yet contacted reached out to us. This inbound interest was incredibly valuable, and we found it far more effective than our outbound efforts. So, if you can, try to generate some PR attention—it can help create inbound interest from VCs, and that interest is typically much stronger and more promising than any outbound strategies because they've gone out of their way to reach out to you. Be well-prepared and deliberate about fundraising. Keep trying and don't lose heart When pitching to VCs, it’s crucial to be thoroughly prepared, as you typically only get one shot at making an impression. If you mess up, it’s unlikely they’ll give you a second chance. You need to have key metrics at your fingertips, especially if you're running a SaaS company. Be ready to answer questions like: What’s your retention rate? What are your projections for the year? How much will you close? What’s your average contract value? These numbers should be at the top of your mind. Additionally, fundraising should be treated as a structured process, not something you do on the side while juggling other tasks. When you start, create a clear plan: identify 20 VCs to reach out to each week. By planning ahead, you’ll maintain momentum and speed up the process. Fundraising can be exhausting and disheartening, especially when you face multiple rejections. Remember, you just need one investor to say yes to make it all worthwhile. When using funds, prioritise profitability and grow only when necessary. Don't rely on funding to survive. In the past, the common advice for startups was to raise money, burn through it quickly, and use it to boost revenue numbers, even if that meant operating at a loss. The idea was that profitability wasn’t the main focus, and the goal was to show rapid growth for the next funding round. However, times have changed, especially with the shift from “funding summer” to “funding winter.” My advice now is to aim for profitability as soon as possible and grow only when it's truly needed. For example, it’s tempting to hire a large team when you have substantial funds in the bank, but ask yourself: Do you really need 10 new hires, or could you get by with just four? Growing too quickly can lead to unnecessary expenses, so focus on reaching profitability as soon as possible, rather than just inflating your team or burn rate. The key takeaway is to spend your funds wisely and only when absolutely necessary to reach profitability. You want to avoid becoming dependent on future VC investments to keep your company afloat. Instead, prioritize reaching break-even as quickly as you can, so you're not reliant on external funding to survive in the long run. Team-Building & Leadership Look for complementary skill sets in co-founders When choosing a co-founder, it’s important to find someone with a complementary skill set, not just someone you’re close to. For example, I come from a business and commercial background, so I needed someone with technical expertise. That’s when I found my co-founder, Himanshu, who had experience in machine learning and AI. He was a great match because his technical knowledge complemented my business skills, and together we formed a strong team. It might seem natural to choose your best friend as your co-founder, but this can often lead to conflict. Chances are, you and your best friend share similar interests, skills, and backgrounds, which doesn’t bring diversity to the table. If both of you come from the same industry or have the same strengths, you may end up butting heads on how things should be done. Having diverse skill sets helps avoid this and fosters a more collaborative working relationship. Himanshu (left) and Somsubhra (right) co-founded AI Palette in 2018 Define roles clearly to prevent co-founder conflict To avoid conflict, it’s essential that your roles as co-founders are clearly defined from the beginning. If your co-founder and you have distinct responsibilities, there is no room for overlap or disagreement. This ensures that both of you can work without stepping on each other's toes, and there’s mutual respect for each other’s expertise. This is another reason as to why it helps to have a co-founder with a complementary skillset to yours. Not only is having similar industry backgrounds and skillsets not particularly useful when building out your startup, it's also more likely to lead to conflicts since you both have similar subject expertise. On the other hand, if your co-founder is an expert in something that you're not, you're less likely to argue with them about their decisions regarding that aspect of the business and vice versa when it comes to your decisions. Look for employees who are driven by your mission, not salary For early-stage startups, the first hires are crucial. These employees need to be highly motivated and excited about the mission. Since the salary will likely be low and the work demanding, they must be driven by something beyond just the paycheck. The right employees are the swash-buckling pirates and romantics, i.e those who are genuinely passionate about the startup’s vision and want to be part of something impactful beyond material gains. When employees are motivated by the mission, they are more likely to stick around and help take the startup to greater heights. A litmus test for hiring: Would you be excited to work with them on a Sunday? One of the most important rounds in the hiring process is the culture fit round. This is where you assess whether a candidate shares the same values as you and your team. A key question to ask yourself is: "Would I be excited to work with this person on a Sunday?" If there’s any doubt about your answer, it’s likely not a good fit. The idea is that you want employees who align with the company's culture and values and who you would enjoy collaborating with even outside of regular work hours. How we structure the team at AI Palette We have three broad functions in our organization. The first two are the big ones: Technical Team – This is the core of our product and technology. This team is responsible for product development and incorporating customer feedback into improving the technology Commercial Team – This includes sales, marketing, customer service, account managers, and so on, handling everything related to business growth and customer relations. General and Administrative Team – This smaller team supports functions like finance, HR, and administration. As with almost all businesses, we have teams that address the two core tasks of building (technical team) and selling (commercial team), but given the size we're at now, having the administrative team helps smoothen operations. Set broad goals but let your teams decide on execution What I've done is recruit highly skilled people who don't need me to micromanage them on a day-to-day basis. They're experts in their roles, and as Steve Jobs said, when you hire the right person, you don't have to tell them what to do—they understand the purpose and tell you what to do. So, my job as the CEO is to set the broader goals for them, review the plans they have to achieve those goals, and periodically check in on progress. For example, if our broad goal is to meet a certain revenue target, I break it down across teams: For the sales team, I’ll look at how they plan to hit that target—how many customers they need to sell to, how many salespeople they need, and what tactics and strategies they plan to use. For the technical team, I’ll evaluate our product offerings—whether they think we need to build new products to attract more customers, and whether they think it's scalable for the number of customers we plan to serve. This way, the entire organization's tasks are cascaded in alignment with our overarching goals, with me setting the direction and leaving the details of execution to the skilled team members that I hire.

Behind the scene : fundraising pre-seed of an AI startup
reddit
LLM Vibe Score0
Human Vibe Score1
Consistent-Wafer7325This week

Behind the scene : fundraising pre-seed of an AI startup

A bit of feedback from our journey at our AI startup. We started prototyping stuff around agentic AI last winter with very cool underlying tech research based on some academic papers (I can send you links if you're interested in LLM orchestration). I'm a serial entrepreneur with 2x exits, nothing went fancy but enough to keep going into the next topic. This time, running an AI project has been a bit different and unique due to the huge interest around the topic. Here are a few insights. Jan \~ Mar: Research Nothing was serious, just a side project with a friend on weekends (the guy became our lead SWE). Market was promising and we had the convinction that our tech can be game changer in computer systems workflows. March \~ April: Market Waking Up Devin published their pre-seed $20m fundraising led by Founders Fund; they paved the market with legitimacy. I decided to launch some coffee meetings with a few angels in my network. Interest confirmed. Back to work on some more serious early prototyping; hard work started here. April \~ May: YC S24 (Fail) Pumped up by our prospective angels and the market waking up on the agentic topic, I applied to YC as a solo founder (was still looking for funds and co-founders). Eventually got rejected (no co-founder and not US-based). May \~ July: VC Dance (Momentum 1) Almost randomly at the same time we got rejected from YC, I got introduced to key members of the VC community by one of our prospective angels. Interest went crazy... tons of calls. Brace yourself here, we probably met 30\~40 funds (+ angels). Got strong interests from 4\~5 of them (3 to 5 meetings each), ultimately closed 1 and some interests which might convert later in the next stage. The legend of AI being hype is true. Majority of our calls went only by word of mouth, lots of inbounds, people even not having the deck would book us a call in the next 48h after saying hi. Also lots of "tourists," just looking because of AI but with no strong opinion on the subject to move further. The hearsay about 90% rejection is true. You'll have a lot of nos, ending some days exhausted and unmotivated. End July: Closing, the Hard Part The VC roadshow is kind of an art you need to master. You need to keep momentum high enough and looking over-subscribed. Good pre-seed VC deals are over-competitive, and good funds only focus on them; they will have opportunities to catch up on lost chances at the seed stage later. We succeeded (arduously) to close our 18\~24mo budget with 1 VC, a few angels, and some state-guaranteed debt. Cash in bank just on time for payday in August (don't under-estimate time of processing) Now: Launching and Prepping the Seed Round We're now in our first weeks of go-to-market with a lot of uncertainty but a very ambitious plan ahead. The good part of having met TONS of VCs during the pre-seed roadshow is that we met probably our future lead investors in these. What would look like a loss of time in the initial pre-seed VC meetings has been finally very prolific, helping us to refine our strategy, assessing more in-depth the market (investors have a lot of insights, they meet a lot of people... that's their full-time job). We now have clear milestones and are heading to raise our seed round by end of year/Q1 if stars stay aligned :) Don't give up, the show must go on.

Building in the open with Founder University - I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
Tim-SylvesterThis week

Building in the open with Founder University - I will not promote

Published Oct 30, 2024 I am on my fifth startup. I ran the last one for a decade, that’s a whole story. A hell of a story. But a different story. I’ll tell it to you when I can, but not right now. The one before that was an e-commerce site that did pretty well but I didn’t love it. Before that were two service businesses. The first one I did for the love of the game, the second one was an attempt to make people stop asking me to fix their computer by charging them outrageous prices, which backfired horribly when they were eager to pay. None are relevant except to say I’ve been around the block and have the scars to prove it. When it was time to get back out there, I wanted to use all I’ve learned to do better. Before I talk about what those lessons produced, I’m going to talk about what those lessons were. Cause before effect, after all. One thing I wanted to do better this time was pattern matching - making the startup look the way that the industry and investors “expect” a startup to look. My last startup was an awesome idea with awesome tech (still is, but like I said, another story), but that one didn’t match patterns. It didn’t match investor patterns, industry buying patterns, patterns of existing, immediate, recognized and admitted needs. Because it didn’t “look” right to anyone, everything about it was way harder than necessary. The “make it look right” approach runs the risk of building a cargo cult, imitating the trappings of something but without understanding the essence of that something, but then again, a thing that looks like a knife is going to make a better knife that a thing that looks like a bowling ball, so sometimes just sharing apparent similarities can get you pretty far, even if it doesn’t get you all the way there. Like how mimicking someone’s accent makes it easier for them to understand you. For this one, I wanted to adopt every tool, method, and pattern that I knew “the industry” wanted to see to minimize the friction from development, go-to-market, scaling, adoption, and that would make investment optional (and, therefore, available if desired) instead of necessary (and, therefore, largely unavailable). That required establishing some expectations for successful patterns I could match against. What patterns am I matching to? Here’s a general sketch of my pattern matching thought process: Software first and software only. It’s the easiest industry to start a business in, lowest startup costs, and easiest customer acquisition. I wanted to build software for an element of the industry that’s actively emerging (and therefore has room to grow) and part of an optimistic investor thesis (and therefore has a cohort of people who are intent on injecting capital into the market to help it grow). It needs to fills a niche that is underexplored (low competition) and highly potent (lots of opportunity), while being aligned to recognized and emerging needs within the industry (readily adopted). I wanted it to have evidence supporting the business thesis that proves the demand exists, but demonstrates that the demand is unanswered (as of yet) by sufficient or adequate supply.* I wanted the lowest number of dominoes to line up and tip for everything to work correctly - the more dominoes in the line, the less likely the last one will fall. I wanted to implement modern toolsets for everything, wherever possible. I wanted to obey the maxim, “When there’s a gold rush, don’t mine the gold, sell the picks and shovels.” Whatever I chose would need to produce cash flow almost immediately with minimal development time or go-to-market delays, because the end of ZIRP killed the “trust me bro” investment thesis predominant over the last 15 years. I wanted to match to YC best practices, not because YC can predict what will definitely work, but because they’ve churned through so many startups in the last 15 years that they have a good sense of what will definitely not work. And I wanted to build client-centric, because if my intent is to to produce cash flow immediately, we need to get clients immediately, and if we need to get clients immediately, we need to focus on what clients need right now. Extra credit: What’s the difference between a customer and a client? Note: Competition is awesome! Competition is validating and not scary, because competition proves a market exists. But competition, especially mature competition against an immature startup, makes it harder to break into a space. A first mover advantage isn’t everything, but seeing demand before it’s sufficiently supplied is a great advantage if you’re capital constrained or otherwise unproven. Think about how much money the first guy to sell fidget spinners or Silly Bandz made versus how much money the last guy to order a pallet of each made. Finding demand that exists already but is as of yet insufficiently satisfied is a great place to start. What opportunity spaces are most relevant? The industries and markets I chose to observe were: AI, because if I’m following a theme & pattern for today, it’s AI. Fintech, because cash is king, and fintech puts your hands on cash flow. Crypto/blockchain, because that’s the “new” fintech (or maybe the “old-new” fintech?), and crypto creates powerful incentives and capital formation strategies, along with a lot of flexibility for transaction systems. Tools, particularly unmet demand in tools, that enable these industries. If you wanted to do some brief and simple homework, you could map each of those bullets to several of the numbered list items preceding them. The reasoning was pretty simplistic - AI is what people want to build and invest in now, while fintech and crypto/blockchain are what people were building and investing in for the last major investment thesis. That means that there’s demand in the market for AI and AI-adjacent startups, while there’s a glut of underutilized and highly developed tools within fintech and crypto/blockchain, with a lot of motivated capital behind the adoption. When someone is thinking “I built this thing and not enough people are using it”, and you then build something that uses it creates a great way to find allies. This rationale harnesses technology that is being built and financed now (which means it needs tools and support methods, and a lot of other “picks and shovels”), while leveraging technology that was recently built and financed and is eager for more widespread adoption of the existing toolkits, which makes it suitable for using to build the AI-adjacent tools that are in demand now. It’s like two harmonics producing constructive interference - it makes two waves into one larger wave, which gives me more momentum to surf against. This was a learning process, and I iterated against my general paradigm repeatedly as I learned more. Neither of us have the patience to go through that in excruciating detail, so I’ll cover the highlights in my next post. Extra credit answer: A customer gets a product, a client gets a service. Challenge: Is software a product or a service?

Looking for a tech cofounder. Revoltionary (yes really!) gig economy app. I will not promote.
reddit
LLM Vibe Score0
Human Vibe Score1
sweetpea___This week

Looking for a tech cofounder. Revoltionary (yes really!) gig economy app. I will not promote.

Hey everyone! I’m building a new gig-work app that cuts out the hassles of interviews, applications, and sky-high fees. We’re aiming to make it easy for businesses to hire qualified freelancers for short shifts or one-off tasks—and for freelancers to set their own rates and get paid quickly. Why This App? Time-Saving Model: Instead of posting jobs and conducting multiple interviews, employers can instantly book from a list of KYC-verified freelancers who showcase their skills via 30-second video bios. Cost Leadership: We plan to charge only 5%, far below the 15–50% common in other gig platforms. This keeps more money in the pockets of both freelancers and businesses. Proven Demand: A beta test in 2018 drew nearly 600 active users, validating that there’s appetite for a simpler, fairer way to fill short shifts. About Me 20+ years’ experience in payroll, workforce management, and operations for Fortune 500 companies. Led cross-functional teams, implemented large-scale solutions, and believe in building with a user-first mindset. Offering meaningful equity—I want a true partner, not a hired gun. Who I’m Looking For Full-Stack Developer (comfortable with Node.js, React, Python, or similar and ML/Ai) who can manage everything from front-end to database integration (ideally Postgres/MySQL) and build a same day payments system. Passion for creating solutions that genuinely help gig workers and small businesses. Excitement to collaborate on the product roadmap, from the booking interface to same-day payment features. The Opportunity Major Market: The gig economy is huge and still growing. If we nail speed, cost-effectiveness, and ease of use, we can capture a significant share of it. Remote-Friendly: We can work together from anywhere, though I’m planning to relaunch in London where the initial beta gained momentum. If this sounds like your kind of challenge, drop a comment or DM me. Let’s chat about how we can merge our strengths—my operations background and your technical expertise—to build a platform that truly transforms the gig-work experience. Thanks for reading, and I look forward to creating something impactful together!

Lessons from 139 YC AI startups (S23)
reddit
LLM Vibe Score0
Human Vibe Score0.333
minophenThis week

Lessons from 139 YC AI startups (S23)

YC's Demo Day was last week, and with it comes another deluge of AI companies. A record-breaking 139 startups were in some way related to AI or ML - up from 112 in the last batch. Here are 5 of my biggest takeaways: AI is (still) eating the world. It's remarkable how diverse the industries are - over two dozen verticals were represented, from materials science to social media to security. However, the top four categories were: AI Ops: Tooling and platforms to help companies deploy working AI models. We'll discuss more below, but AI Ops has become a huge category, primarily focused on LLMs and taming them for production use cases. Developer Tools: Apps, plugins, and SDKs making it easier to write code. There were plenty of examples of integrating third-party data, auto-generating code/tests, and working with agents/chatbots to build and debug code. Healthcare + Biotech: It seems like healthcare has a lot of room for automation, with companies working on note-taking, billing, training, and prescribing. And on the biotech side, there are some seriously cool companies building autonomous surgery robots and at-home cancer detection. Finance + Payments: Startups targeting banks, fintechs, and compliance departments. This was a wide range of companies, from automated collections to AI due diligence to "Copilot for bankers." Those four areas covered over half of the startups. The first two make sense: YC has always filtered for technical founders, and many are using AI to do what they know - improve the software developer workflow. But it's interesting to see healthcare and finance not far behind. Previously, I wrote: Large enterprises, healthcare, and government are not going to send sensitive data to OpenAI. This leaves a gap for startups to build on-premise, compliant \[LLMs\] for these verticals. And we're now seeing exactly that - LLMs focused on healthcare and finance and AI Ops companies targeting on-prem use cases. It also helps that one of the major selling points of generative AI right now is cost-cutting - an enticing use case for healthcare and finance. Copilots are king. In the last batch, a lot of startups positioned themselves as "ChatGPT for X," with a consumer focus. It seems the current trend, though, is "Copilot for X" - B2B AI assistants to help you do everything from KYC checks to corporate event planning to chip design to negotiate contracts. Nearly two dozen companies were working on some sort of artificial companion for businesses - and a couple for consumers. It's more evidence for the argument that AI will not outright replace workers - instead, existing workers will collaborate with AI to be more productive. And as AI becomes more mainstream, this trend of making specialized tools for specific industries or tasks will only grow. That being said - a Bing-style AI that lives in a sidebar and is only accessible via chat probably isn't the most useful form factor for AI. But until OpenAI, Microsoft, and Google change their approach (or until another company steps up), we'll probably see many more Copilots. AI Ops is becoming a key sector. "AI Ops" has been a term for only a few years. "LLM Ops" has existed for barely a year. And yet, so many companies are focused on training, fine-tuning, deploying, hosting, and post-processing LLMs it's quickly becoming a critical piece of the AI space. It's a vast industry that's sprung up seemingly overnight, and it was pretty interesting to see some of the problems being solved at the bleeding edge. For example: Adding context to language models with as few as ten samples. Pausing and moving training runs in real-time. Managing training data ownership and permissions. Faster vector databases. Fine-tuning models with synthetic data. But as much ~~hype~~ enthusiasm and opportunity as there might be, the size of the AI Ops space also shows how much work is needed to really productionalize LLMs and other models. There are still many open questions about reliability, privacy, observability, usability, and safety when it comes to using LLMs in the wild. Who owns the model? Does it matter? Nine months ago, anyone building an LLM company was doing one of three things: Training their own model from scratch. Fine-tuning a version of GPT-3. Building a wrapper around ChatGPT. Thanks to Meta, the open-source community, and the legions of competitors trying to catch up to OpenAI, there are now dozens of ways to integrate LLMs. However, I found it interesting how few B2B companies mentioned whether or not they trained their own model. If I had to guess, I'd say many are using ChatGPT or a fine-tuned version of Llama 2. But it raises an interesting question - if the AI provides value, does it matter if it's "just" ChatGPT behind the scenes? And once ChatGPT becomes fine-tuneable, when (if ever) will startups decide to ditch OpenAI and use their own model instead? "AI" isn't a silver bullet. At the end of the day, perhaps the biggest lesson is that "AI" isn't a magical cure-all - you still need to build a defensible company. At the beginning of the post-ChatGPT hype wave, it seemed like you just had to say "we're adding AI" to raise your next round or boost your stock price. But competition is extremely fierce. Even within this batch, there were multiple companies with nearly identical pitches, including: Solving customer support tickets. Negotiating sales contracts. Writing drafts of legal documents. Building no-code LLM workflows. On-prem LLM deployment. Automating trust and safety moderation. As it turns out, AI can be a competitive advantage, but it can't make up for a bad business. The most interesting (and likely valuable) companies are the ones that take boring industries and find non-obvious use cases for AI. In those cases, the key is having a team that can effectively distribute a product to users, with or without AI. Where we’re headed I'll be honest - 139 companies is a lot. In reviewing them all, there were points where it just felt completely overwhelming. But after taking a step back, seeing them all together paints an incredibly vivid picture of the current AI landscape: one that is diverse, rapidly evolving, and increasingly integrated into professional and personal tasks. These startups aren't just building AI for the sake of technology or academic research, but are trying to address real-world problems. Technology is always a double-edged sword - and some of the startups felt a little too dystopian for my taste - but I'm still hopeful about AI's ability to improve productivity and the human experience.

Why raise in 2025? - I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
Able_Swimming_4909This week

Why raise in 2025? - I will not promote

I will not promote Lately, I've been thinking about how AI tools are completely reshaping what it means to bootstrap a startup. It honestly feels like we're living through a golden age for entrepreneurs where you don't necessarily need venture capital to build something big or meaningful. At my company, we're a small team of just four people, bootstrapping our AI-focused startup. Thanks to AI-powered tools, we're able to keep our burn rate ridiculously low, quickly test new ideas, and scale our operations way faster than we ever expected. It’s honestly pretty incredible how accessible advanced technology has become, even compared to just a few years ago. Of course, bootstrapping definitely comes with its own share of headaches. For example, we've noticed that funded startups get significantly better access to cloud credits, advertising budgets, and enterprise-level tools. We do have access to some discounts and free resources, but it rarely compares to what funded startups enjoy. This can feel frustrating, especially when you know you're competing directly with businesses that have those extra advantages. Visibility is another major challenge we've noticed. Without big funding announcements or a well-connected investor backing us, getting attention from media or even early adopters can be tough. It's just harder to make a splash without someone else's endorsement. We've had to accept and work around creatively. That said, there's something genuinely empowering about staying bootstrapped, prioritizing profitability, and maintaining control over our vision. After speaking with several investors, we've become aware of how investors can significantly influence or even redirect the trajectory of a business. We've heard stories where investors gained enough leverage to replace the original founders or have killed perfectly profitable businesses that were not growing "fast enough", which certainly gave us pause. They can definitely be helpful but giving the control over the future of my business to someone else would definitely make me feel anxious. At this time, we simply don't feel raising external capital aligns with our current goals, but we're also aware that this could change in the future. For now, maintaining autonomy and staying close to our original vision remains a priority. I'm curious to hear from others here who've been through this. Have you successfully bootstrapped an AI a tech business? What obstacles did you encounter, and how did you overcome them? EDIT: To give you a bit of perspective, my company is a B2B SaaS in the finance industry based in Europe. We have received VC funding in the past but it was an exceptionally good deal and we don't plan to raise in the near future even-thought it may change if we see the need to help us scale. We have also raised a significant amount in soft funding. Right now, we are growing on our revenues, and we plan to continue this trajectory. Recently, one of our developers left, and although we are a small team, we noticed that it had little to no impact on our productivity.

Behind the scene : fundraising pre-seed of an AI startup
reddit
LLM Vibe Score0
Human Vibe Score1
Consistent-Wafer7325This week

Behind the scene : fundraising pre-seed of an AI startup

A bit of feedback from our journey at our AI startup. We started prototyping stuff around agentic AI last winter with very cool underlying tech research based on some academic papers (I can send you links if you're interested in LLM orchestration). I'm a serial entrepreneur with 2x exits, nothing went fancy but enough to keep going into the next topic. This time, running an AI project has been a bit different and unique due to the huge interest around the topic. Here are a few insights. Jan \~ Mar: Research Nothing was serious, just a side project with a friend on weekends (the guy became our lead SWE). Market was promising and we had the convinction that our tech can be game changer in computer systems workflows. March \~ April: Market Waking Up Devin published their pre-seed $20m fundraising led by Founders Fund; they paved the market with legitimacy. I decided to launch some coffee meetings with a few angels in my network. Interest confirmed. Back to work on some more serious early prototyping; hard work started here. April \~ May: YC S24 (Fail) Pumped up by our prospective angels and the market waking up on the agentic topic, I applied to YC as a solo founder (was still looking for funds and co-founders). Eventually got rejected (no co-founder and not US-based). May \~ July: VC Dance (Momentum 1) Almost randomly at the same time we got rejected from YC, I got introduced to key members of the VC community by one of our prospective angels. Interest went crazy... tons of calls. Brace yourself here, we probably met 30\~40 funds (+ angels). Got strong interests from 4\~5 of them (3 to 5 meetings each), ultimately closed 1 and some interests which might convert later in the next stage. The legend of AI being hype is true. Majority of our calls went only by word of mouth, lots of inbounds, people even not having the deck would book us a call in the next 48h after saying hi. Also lots of "tourists," just looking because of AI but with no strong opinion on the subject to move further. The hearsay about 90% rejection is true. You'll have a lot of nos, ending some days exhausted and unmotivated. End July: Closing, the Hard Part The VC roadshow is kind of an art you need to master. You need to keep momentum high enough and looking over-subscribed. Good pre-seed VC deals are over-competitive, and good funds only focus on them; they will have opportunities to catch up on lost chances at the seed stage later. We succeeded (arduously) to close our 18\~24mo budget with 1 VC, a few angels, and some state-guaranteed debt. Cash in bank just on time for payday in August (don't under-estimate time of processing) Now: Launching and Prepping the Seed Round We're now in our first weeks of go-to-market with a lot of uncertainty but a very ambitious plan ahead. The good part of having met TONS of VCs during the pre-seed roadshow is that we met probably our future lead investors in these. What would look like a loss of time in the initial pre-seed VC meetings has been finally very prolific, helping us to refine our strategy, assessing more in-depth the market (investors have a lot of insights, they meet a lot of people... that's their full-time job). We now have clear milestones and are heading to raise our seed round by end of year/Q1 if stars stay aligned :) Don't give up, the show must go on.

From "There's an App for that" to "There's YOUR App for that" - AI workflows will transform generic apps into deeply personalized experiences
reddit
LLM Vibe Score0
Human Vibe Score1
Important-Ostrich69This week

From "There's an App for that" to "There's YOUR App for that" - AI workflows will transform generic apps into deeply personalized experiences

I will not promote. For the past decade mobile apps were a core element of daily life for entertainment, productivity and connectivity. However, as the ecosystem saturated the general desire to download "just one more app" became apprehensive. There were clear monopolistic winners in different categories, such as Instagram and TikTok, which completely captured the majority of people's screentime. The golden age of creating indie apps and becoming a millionaire from them was dead. Conceptual models of these popular apps became ingrained in the general consciousness, and downloading new apps where re-learning new UI layouts was required, became a major friction point. There is high reluctance to download a new app rather than just utilizing the tooling of the growing market share of the existing winners. Content marketing and white labeled apps saw a resurgence of new app downloads, as users with parasympathetic relationships with influencers could be more easily persuaded to download them. However, this has led to a series of genericized tooling that lacks the soul of the early indie developer apps from the 2010s (Flappy bird comes to mind). A seemingly grim spot to be in, until everything changed on November 30th 2022. Sam Altman, Ilya Sutskever and team announced chatGPT, a Large Language Model that was the first publicly available generative AI tool. The first non-deterministic tool that could reason probablisitically in a similar (if flawed) way, to the human mind. At first, it was a clear paradigm shift in the world of computing, this was obvious from the fact that it climbed to 1 Million users within the first 5 days of its launch. However, despite the insane hype around the AI, its utility was constrained to chatbot interfaces for another year or more. As the models reasoning abilities got better and better, engineers began to look for other ways of utilizing this new paradigm shift, beyond chatbots. It became clear that, despite the powerful abilities to generate responses to prompts, the LLMs suffered from false hallucinations with extreme confidence, significantly impacting the reliability of their use, in search, coding and general utility. Retrieval Augmented Generation (RAG) was coined to provide a solution to this. Now, the LLM would apply a traditional search for data, via a database, a browser or other source of truth, and then feed that information into the prompt as it generates, allowing for more accurate results. Furthermore, it became clear that you could enhance an LLM by providing them metadata to interact with tools such as APIs for other services, allowing LLMs to perform actions typically reserved for humans, like fetching data, manipulating it and acting as an independent Agent. This prompted engineers to start treating LLMs, not as a database and a search engine, but rather a reasoning system, that could be part of a larger system of inputs and feedback to handle workflows independently. These "AI Agents" are poised to become the core technology in the next few years for hyper-personalizing and automating processes for specific users. Rather than having a generic B2B SaaS product that is somewhat useful for a team, one could standup a modular system of Agents that can handle the exactly specified workflow for that team. Frameworks such as LlangChain and LLamaIndex will help enable this for companies worldwide. The power is back in the hands of the people. However, it's not just big tech that is going to benefit from this revolution. AI Agentic workflows will allow for a resurgence in personalized applications that work like personal digital employee's. One could have a Personal Finance agent keeping track of their budgets, a Personal Trainer accountability coaching you making sure you meet your goals, or even a silly companion that roasts you when you're procrastinating. The options are endless ! At the core of this technology is the fact that these agents will be able to recall all of your previous data and actions, so they will get better at understanding you and your needs as a function of time. We are at the beginning of an exciting period in history, and I'm looking forward to this new period of deeply personalized experiences. What are your thoughts ? Let me know in the comments !

I am selling my tool which converts websites into android and iOS apps within 5 minutes.
reddit
LLM Vibe Score0
Human Vibe Score1
Latter-Row-5719This week

I am selling my tool which converts websites into android and iOS apps within 5 minutes.

Hi, my name is Toshit Garg. I started working on SaaS products around April 2023. The plan was simple: to create tools that help entrepreneurs easily grow their businesses. My first tool was "Convertixo", inspired by my work as a Fiverr seller where I converted websites into apps for clients, earning around $1,000 per month. I thought, why not automate this process? Following Convertixo, I created a few other tools like "Web to PWA". At one point, I developed an AI-based tool called "AppMintAI" , a productized service named "Engage Enhance", and even a WordPress plugin that lets users create pragmatic pages for SEO and a boilerplates. Unfortunately, none of these tools gained significant traction. I would launch them on Product Hunt, get a few users, and then nothing. Other than Convertixo, all my other tools only received a handful of free users. I believe this happened because I’m not very passionate about marketing. So, I decided to pivot and focus on content creation, which is where my true passion lies. Currently, I’m selling all my products one by one. As for Convertixo, it now has 800 users, a $20 MRR, and an email subscriber list of 100+. It was also the third Product of the Day on Product Hunt in January of this year. While the product has gained some traction, I’ve realized my focus is on content creation. However, with the right marketing and drive, I believe Convertixo has great potential to grow. If you’re interested in taking Convertixo to the next level, let’s chat! Here are some key statistics: In the last 20 days, Convertixo has received 4.9K impressions from Google and 338 visitors. More about the product: Convertixo can convert any website into Android and iOS apps using a custom webview. The apps are generated in Android Studio and Xcode. You receive both the APK and the source code for the Android app, along with the source code for the iOS app. The converted apps require no maintenance, and they update exactly like the website. A major benefit is the ability to add push notifications via OneSignal for free, allowing you to re-target your customers at no cost. Feel free to ask if you have any questions!

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies) (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies) (I will not promote)

AI Palette is an AI-driven platform that helps food and beverage companies predict emerging product trends. I had the opportunity recently to sit down with the founder to get his advice on building an AI-first startup, which he'll be going through in this post. (I will not promote) About AI Palette: Co-founders: >!2 (Somsubhra GanChoudhuri, Himanshu Upreti)!!100+!!$12.7M USD!!AI-powered predictive analytics for the CPG (Consumer Packaged Goods) industry!!Signed first paying customer in the first year!!65+ global brands, including Cargill, Diageo, Ajinomoto, Symrise, Mondelez, and L’Oréal, use AI Palette!!Every new product launched has secured a paying client within months!!Expanded into Beauty & Personal Care (BPC), onboarding one of India’s largest BPC companies within weeks!!Launched multiple new product lines in the last two years, creating a unified suite for brand innovation!Identify the pain points in your industry for ideas* When I was working in the flavour and fragrance industry, I noticed a major issue CPG companies faced: launching a product took at least one to two years. For instance, if a company decided today to launch a new juice, it wouldn’t hit the market until 2027. This long timeline made it difficult to stay relevant and on top of trends. Another big problem I noticed was that companies relied heavily on market research to determine what products to launch. While this might work for current consumer preferences, it was highly inefficient since the product wouldn’t actually reach the market for several years. By the time the product launched, the consumer trends had already shifted, making that research outdated. That’s where AI can play a crucial role. Instead of looking at what consumers like today, we realised that companies should use AI to predict what they will want next. This allows businesses to create products that are ahead of the curve. Right now, the failure rate for new product launches is alarmingly high, with 8 out of 10 products failing. By leveraging AI, companies can avoid wasting resources on products that won’t succeed, leading to better, more successful launches. Start by talking to as many industry experts as possible to identify the real problems When we first had the idea for AI Palette, it was just a hunch, a gut feeling—we had no idea whether people would actually pay for it. To validate the idea, we reached out to as many people as we could within the industry. Since our focus area was all about consumer insights, we spoke to professionals in the CPG sector, particularly those in the insights departments of CPG companies. Through these early conversations, we began to see a common pattern emerge and identified the exact problem we wanted to solve. Don’t tell people what you’re building—listen to their frustrations and challenges first. Going into these early customer conversations, our goal was to listen and understand their challenges without telling them what we were trying to build. This is crucial as it ensures that you can gather as much data about the problem to truly understand it and that you aren't biasing their answers by showing your solution. This process helped us in two key ways: First, it validated that there was a real problem in the industry through the number of people who spoke about experiencing the same problem. Second, it allowed us to understand the exact scale and depth of the problem—e.g., how much money companies were spending on consumer research, what kind of tools they were currently using, etc. Narrow down your focus to a small, actionable area to solve initially. Once we were certain that there was a clear problem worth solving, we didn’t try to tackle everything at once. As a small team of two people, we started by focusing on a specific area of the problem—something big enough to matter but small enough for us to handle. Then, we approached customers with a potential solution and asked them for feedback. We learnt that our solution seemed promising, but we wanted to validate it further. If customers are willing to pay you for the solution, it’s a strong validation signal for market demand. One of our early customer interviewees even asked us to deliver the solution, which we did manually at first. We used machine learning models to analyse the data and presented the results in a slide deck. They paid us for the work, which was a critical moment. It meant we had something with real potential, and we had customers willing to pay us before we had even built the full product. This was the key validation that we needed. By the time we were ready to build the product, we had already gathered crucial insights from our early customers. We understood the specific information they wanted and how they wanted the results to be presented. This input was invaluable in shaping the development of our final product. Building & Product Development Start with a simple concept/design to validate with customers before building When we realised the problem and solution, we began by designing the product, but not by jumping straight into coding. Instead, we created wireframes and user interfaces using tools like InVision and Figma. This allowed us to visually represent the product without the need for backend or frontend development at first. The goal was to showcase how the product would look and feel, helping potential customers understand its value before we even started building. We showed these designs to potential customers and asked for feedback. Would they want to buy this product? Would they pay for it? We didn’t dive into actual development until we found a customer willing to pay a significant amount for the solution. This approach helped us ensure we were on the right track and didn’t waste time or resources building something customers didn’t actually want. Deliver your solution using a manual consulting approach before developing an automated product Initially, we solved problems for customers in a more "consulting" manner, delivering insights manually. Recall how I mentioned that when one of our early customer interviewees asked us to deliver the solution, we initially did it manually by using machine learning models to analyse the data and presenting the results to them in a slide deck. This works for the initial stages of validating your solution, as you don't want to invest too much time into building a full-blown MVP before understanding the exact features and functionalities that your users want. However, after confirming that customers were willing to pay for what we provided, we moved forward with actual product development. This shift from a manual service to product development was key to scaling in a sustainable manner, as our building was guided by real-world feedback and insights rather than intuition. Let ongoing customer feedback drive iteration and the product roadmap Once we built the first version of the product, it was basic, solving only one problem. But as we worked closely with customers, they requested additional features and functionalities to make it more useful. As a result, we continued to evolve the product to handle more complex use cases, gradually developing new modules based on customer feedback. Product development is a continuous process. Our early customers pushed us to expand features and modules, from solving just 20% of their problems to tackling 50–60% of their needs. These demands shaped our product roadmap and guided the development of new features, ultimately resulting in a more complete solution. Revenue and user numbers are key metrics for assessing product-market fit. However, critical mass varies across industries Product-market fit (PMF) can often be gauged by looking at the size of your revenue and the number of customers you're serving. Once you've reached a certain critical mass of customers, you can usually tell that you're starting to hit product-market fit. However, this critical mass varies by industry and the type of customers you're targeting. For example, if you're building an app for a broad consumer market, you may need thousands of users. But for enterprise software, product-market fit may be reached with just a few dozen key customers. Compare customer engagement and retention with other available solutions on the market for product-market fit Revenue and the number of customers alone isn't always enough to determine if you're reaching product-market fit. The type of customer and the use case for your product also matter. The level of engagement with your product—how much time users are spending on the platform—is also an important metric to track. The more time they spend, the more likely it is that your product is meeting a crucial need. Another way to evaluate product-market fit is by assessing retention, i.e whether users are returning to your platform and relying on it consistently, as compared to other solutions available. That's another key indication that your solution is gaining traction in the market. Business Model & Monetisation Prioritise scalability Initially, we started with a consulting-type model where we tailor-made specific solutions for each customer use-case we encountered and delivered the CPG insights manually, but we soon realized that this wasn't scalable. The problem with consulting is that you need to do the same work repeatedly for every new project, which requires a large team to handle the workload. That is not how you sustain a high-growth startup. To solve this, we focused on building a product that would address the most common problems faced by our customers. Once built, this product could be sold to thousands of customers without significant overheads, making the business scalable. With this in mind, we decided on a SaaS (Software as a Service) business model. The benefit of SaaS is that once you create the software, you can sell it to many customers without adding extra overhead. This results in a business with higher margins, where the same product can serve many customers simultaneously, making it much more efficient than the consulting model. Adopt a predictable, simplistic business model for efficiency. Look to industry practices for guidance When it came to monetisation, we considered the needs of our CPG customers, who I knew from experience were already accustomed to paying annual subscriptions for sales databases and other software services. We decided to adopt the same model and charge our customers an annual upfront fee. This model worked well for our target market, aligning with industry standards and ensuring stable, recurring revenue. Moreover, our target CPG customers were already used to this business model and didn't have to choose from a huge variety of payment options, making closing sales a straightforward and efficient process. Marketing & Sales Educate the market to position yourself as a thought leader When we started, AI was not widely understood, especially in the CPG industry. We had to create awareness around both AI and its potential value. Our strategy focused on educating potential users and customers about AI, its relevance, and why they should invest in it. This education was crucial to the success of our marketing efforts. To establish credibility, we adopted a thought leadership approach. We wrote blogs on the importance of AI and how it could solve problems for CPG companies. We also participated in events and conferences to demonstrate our expertise in applying AI to the industry. This helped us build our brand and reputation as leaders in the AI space for CPG, and word-of-mouth spread as customers recognized us as the go-to company for AI solutions. It’s tempting for startups to offer products for free in the hopes of gaining early traction with customers, but this approach doesn't work in the long run. Free offerings don’t establish the value of your product, and customers may not take them seriously. You should always charge for pilots, even if the fee is minimal, to ensure that the customer is serious about potentially working with you, and that they are committed and engaged with the product. Pilots/POCs/Demos should aim to give a "flavour" of what you can deliver A paid pilot/POC trial also gives you the opportunity to provide a “flavour” of what your product can deliver, helping to build confidence and trust with the client. It allows customers to experience a detailed preview of what your product can do, which builds anticipation and desire for the full functionality. During this phase, ensure your product is built to give them a taste of the value you can provide, which sets the stage for a broader, more impactful adoption down the line. Fundraising & Financial Management Leverage PR to generate inbound interest from VCs When it comes to fundraising, our approach was fairly traditional—we reached out to VCs and used connections from existing investors to make introductions. However, looking back, one thing that really helped us build momentum during our fundraising process was getting featured in Tech in Asia. This wasn’t planned; it just so happened that Tech in Asia was doing a series on AI startups in Southeast Asia and they reached out to us for an article. During the interview, they asked if we were fundraising, and we mentioned that we were. As a result, several VCs we hadn’t yet contacted reached out to us. This inbound interest was incredibly valuable, and we found it far more effective than our outbound efforts. So, if you can, try to generate some PR attention—it can help create inbound interest from VCs, and that interest is typically much stronger and more promising than any outbound strategies because they've gone out of their way to reach out to you. Be well-prepared and deliberate about fundraising. Keep trying and don't lose heart When pitching to VCs, it’s crucial to be thoroughly prepared, as you typically only get one shot at making an impression. If you mess up, it’s unlikely they’ll give you a second chance. You need to have key metrics at your fingertips, especially if you're running a SaaS company. Be ready to answer questions like: What’s your retention rate? What are your projections for the year? How much will you close? What’s your average contract value? These numbers should be at the top of your mind. Additionally, fundraising should be treated as a structured process, not something you do on the side while juggling other tasks. When you start, create a clear plan: identify 20 VCs to reach out to each week. By planning ahead, you’ll maintain momentum and speed up the process. Fundraising can be exhausting and disheartening, especially when you face multiple rejections. Remember, you just need one investor to say yes to make it all worthwhile. When using funds, prioritise profitability and grow only when necessary. Don't rely on funding to survive. In the past, the common advice for startups was to raise money, burn through it quickly, and use it to boost revenue numbers, even if that meant operating at a loss. The idea was that profitability wasn’t the main focus, and the goal was to show rapid growth for the next funding round. However, times have changed, especially with the shift from “funding summer” to “funding winter.” My advice now is to aim for profitability as soon as possible and grow only when it's truly needed. For example, it’s tempting to hire a large team when you have substantial funds in the bank, but ask yourself: Do you really need 10 new hires, or could you get by with just four? Growing too quickly can lead to unnecessary expenses, so focus on reaching profitability as soon as possible, rather than just inflating your team or burn rate. The key takeaway is to spend your funds wisely and only when absolutely necessary to reach profitability. You want to avoid becoming dependent on future VC investments to keep your company afloat. Instead, prioritize reaching break-even as quickly as you can, so you're not reliant on external funding to survive in the long run. Team-Building & Leadership Look for complementary skill sets in co-founders When choosing a co-founder, it’s important to find someone with a complementary skill set, not just someone you’re close to. For example, I come from a business and commercial background, so I needed someone with technical expertise. That’s when I found my co-founder, Himanshu, who had experience in machine learning and AI. He was a great match because his technical knowledge complemented my business skills, and together we formed a strong team. It might seem natural to choose your best friend as your co-founder, but this can often lead to conflict. Chances are, you and your best friend share similar interests, skills, and backgrounds, which doesn’t bring diversity to the table. If both of you come from the same industry or have the same strengths, you may end up butting heads on how things should be done. Having diverse skill sets helps avoid this and fosters a more collaborative working relationship. Himanshu (left) and Somsubhra (right) co-founded AI Palette in 2018 Define roles clearly to prevent co-founder conflict To avoid conflict, it’s essential that your roles as co-founders are clearly defined from the beginning. If your co-founder and you have distinct responsibilities, there is no room for overlap or disagreement. This ensures that both of you can work without stepping on each other's toes, and there’s mutual respect for each other’s expertise. This is another reason as to why it helps to have a co-founder with a complementary skillset to yours. Not only is having similar industry backgrounds and skillsets not particularly useful when building out your startup, it's also more likely to lead to conflicts since you both have similar subject expertise. On the other hand, if your co-founder is an expert in something that you're not, you're less likely to argue with them about their decisions regarding that aspect of the business and vice versa when it comes to your decisions. Look for employees who are driven by your mission, not salary For early-stage startups, the first hires are crucial. These employees need to be highly motivated and excited about the mission. Since the salary will likely be low and the work demanding, they must be driven by something beyond just the paycheck. The right employees are the swash-buckling pirates and romantics, i.e those who are genuinely passionate about the startup’s vision and want to be part of something impactful beyond material gains. When employees are motivated by the mission, they are more likely to stick around and help take the startup to greater heights. A litmus test for hiring: Would you be excited to work with them on a Sunday? One of the most important rounds in the hiring process is the culture fit round. This is where you assess whether a candidate shares the same values as you and your team. A key question to ask yourself is: "Would I be excited to work with this person on a Sunday?" If there’s any doubt about your answer, it’s likely not a good fit. The idea is that you want employees who align with the company's culture and values and who you would enjoy collaborating with even outside of regular work hours. How we structure the team at AI Palette We have three broad functions in our organization. The first two are the big ones: Technical Team – This is the core of our product and technology. This team is responsible for product development and incorporating customer feedback into improving the technology Commercial Team – This includes sales, marketing, customer service, account managers, and so on, handling everything related to business growth and customer relations. General and Administrative Team – This smaller team supports functions like finance, HR, and administration. As with almost all businesses, we have teams that address the two core tasks of building (technical team) and selling (commercial team), but given the size we're at now, having the administrative team helps smoothen operations. Set broad goals but let your teams decide on execution What I've done is recruit highly skilled people who don't need me to micromanage them on a day-to-day basis. They're experts in their roles, and as Steve Jobs said, when you hire the right person, you don't have to tell them what to do—they understand the purpose and tell you what to do. So, my job as the CEO is to set the broader goals for them, review the plans they have to achieve those goals, and periodically check in on progress. For example, if our broad goal is to meet a certain revenue target, I break it down across teams: For the sales team, I’ll look at how they plan to hit that target—how many customers they need to sell to, how many salespeople they need, and what tactics and strategies they plan to use. For the technical team, I’ll evaluate our product offerings—whether they think we need to build new products to attract more customers, and whether they think it's scalable for the number of customers we plan to serve. This way, the entire organization's tasks are cascaded in alignment with our overarching goals, with me setting the direction and leaving the details of execution to the skilled team members that I hire.

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
GummyBear8659This week

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)

Hey everyone! Long-time lurker, first-time founder here. I’m reaching out to get feedback on a recent startup experience—what went wrong, what I could have done better, and how I should approach future opportunities. The Background There were three founders in this venture: • Founder A (CEO, 50%) – The product/growth guy who identified the problem space. • Founder B (Me, CTO, 37.5%) – A software engineer with a software dev shop and multiple clients. I wanted to diversify into building my own products but am not inherently a “product person.” • Founder C (COO, 12.5%) – Brought into the mix by Founder A, with the goal of leveraging his network for traction once the product was built. The idea was to create Product X, a solution targeting the SMB space while competitors were moving upmarket. It wasn’t revolutionary—more of a strategic market play. The Initial Plan & My Role • Founder A would define and prioritize product specs, guiding what needed to be built. • I (Founder B) didn’t have time to code myself, so I allocated engineers from my dev shop (which I personally paid for). My stake was adjusted from 32.5% to 37.5% to reflect this contribution. • Founder C was more of an observer early on, planning to help with traction once we had a product ready. We agreed on a 1-year cliff and a 4-year vesting schedule for equity. Where Things Started to Go Wrong • Lack of a Clear Product Roadmap – Founder A was very focused on getting something built fast, but we never signed off on a structured roadmap or milestones. I underestimated the complexity of what was actually needed for customer conversations. • Engineering Expectations vs. Reality – The team (one part-time lead + two full-time juniors from my dev shop) faced early feedback that development was too slow. In response, I ramped up the lead to full-time and added a part-time PM. But Founder A continued pushing for speed, despite real hurdles (OAuth integrations, etc.). • Shifting MVP Goalposts – Midway, Founder A concluded that an MVP wouldn’t cut it—we needed a more complete product to be competitive. This meant more engineering, more delays, and more of my own money spent on development. The Breaking Point Near the 1-year vesting mark, we had an opportunity: a paying client willing to fund an app. I didn’t have devs on the bench, so I asked Founder A to hold off our project briefly while I hired more engineers to avoid stalling either effort. This was the final straw. Founder A (with Founder C somewhat aligned) decided the arrangement wasn’t working—citing past disagreements and the “slowness” issue. The decision was made to end the partnership. Now, Founder A, as majority holder, is requesting a full handover of the code, Founder C is indifferent, and all engineering costs I covered are essentially lost. Key Takeaways (So Far) Crystal-Clear Agreements Upfront – A formalized product roadmap and timeline should’ve been locked in from day one. Business Needs > Engineering Standards – I wanted to build something solid and scalable, but in an early-stage startup, speed to market is king. This was before AI tools became mainstream, so our approach wasn’t as optimized. Don’t Overextend Without Protection – I personally financed all engineering, but without clear safeguards, that investment became a sunk cost. Expenses Must Be Distributed – I was solely covering engineering salaries, which created an imbalance in financial risk. Future partnerships should ensure costs are shared proportionally, rather than one person shouldering the burden. Where I Need Advice Looking back, I want to improve as an engineer, CEO, and co-founder. • What should I have done differently in structuring this partnership? • How do you balance engineering quality with the startup need for speed? • As a dev shop owner, how can I better navigate equity deals where I’m also bringing in engineering resources? I really appreciate everyone who went through this long post and provide any insights from founders, engineers, or anyone who has been in a similar situation. Thanks for reading! ===================================================================== For readers who might be thinking what set this type of expectation? Because I had a dev shop and I thought my co-founders will be understanding of my business circumstance and I was a bit trigger to build a product with a C-exec team, I gave the impression of "unlimited" engineering which I later realized down the line that it was not feasible for me. Something I learned that I have to be more careful with and set expectations accordingly from the very beginning. And from the feedback of the commenters here, I am much more aware what I should offer and how to set expectations, esp. in the early stages of execution. So thank you all! 🙏🏾 EDIT: I would like to thank everyone who contributed to this thread. You not only helped me but future founders who are considering to get into the startup scene!

Looking for a tech cofounder. Revoltionary (yes really!) gig economy app. I will not promote.
reddit
LLM Vibe Score0
Human Vibe Score1
sweetpea___This week

Looking for a tech cofounder. Revoltionary (yes really!) gig economy app. I will not promote.

Hey everyone! I’m building a new gig-work app that cuts out the hassles of interviews, applications, and sky-high fees. We’re aiming to make it easy for businesses to hire qualified freelancers for short shifts or one-off tasks—and for freelancers to set their own rates and get paid quickly. Why This App? Time-Saving Model: Instead of posting jobs and conducting multiple interviews, employers can instantly book from a list of KYC-verified freelancers who showcase their skills via 30-second video bios. Cost Leadership: We plan to charge only 5%, far below the 15–50% common in other gig platforms. This keeps more money in the pockets of both freelancers and businesses. Proven Demand: A beta test in 2018 drew nearly 600 active users, validating that there’s appetite for a simpler, fairer way to fill short shifts. About Me 20+ years’ experience in payroll, workforce management, and operations for Fortune 500 companies. Led cross-functional teams, implemented large-scale solutions, and believe in building with a user-first mindset. Offering meaningful equity—I want a true partner, not a hired gun. Who I’m Looking For Full-Stack Developer (comfortable with Node.js, React, Python, or similar and ML/Ai) who can manage everything from front-end to database integration (ideally Postgres/MySQL) and build a same day payments system. Passion for creating solutions that genuinely help gig workers and small businesses. Excitement to collaborate on the product roadmap, from the booking interface to same-day payment features. The Opportunity Major Market: The gig economy is huge and still growing. If we nail speed, cost-effectiveness, and ease of use, we can capture a significant share of it. Remote-Friendly: We can work together from anywhere, though I’m planning to relaunch in London where the initial beta gained momentum. If this sounds like your kind of challenge, drop a comment or DM me. Let’s chat about how we can merge our strengths—my operations background and your technical expertise—to build a platform that truly transforms the gig-work experience. Thanks for reading, and I look forward to creating something impactful together!

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
GummyBear8659This week

What I Learned from a Failed Startup: Seeking Advice on Engineering, Co-Founder Agreements & Execution (i will not promote)

Hey everyone! Long-time lurker, first-time founder here. I’m reaching out to get feedback on a recent startup experience—what went wrong, what I could have done better, and how I should approach future opportunities. The Background There were three founders in this venture: • Founder A (CEO, 50%) – The product/growth guy who identified the problem space. • Founder B (Me, CTO, 37.5%) – A software engineer with a software dev shop and multiple clients. I wanted to diversify into building my own products but am not inherently a “product person.” • Founder C (COO, 12.5%) – Brought into the mix by Founder A, with the goal of leveraging his network for traction once the product was built. The idea was to create Product X, a solution targeting the SMB space while competitors were moving upmarket. It wasn’t revolutionary—more of a strategic market play. The Initial Plan & My Role • Founder A would define and prioritize product specs, guiding what needed to be built. • I (Founder B) didn’t have time to code myself, so I allocated engineers from my dev shop (which I personally paid for). My stake was adjusted from 32.5% to 37.5% to reflect this contribution. • Founder C was more of an observer early on, planning to help with traction once we had a product ready. We agreed on a 1-year cliff and a 4-year vesting schedule for equity. Where Things Started to Go Wrong • Lack of a Clear Product Roadmap – Founder A was very focused on getting something built fast, but we never signed off on a structured roadmap or milestones. I underestimated the complexity of what was actually needed for customer conversations. • Engineering Expectations vs. Reality – The team (one part-time lead + two full-time juniors from my dev shop) faced early feedback that development was too slow. In response, I ramped up the lead to full-time and added a part-time PM. But Founder A continued pushing for speed, despite real hurdles (OAuth integrations, etc.). • Shifting MVP Goalposts – Midway, Founder A concluded that an MVP wouldn’t cut it—we needed a more complete product to be competitive. This meant more engineering, more delays, and more of my own money spent on development. The Breaking Point Near the 1-year vesting mark, we had an opportunity: a paying client willing to fund an app. I didn’t have devs on the bench, so I asked Founder A to hold off our project briefly while I hired more engineers to avoid stalling either effort. This was the final straw. Founder A (with Founder C somewhat aligned) decided the arrangement wasn’t working—citing past disagreements and the “slowness” issue. The decision was made to end the partnership. Now, Founder A, as majority holder, is requesting a full handover of the code, Founder C is indifferent, and all engineering costs I covered are essentially lost. Key Takeaways (So Far) Crystal-Clear Agreements Upfront – A formalized product roadmap and timeline should’ve been locked in from day one. Business Needs > Engineering Standards – I wanted to build something solid and scalable, but in an early-stage startup, speed to market is king. This was before AI tools became mainstream, so our approach wasn’t as optimized. Don’t Overextend Without Protection – I personally financed all engineering, but without clear safeguards, that investment became a sunk cost. Expenses Must Be Distributed – I was solely covering engineering salaries, which created an imbalance in financial risk. Future partnerships should ensure costs are shared proportionally, rather than one person shouldering the burden. Where I Need Advice Looking back, I want to improve as an engineer, CEO, and co-founder. • What should I have done differently in structuring this partnership? • How do you balance engineering quality with the startup need for speed? • As a dev shop owner, how can I better navigate equity deals where I’m also bringing in engineering resources? I really appreciate everyone who went through this long post and provide any insights from founders, engineers, or anyone who has been in a similar situation. Thanks for reading! ===================================================================== For readers who might be thinking what set this type of expectation? Because I had a dev shop and I thought my co-founders will be understanding of my business circumstance and I was a bit trigger to build a product with a C-exec team, I gave the impression of "unlimited" engineering which I later realized down the line that it was not feasible for me. Something I learned that I have to be more careful with and set expectations accordingly from the very beginning. And from the feedback of the commenters here, I am much more aware what I should offer and how to set expectations, esp. in the early stages of execution. So thank you all! 🙏🏾 EDIT: I would like to thank everyone who contributed to this thread. You not only helped me but future founders who are considering to get into the startup scene!

Scratch Machine Learning Algorithms Implementations
reddit
LLM Vibe Score0
Human Vibe Score1
ParkMountainThis week

Scratch Machine Learning Algorithms Implementations

Hi there, other Redditors! Like many of you, when I first started working in the AI field, I wanted to build some basic Machine Learning models from scratch in order to better understand how each algorithm works, improve my programming and math skills, or simply produce an eye-catching, difficult project to put in the résumé. After spending some time searching for resources that could help me guide my studies, I discovered that the majority of scratch implementations that are currently available are either i) outdated (having been implemented years ago using Python 2 or an earlier version of Python 3); ii) too difficult to understand (using a lot of difficult, unfriendly optimization techniques or with poorly written code); or iii) too simple (only covering binary classification). With that in mind, I made the decision to develop user-friendly, uncomplicated, organized, and simple implementations from scratch. Aside from all of that, I've always wanted to create an open-source project so that others, particularly novices and those with less than a year's experience (like me), can collaborate with others, contribute to public projects, and experience Git firsthand (some of these implementations were made by other contributors!). Here are some implementations that are available: Algorithms (Random Forest Classifier and Regressor, Decision Tree Classifier and Regressor, KMeans, KNN Classifier and Regressor, Gaussian Naive Bayes, Linear Regression, Logistic Regression, PCA, Perceptron, MLP Classifier and Regressor, SVM Classifier and Regressor); Regression and classification metrics; Distance metrics (such as Euclidean); Data split functions (such as KFold); Activation and loss functions; Scalers (such as MinMaxScaler) and encoders (such as One Hot Encoder); and a few things more! Project's link: https://github.com/rafaelgreca/scratchml Disclaimer: The goal of this library is to provide code that is simpler, easier to understand, and more approachable for artificial intelligence enthusiasts and beginners who want to contribute to an open-source repository or who want to learn more about how algorithms work. It is not meant to replace existing libraries that are better, more optimized, and have a wider variety of implemented algorithms (such as scikit-learn, PyTorch, Keras, and Tensorflow). If you want to use optimized implementations with accurate results, please use one of the previously mentioned libraries. P.S.: I accidentally deleted the other post, so I am posting again. :-)

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.
reddit
LLM Vibe Score0
Human Vibe Score0.714
level6-killjoyThis week

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.

This is a recap covering the major news from last week. 🔥Top 3 news - OpenAI’s updates, Meta’s upcoming free LLM and EU Regulation 🗞️Interesting reads include PSA about protecting your keys, The GPT ouroboros, Reddit - OpenAI’s moat, and more.. 🧑‍🎓Learning includes a Step-by-step guide from a non-technical founder who launched his MVP, Chatbot for your Gdrive and more 🔥Top 3 AI news in the past week OpenAI: New Pricing, Models, & Functions OpenAI has been on a roll. Last week we saw the release of OpenAI best practice on using GPT. This week we saw some amazing updates. Three major buckets were: First, the price decreases for both embeddings and GPT-3.5 tokens. Second, new models for gpt-4 and gpt-3.5. A new longer context model for gpt-3.5. Third, a new function calling capability. Why is it important? Previously, the output from OpenAI was all text. So, calling an external API from GPT was quite difficult. You had to parse the text data and things were often incorrect. Langchain created the Agents and Tools feature to tackle this problem. It was still unreliable and prone to issues. Now you get native support to generate a fixed format output. You can use the output to generate functional calls and also pass functions which need to be called. For example, if your app has multiple API endpoints then you can use GPT to generate the API calls with parameters. You can also pass the endpoints as function calls to ensure the correct function is executed. This functionality can further be used to generate structured data (JSON) out of GPT. So, you can generate data from GPT and load it into your backend. What’s next? This functionality allows turning natural language responses into structured data. This can be used to create “intelligent” backends using LLMs. We might see implementations in no-code tools to allow more robust and natural-language tools for non-technical folks. The structured data process goes both ways. You can also feed structured data into GPT for better responses. This feature also has its share of issues. Function calling suffers from the same prompt injection issues. Malicious actors can pass malicious code in function or the responses. For example, creation of queries using functions might contain malicious code to delete data. Without proper user validation this code will be executed automatically and delete data. So, using LLM as the back-end layer needs proper security implementation. Meta's LLM: Commercial Use Ahead Llama has been a boon for the open source community. Many of the open source models rely on Llama. The issue is that Llama is research-only and cannot be used commercially. So, no one can use it to build any product. Meta is now working on the next version of the model. This model will be available for commercial use. This is in stark contrast to both OpenAI and Google. Both safe-guarde their models and make it available through API. Why is it important? Certain industries cannot use LLM APIs because of strict restrictions on data privacy. These companies would want to run their own instance of a foundational model. A commercially available foundational model is also going to help people who want to keep their “API call” costs next to 0. A commercially available free-for-all model will also help push the open source community further. Just like Llama. What’s next? Sam Altman has said OpenAI didn’t release GPT-3 as open-source because they didn’t think people would be able to run it. Now OpenAI is working on an open-source model. This is going to be weaker than GPT-4. Let the battle of LLMs begin. EU's Proposed Legislation and Its Impact on AI Usage The EU parliament voted to move ahead with the E.U. AI Act. This act aims to ensure consumer protection against the dangers of AI. Why is it important? OpenAI and Sam Altman want regulations for models. They have proposed a IAEA-type of agency to stop the proliferation of LLM models. As per OpenAI, all models should be regulated and monitored. The suggestion of a license based regulation has led to significant backlash. Many people have called it “regulatory capture” - with the aim of shutting down competing LLMs. Licensing based regulations might not really be effective. The EU is approaching regulation from a different angle. It doesn’t focus on how models are developed. Rather focuses on how AI will/can be used. They have broken down use cases into 4 categories - unacceptable (prohibited), high, medium and low risk. For example, Building a Pre-Crime software,on%20crimes%20not%20yet%20committed.) to predict crimes? Building a Social credit system? Unacceptable. Using tools to influence elections or recommendation algorithms? High (Highly regulated). Using generative AI tools to create text or images on news sites? Medium (Add label that the content is AI generated) AI providers also need to disclose their training source. To me this sounds like good legislation. What do you guys think? But, OpenAI has warned that EU regulations might force them to pull out completely. What’s next? The disclosure requirements might help various publishing companies. AI and media companies are in talks to pay for training data. Google has been leading the charge. Additionally, OpenAI and Deepmind will open their models for safety and research purposes to the UK government. 🗞️10 AI news highlights and interesting reads PSA: If you are using Repl to write code, you might want to check your OpenAI API keys. If you have left them embedded then people can pirate and steal the keys. LLMs rely on human annotation or human feedback to learn. And one way to generate human annotation is crowdsourcing. But what if the crowdsource human annotators use LLMs? Research shows 33-46% workers used LLMs. So, basically we go from Human -> AI -> Human -> AI. The AI ouroboros. Researchers also say generated data to train models might cause serious issue. All the talks about moats \- Reddit might be OpenAI’s \future\ moat. Given the amount of complaints about how Google search experience has deteriorated during the blackout, this might be true? Doctors are using ChatGPT but not to diagnose.Rather to be more empathetic. We discussed this just a month ago. And guess where the data for this study came from? Reddit AskDocs. Moat FTW?! Beatles to make a comeback…using Generative AI. SnapFusion - Text to Image diffusion on mobile phones. Large context lengths are important for better GPT experience. The secret sauce for 100k context length. There is a lot of bad AI research out there. Some border on snake oil. Most AI “research” should be double checked and challenged. A new research on huggingface said that GPT-4 can ace MIT curriculum. Now someone is replicating the results and say that GPT-4 can’t beat MIT. Are we seeing peak AI? Especially when people from Deepmind and Meta are involved? Mistral AI raised $113 million in seed round with no product. Some might say this funding is for the team and the team is really solid. The issue though is whether the valuation is justified when OpenAI and Google already have a head start. The AI Hype Wall of Shame. \- Collection of articles which mislead people about AI in various aspects. 🧑‍🎓3 Learning Resources Building and Launching a company using GPT-4 with prompts. (The author didn’t know how to code but created and launched the MVP in a month). Chatbot for your Gdrive - https://www.haihai.ai/gpt-gdrive/ Building ChatGPT plugin using Supabase - https://supabase.com/blog/building-chatgpt-plugins-template That’s it folks. Thank you for reading and have a great week ahead. If you are interested in a focused weekly recap delivered to your inbox on Mondays you can subscribe here. It is FREE!

Randomly asked ChatGPT and Claude for a 4 year roadmap for an ML Engineer
reddit
LLM Vibe Score0
Human Vibe Score1
Brilliant_Fishing110This week

Randomly asked ChatGPT and Claude for a 4 year roadmap for an ML Engineer

Title, Is it actually a good plan ?? If no, why not ?? \\🚀 4-Year Roadmap to Becoming a High-Earning ML Engineer & Entrepreneur\\ \\(With Smartwork & Realistic 60-70% Execution Feasibility)\\ \\🟢 Year 1: Strong Foundation & Initial Projects (0-12 Months)\\ 🎯 \\Goal: Master Python & ML Fundamentals\\ \\🔹 1-4 Months (Python & Math Strengthening)\\ ✅ Python Mastery \- Daily LeetCode Easy problems (minimum 2) \- Build automation projects \- NumPy & Pandas mastery \- DSA fundamentals ✅ Mathematics Foundation \- Linear Algebra basics \- Statistics fundamentals \- Basic calculus concepts ✅ First Mini-Hackathon Participation \- Join beginner-friendly hackathons \- Focus on Python-based challenges \- Team up with other beginners 💡 \\Smart Move:\\ \- Join Discord/Slack hackathon communities \- Practice collaborative coding \- Build network with fellow participants \\🔹 5-8 Months (ML Foundations)\\ ✅ Machine Learning Basics \- Supervised Learning \- Model evaluation \- Feature engineering \- scikit-learn projects ✅ Participate in 2-3 ML Hackathons \- Kaggle Getting Started competitions \- Local ML hackathons \- University hackathons ✅ Start LinkedIn & GitHub Portfolio 💡 \\Smart Move:\\ \- Document hackathon experiences \- Share learnings on LinkedIn \- Focus on completion over winning \\🔹 9-12 Months (Deep Learning Introduction)\\ ✅ Basic Deep Learning \- Neural network fundamentals \- PyTorch basics \- Computer vision tasks \- Basic NLP ✅ Advanced Hackathon Participation \- AI/ML specific hackathons \- Team lead in 1-2 hackathons \- Start mentoring beginners \\🔵 Year 1 Expected Outcome (60-70% Execution)\\ ✔ \\Strong Python & ML foundations\\ ✔ \\5-6 hackathon participations\\ ✔ \\Active GitHub (100+ commits)\\ ✔ \\Growing LinkedIn (300+ connections)\\ 💰 \\Earning Expectation → ₹8K-₹20K per month (Projects/Internship)\\ \\🟢 Year 2: Professional Growth & Specialization (12-24 Months)\\ 🎯 \\Goal: Build Professional Experience & Recognition\\ \\🔹 1-6 Months (Technical Depth)\\ ✅ Advanced ML Topics \- Deep Learning architectures \- Computer Vision OR NLP \- MLOps basics (Docker, FastAPI) \- Cloud fundamentals (AWS/GCP) ✅ Hackathon Achievements \- Win minor prizes in 2-3 hackathons \- Lead teams in major hackathons \- Network with sponsors ✅ Start Technical Blogging 💡 \\Smart Move:\\ \- Focus on hackathon projects that align with career goals \- Build relationships with companies at hackathons \- Create detailed project documentation \\🔹 7-12 Months (Professional Experience)\\ ✅ Secure ML Role/Internship ✅ Advanced Project Building ✅ Open Source Contributions ✅ Organize Small Hackathons 💡 \\Smart Move:\\ \- Use hackathon network for job referrals \- Convert hackathon projects into full products \- Build mentor reputation \\🔵 Year 2 Expected Outcome (60-70% Execution)\\ ✔ \\Professional ML experience\\ ✔ \\10+ hackathon participations\\ ✔ \\1-2 hackathon wins\\ ✔ \\Strong industry network\\ 💰 \\Earning Expectation → ₹40K-₹70K per month (Job/Freelancing)\\ \\🟢 Year 3: Scaling & Business Foundation (24-36 Months)\\ 🎯 \\Goal: Establish Multiple Income Streams\\ \\🔹 1-4 Months (Expertise Building)\\ ✅ Choose Specialization \- MLOps \- Computer Vision \- NLP/LLMs \- Generative AI ✅ Advanced Competitions \- International hackathons \- High-prize competitions \- Corporate ML challenges ✅ Start Consulting Services 💡 \\Smart Move:\\ \- Use hackathon wins for marketing \- Build service packages around expertise \- Network with corporate sponsors \\🔹 5-8 Months (Business Development)\\ ✅ Scale Services ✅ Build Client Network ✅ Create Training Programs ✅ Hackathon Mentorship Program 💡 \\Smart Move:\\ \- Convert hackathon projects to products \- Use event networks for client acquisition \- Build authority through speaking \\🔹 9-12 Months (Growth & Innovation)\\ ✅ Product Development ✅ Team Building ✅ Innovation Focus ✅ Hackathon Organization \\🔵 Year 3 Expected Outcome (60-70% Execution)\\ ✔ \\Established ML business/career\\ ✔ \\Known in hackathon community\\ ✔ \\Multiple income streams\\ ✔ \\Strong industry presence\\ 💰 \\Earning Expectation → ₹1L-₹2L per month (Multiple Streams)\\ \\🟢 Year 4: Scale & Leadership (36-48 Months)\\ 🎯 \\Goal: Build AI Company & Achieve Financial Freedom\\ \\🔹 1-4 Months (Business Scaling)\\ ✅ Company Formation \- AI consulting firm \- Product development \- Training programs ✅ Hackathon Innovation \- Launch own hackathon series \- Corporate partnerships \- Prize sponsorships ✅ Team Expansion 💡 \\Smart Move:\\ \- Use hackathon network for hiring \- Create unique event formats \- Build corporate relationships \\🔹 5-8 Months (Market Leadership)\\ ✅ Product Launch ✅ Service Expansion ✅ International Presence ✅ Innovation Hub Creation 💡 \\Smart Move:\\ \- Create hackathon-to-hiring pipeline \- Build educational programs \- Establish thought leadership \\🔹 9-12 Months (Empire Building)\\ ✅ Multiple Revenue Streams \- AI products \- Consulting services \- Educational programs \- Event organization \- Investment returns ✅ Industry Leadership \- Conference speaking \- Published content \- Community leadership \\🔵 Year 4 Expected Outcome (60-70% Execution)\\ ✔ \\Established AI company\\ ✔ \\Major hackathon organizer\\ ✔ \\Multiple product lines\\ ✔ \\Industry authority status\\ 💰 \\Earning Expectation → ₹3L-₹5L+ per month (Business Income)\\ \\📊 FINAL RATING\\ ✅ \\Comprehensive growth plan\\ ✅ \\Strong community focus\\ ✅ \\Multiple income pathways\\ 💡 \\If 100% Execution → 8.5/10 Feasibility\\ 💡 \\If 50% Execution → 6/10 Feasibility\\ 🔥 \\Conclusion: A balanced path to ML mastery and entrepreneurship, built through consistent growth and community engagement!\\ 🚀 \\Key Success Factors:\\ Regular hackathon participation Strong community involvement Consistent skill development Strategic network building Focus on both technical and business growth

Randomly asked ChatGPT and Claude for a 4 year roadmap for an ML Engineer
reddit
LLM Vibe Score0
Human Vibe Score1
Brilliant_Fishing110This week

Randomly asked ChatGPT and Claude for a 4 year roadmap for an ML Engineer

Title, Is it actually a good plan ?? If no, why not ?? \\🚀 4-Year Roadmap to Becoming a High-Earning ML Engineer & Entrepreneur\\ \\(With Smartwork & Realistic 60-70% Execution Feasibility)\\ \\🟢 Year 1: Strong Foundation & Initial Projects (0-12 Months)\\ 🎯 \\Goal: Master Python & ML Fundamentals\\ \\🔹 1-4 Months (Python & Math Strengthening)\\ ✅ Python Mastery \- Daily LeetCode Easy problems (minimum 2) \- Build automation projects \- NumPy & Pandas mastery \- DSA fundamentals ✅ Mathematics Foundation \- Linear Algebra basics \- Statistics fundamentals \- Basic calculus concepts ✅ First Mini-Hackathon Participation \- Join beginner-friendly hackathons \- Focus on Python-based challenges \- Team up with other beginners 💡 \\Smart Move:\\ \- Join Discord/Slack hackathon communities \- Practice collaborative coding \- Build network with fellow participants \\🔹 5-8 Months (ML Foundations)\\ ✅ Machine Learning Basics \- Supervised Learning \- Model evaluation \- Feature engineering \- scikit-learn projects ✅ Participate in 2-3 ML Hackathons \- Kaggle Getting Started competitions \- Local ML hackathons \- University hackathons ✅ Start LinkedIn & GitHub Portfolio 💡 \\Smart Move:\\ \- Document hackathon experiences \- Share learnings on LinkedIn \- Focus on completion over winning \\🔹 9-12 Months (Deep Learning Introduction)\\ ✅ Basic Deep Learning \- Neural network fundamentals \- PyTorch basics \- Computer vision tasks \- Basic NLP ✅ Advanced Hackathon Participation \- AI/ML specific hackathons \- Team lead in 1-2 hackathons \- Start mentoring beginners \\🔵 Year 1 Expected Outcome (60-70% Execution)\\ ✔ \\Strong Python & ML foundations\\ ✔ \\5-6 hackathon participations\\ ✔ \\Active GitHub (100+ commits)\\ ✔ \\Growing LinkedIn (300+ connections)\\ 💰 \\Earning Expectation → ₹8K-₹20K per month (Projects/Internship)\\ \\🟢 Year 2: Professional Growth & Specialization (12-24 Months)\\ 🎯 \\Goal: Build Professional Experience & Recognition\\ \\🔹 1-6 Months (Technical Depth)\\ ✅ Advanced ML Topics \- Deep Learning architectures \- Computer Vision OR NLP \- MLOps basics (Docker, FastAPI) \- Cloud fundamentals (AWS/GCP) ✅ Hackathon Achievements \- Win minor prizes in 2-3 hackathons \- Lead teams in major hackathons \- Network with sponsors ✅ Start Technical Blogging 💡 \\Smart Move:\\ \- Focus on hackathon projects that align with career goals \- Build relationships with companies at hackathons \- Create detailed project documentation \\🔹 7-12 Months (Professional Experience)\\ ✅ Secure ML Role/Internship ✅ Advanced Project Building ✅ Open Source Contributions ✅ Organize Small Hackathons 💡 \\Smart Move:\\ \- Use hackathon network for job referrals \- Convert hackathon projects into full products \- Build mentor reputation \\🔵 Year 2 Expected Outcome (60-70% Execution)\\ ✔ \\Professional ML experience\\ ✔ \\10+ hackathon participations\\ ✔ \\1-2 hackathon wins\\ ✔ \\Strong industry network\\ 💰 \\Earning Expectation → ₹40K-₹70K per month (Job/Freelancing)\\ \\🟢 Year 3: Scaling & Business Foundation (24-36 Months)\\ 🎯 \\Goal: Establish Multiple Income Streams\\ \\🔹 1-4 Months (Expertise Building)\\ ✅ Choose Specialization \- MLOps \- Computer Vision \- NLP/LLMs \- Generative AI ✅ Advanced Competitions \- International hackathons \- High-prize competitions \- Corporate ML challenges ✅ Start Consulting Services 💡 \\Smart Move:\\ \- Use hackathon wins for marketing \- Build service packages around expertise \- Network with corporate sponsors \\🔹 5-8 Months (Business Development)\\ ✅ Scale Services ✅ Build Client Network ✅ Create Training Programs ✅ Hackathon Mentorship Program 💡 \\Smart Move:\\ \- Convert hackathon projects to products \- Use event networks for client acquisition \- Build authority through speaking \\🔹 9-12 Months (Growth & Innovation)\\ ✅ Product Development ✅ Team Building ✅ Innovation Focus ✅ Hackathon Organization \\🔵 Year 3 Expected Outcome (60-70% Execution)\\ ✔ \\Established ML business/career\\ ✔ \\Known in hackathon community\\ ✔ \\Multiple income streams\\ ✔ \\Strong industry presence\\ 💰 \\Earning Expectation → ₹1L-₹2L per month (Multiple Streams)\\ \\🟢 Year 4: Scale & Leadership (36-48 Months)\\ 🎯 \\Goal: Build AI Company & Achieve Financial Freedom\\ \\🔹 1-4 Months (Business Scaling)\\ ✅ Company Formation \- AI consulting firm \- Product development \- Training programs ✅ Hackathon Innovation \- Launch own hackathon series \- Corporate partnerships \- Prize sponsorships ✅ Team Expansion 💡 \\Smart Move:\\ \- Use hackathon network for hiring \- Create unique event formats \- Build corporate relationships \\🔹 5-8 Months (Market Leadership)\\ ✅ Product Launch ✅ Service Expansion ✅ International Presence ✅ Innovation Hub Creation 💡 \\Smart Move:\\ \- Create hackathon-to-hiring pipeline \- Build educational programs \- Establish thought leadership \\🔹 9-12 Months (Empire Building)\\ ✅ Multiple Revenue Streams \- AI products \- Consulting services \- Educational programs \- Event organization \- Investment returns ✅ Industry Leadership \- Conference speaking \- Published content \- Community leadership \\🔵 Year 4 Expected Outcome (60-70% Execution)\\ ✔ \\Established AI company\\ ✔ \\Major hackathon organizer\\ ✔ \\Multiple product lines\\ ✔ \\Industry authority status\\ 💰 \\Earning Expectation → ₹3L-₹5L+ per month (Business Income)\\ \\📊 FINAL RATING\\ ✅ \\Comprehensive growth plan\\ ✅ \\Strong community focus\\ ✅ \\Multiple income pathways\\ 💡 \\If 100% Execution → 8.5/10 Feasibility\\ 💡 \\If 50% Execution → 6/10 Feasibility\\ 🔥 \\Conclusion: A balanced path to ML mastery and entrepreneurship, built through consistent growth and community engagement!\\ 🚀 \\Key Success Factors:\\ Regular hackathon participation Strong community involvement Consistent skill development Strategic network building Focus on both technical and business growth

How I built my SaaS and earned $273 MRR in the first month
reddit
LLM Vibe Score0
Human Vibe Score1
Ok_Damage_1764This week

How I built my SaaS and earned $273 MRR in the first month

Hi everyone! I’m Alex Varga, an indie developer. Last year, I focused on accelerating my development speed and launched 10 projects in 12 months. One of them called Bulk Image Generation started growing through SEO, so I decided to focus on it. After one month of SEO efforts, it’s generating $273 MRR. I hope my experience will be useful to others. Concept bulkimagegeneration.com website helps to generate up to 100 images in 15 seconds using AI I was using Google, started with keywords like "Bulk Image ..." a lot of them are Bulk Image Resizer, Downloader etc. But there was no Bulk Image Generator. I thought: yeah, this domain is available, let's buy. So I bought bulkimagegeneration.com and bulkimagegenerator.com So, the app concept is to help people generate images with AI at scale: let\`s say 100 images in 15 seconds. Marketing Gap https://preview.redd.it/4luzib02bbie1.png?width=1905&format=png&auto=webp&s=cbe845107aca46ae5729dfe121fefd5e9cdab9ac Most builders create a product first and figure out how to sell it later. I took a completely different approach with Bulk Image Generator. I identified a market gap and secured a domain name that matched exactly what people were searching for and launched app. https://preview.redd.it/h6vwur34bbie1.png?width=1905&format=png&auto=webp&s=9a163ff6f503be4c175c6e5e82e2003b32df1fe0 Growth Strategy SEO has become the main acquisition channel, so I’ve decided to focus even more on it with this experiment. Almost every day, I publish either a new article or a free micro-app (as a lead magnet) for Bulk Image Generator. I also tried Google Ads, spent $20, and got a $0.35 CPC. https://preview.redd.it/3rhnzvs6bbie1.png?width=1905&format=png&auto=webp&s=f9819d1e82d3e2429d6ccb7b00dcac86a7a351c2 In comparison, the Free Image to Text Prompt Converter (one of the lead magnets) has a $0.011 CPC, which is more than 30 times cheaper than Google Ads. So I decided not to focus now on paid ads. https://preview.redd.it/p333fyl9bbie1.png?width=1905&format=png&auto=webp&s=2e96532d7709b44b7459e7ccf37ef9a0fa784728 After using our free tools, some users explore our main product - a bulk image generation service. Users pay a monthly subscription to get credits, which they can spend on image generation, face swaps, and bulk background removal. Currently, this app generates around $250 in Monthly Recurring Revenue: https://preview.redd.it/9wcm0tjfbbie1.png?width=1905&format=png&auto=webp&s=41bcdd4f7594b09087c51cc5044e4b9c94c129c8 SEO Keyword Research I use Semrush or similar tools to find keywords with a search volume greater than 300 and then write articles targeting those keywords. If the topic has enough potential, I might create a free tool (e.g., a Free Image to Text Prompt Converter) to attract more users. Occasions matter. For instance, I wrote an article about creating images for Super Bowl ads, which led to one paying user who replicated the exact creatives showcased in the article https://preview.redd.it/shpax6mlbbie1.png?width=1905&format=png&auto=webp&s=d491385761df126424c2f9ba14c5da15f8cbb603 AI Tools Aggregators This can be an excellent acquisition channel. When BulkImageGeneration.com was featured in an article on Toolify.ai, I immediately gained three paying users (\~$60). I took 2 more AI Aggregators, and on average I had CPC = $0.2, which is a fair price and usually it has ROAs > 100%. However, some major aggregators are expensive ($300–400 per placement). I want to try it once I reach $500+ MRR. Next Steps bulkimagegeneration.com currently ranks #1 in search results for relevant keywords (e.g., “bulk image generation,” “bulk image generator”). I plan to keep producing content targeting niche keywords and timely occasions. buy more places in AI Aggregators I also want to reach out to YouTubers and ask them to include Bulk in their reviews for free

My Marketing App made $10,000 in 2024. Here is how I target to make $100,000 in 2025:
reddit
LLM Vibe Score0
Human Vibe Score1
MonkDiThis week

My Marketing App made $10,000 in 2024. Here is how I target to make $100,000 in 2025:

You totally get me, I think. It’s a bizarre feeling when you build something, and people appreciate it and are even ready to pay! Pleasant though) In early 2024 my mate and I created a marketing tool that generates ads, content and strategy blocks with a click – Aiter.io. Users can just insert a URL, hit the button and everything is ready. TBH, I built this tool because I’m too lazy to chat with ChatGPT) https://preview.redd.it/ew2kud7ceyde1.png?width=1140&format=png&auto=webp&s=f3fe5b67075858cea3d52278e8063113efa3b97e In 2024 we made $10,000, here is what worked for us: AI directories. Still is the best channel of traffic and clients for us. We listed on TAAFT and other directories scrape TAAFT, so, eventually, we became listed on all major ones. I wrote a Reddit post earlier that explained this process in detail. Email marketing. Gosh, I thought it was dead – I have never been so wrong! We set up automatic emails that share marketing insights and they have a \~25% open rate + consistently convert people. It works great. Product marketing. Having a free version really helps with word-of-mouth and leads, which can be converted via email. Also, we consistently worked on product improvement. I’d say, that our free updates give people a feeling that the devs care about their stuff that’s why they are more confident investing in it. Google Ads. TBH, we had a shitty landing page all the time because were busy with the product. So, Google Ads didn’t work well for us. But we’ve launched the 2.0 version which has a better landing page, and will try it again. Influencers. Worked well for us, but we didn’t pay a dime for this. They just found our tool on directories and created videos about Aiter, so it was a sporadic marketing channel for us. We hope to change it in 2025. We see that our product works and attracts the audience, so we want to deliver and get more in 2025. Here is the plan: Product: add ad banners and video generation. So far, we generate only text data and it’s not so valuable in the time of ChatGPT and Claude. But to generate a high-quality ad banner is still challenging, so we put this on our roadmap. Another feature – one-click market analysis to get marketing insights. Become a TOP50 tool on TAAFT. We’ve become a top tool in our category (content generation) but will need to promote our profile on the profile far more aggressively to get into TOP50 Email marketing. We are fools because we almost didn’t have product emails that explain how it works. Will fix it. Also, we are considering participating more in paid newsletters, like collaborating with Substack influencers. Youtube marketing. Search for low-tail marketing keywords on YouTube and create videos on them, placing my product in them. Blog. Our new platform is Webflow which gives a lot of flexibility in terms of blogging. So, we will repeat the YouTube strategy with blogging. Paid marketing. With an updated landing page, we hope that paid campaigns will work better. We plan to launch campaigns that target different jobs to be done and customer objections to find the right message. Product Management. For 2025, our two key product metrics are retention and product activation rate. For this, we plan to simplify onboarding and make it simpler as well as conduct a lot of in-depth interviews to understand how we can retain users better. Funding. All of this exciting stuff requires money, so we are in the process of securing funding (fingers crossed). Having an indie project is exciting and invigorating. With all these activities, I hope we will achieve the goal of $100,000 in 2025. And what are your goals and marketing steps for 2025? Or maybe you could share some exciting marketing ideas I overlooked?

I grew my mobile app to 1.4 million downloads
reddit
LLM Vibe Score0
Human Vibe Score1
TechPrimoThis week

I grew my mobile app to 1.4 million downloads

I started developing the app in early 2017, well before the AI era, when mobile apps were at their peak popularity. My idea was to create an app for emotional and psychological support in the form of helpful articles and various quizzes, such as personality assessments and life satisfaction tests. I named the app "Emotional Intelligence" because this keyword showed good ASO potential for positioning at the top of mobile stores. This proved to be accurate, and the app quickly gained traction in terms of downloads. A major problem I faced then was monetization. Unfortunately, in my country, it wasn't possible to sell through Google Play then, so I could only display ads. I started with Google AdMob, earning $2000 monthly after just a few months. The app then got about 1500 organic downloads daily and quickly surpassed 500,000. Three years after launching the app, I decided it was time for branding to build recognition. By combining the words "sentiment" and "intelligence," I came up with "Sintelly." I then pushed the app toward a social network, which differed from the right move. Adding features like discussion forums for problems, likes, and comments would result in even more growth, but the opposite happened. The app started declining, and I began investing in advertising campaigns. I managed to maintain a balance between income and expenses but without any profit. Then COVID-19 hit, and everything went downhill. I had to give up development and find a job as a developer to ensure my livelihood. Two years passed since I gave up, and that's when ChatGPT started gaining popularity. This immediately showed me how to steer the app towards active support for well-being questions. As I'm not an expert in psychology, I found several external psychotherapists who helped me put together CBT therapy, which I then implemented through a chatbot. This is how the new Sintelly app was born, with its main feature being a chatbot system composed of 17 AI agents that adapt to the user and guide them through a five-phase CBT therapy (I'll write a post about the technology). In addition to the agents, I added various exercises and tests to provide better personalization for the user. Initially, I made all of this free, which was also a mistake. I followed the principle of first showing what the app can do and gathering enough new users before starting to charge. I started selling subscriptions at the beginning of July, and since then, the app has had stable growth. If you want a check app, here is the link. Lessons learned: If things are working, don't touch them Start selling immediately upon app release; there's no need to wait Regularly test prices and types of subscriptions Onboarding is the most essential part of the app because most users buy subscriptions during onboarding It's essential to listen to user feedback. From day one, have a website and work on content to generate organic visits and redirect users from the web to the mobile app Stats: Over 1.4 million downloads 4.4 rating Only 40,000 active users (I had a massive loss during the period when I gave up) 280 active subscribers $3000 monthly revenue Next steps: Work on improving the Agent AI approach Setting up email campaigns and transactional emails Introducing in-app and push notifications Introducing gamification Potential for B2B I hope you can extract useful information from my example and avoid repeating my mistakes. I'm interested in your thoughts and if you have any recommendations for the next steps. I'm always looking to learn and improve.

How should I implement this local ai into my project?
reddit
LLM Vibe Score0
Human Vibe Score1
TrustingmeerkatThis week

How should I implement this local ai into my project?

I’m currently building a system that takes the text content of news articles about cocktail competitions and then attempts to extract a JSON object from it via using phi4 local ai model. I’m developing it alongside Claude in a project and we’ve built so far a series of qualifying questions that is prompted to phi4 and it’s answered are formatted to JSON I’m attempting to one shot each answer with the specific question and content of the article by asking the same question to phi4 3 times and picking majority answer. Then, the flow of questions are conditional so that the ai is provided a set of questions based on previous answers. I’m getting decent results and anecdotally it’s about 50% correct. So I think I need to begin prompt engineering to get better. Except, I’m wondering if there’s a way to automate these iterations a bit? Currently I’m pasting code and results into 01 preview and asking for detailed analysis, then passing this back into Claude for code revisions all manually. I guess I should design an accuracy test (again with ai) across 10 or so random articles at a time and a/b test until we get something we’re happy with? Does anyone else have any suggestions? I also previously attempted to one shot the entire JSON object rather than elect to flow through a bunch of questions except that didn’t work so well and decided to pivot rather than keep trying to optimise it.

Introducing Vest: Your AI-Powered Due Diligence Partner - Looking for feedback!
reddit
LLM Vibe Score0
Human Vibe Score1
nervousslinkyThis week

Introducing Vest: Your AI-Powered Due Diligence Partner - Looking for feedback!

TLDR; We are introducing Vest, an AI powered due-diligence and stock recommendation platform. We have bootstrapped ourselves so far and are wanting to get as much feedback from Reddit as we can to see where we can improve, but also what we are doing right. So please have a look around, give us feedback and if you like it, feel free to use it. Hi Reddit, My name is Drian and I'm one of the founders of Vest. We believe we are crafting something special at Vest and we want to get the word out and gather as much feedback as possible! Our major goal at Vest is to help new retail investors make sense of the investment landscape and get AI powered assistance, or even help experienced investors get confirmation of their potential moves. Overall, we want people to start their journey to financial freedom and not be daunted by the complexity of it. So how do we do this? Vest is a user-friendly service that harnesses fundamental metrics, social and news sentiment, and technical analysis, that we feed into some advanced AI models to generate clear buy, sell, or hold signals for US-based (for now!) stocks, offering our users transparent due-diligence for confident investing. The service is currently free with no ads - however, at some point we do plan on adding a paid tier. What's included: ​ Financial Metrics. Our financial metrics take all the potentially complex mathematical equations and present the fundamentals of a company to users in a simple 1 pager, with a score displaying if the metric is positive for a stock. We also provide publicly available analyst ratings from investment banks as well as price targets they have set. News Sentiment. We take publications about a specific stock from new articles, journals and socials and give these all a rating to determine if social sentiment is positive around a stock or not. Each article and its rating is visible to our users through through our dashboard. AI assisted Stock Signals. We have developed an algorithm to take all the metrics, sentiment and technical analysis we collate and analyze this with historic performance data for every stock to attempt to figure out if a stock is undervalued (great time to buy) or overvalued (great time to sell). 155 US stock tickers and counting. We currently have trained our models for around 155 US based stocks on the NASDAQ and NYSE exchanges. As we get more funding/runway we do plan on adding more, with the eventual goal to expand to more exchanges, countries and securities. Knowledge base and community. Our knowledge base & community contains explanations and articles for all metrics and the other good stuff behind Vest. We don’t want to just tell users what to do, but to also assist in their financial education. We hope our knowledge base can also become a thriving community where users can interact with us and each, ask questions around investing and keep gaining knowledge. Is it 100% accurate? Absolutely not. While we do a pretty great job at tracking and surfacing signals, we are not presenting a fool-proof, silver bullet with a guarantee here - rather a starting point for users to make more informed decisions, find potential new investment opportunities and hopefully learn about investing as they do so. We encourage our users to do their own research and due-diligence and not just take our signals as gospel - we know each and every person has a different risk appetite and goals, and we encourage you to use Vest in a way that fits with your own financial goals and risk appetite. We also display our win rates, average returns, and comparisons with buy and hold for each stock - and we are transparent about it when we’ve fallen short. Next steps: ​ Hope over to vestapp.ai and sign-up From the dashboard, play around, inspect our stock information and add some stocks to your watchlist. If you like what you see, and you’ve done your homework - use your favourite brokerage account to make an investment and watch Vest for changes in a stocks signals. If you don’t have one, we have a pop-up when you click buy/sell on any given stock with some non-affiliated brokerage options for the US, Australia and New Zealand - we don’t get a kickback from these brokerages, they are just what we’ve personally been using. FEEDBACK - We’re just getting started and we know the value of a fresh pair of eyes - our current mission is to get as much feedback as possible - anything you think of please send it through here or on the dedicated feedback form on our website in the sidebar on the left. Features we’re working on We're quietly thrilled about the direction Vest is headed, and we want to give you a sneak peek of what's in store for the next couple of quarters. Some of these may roll out as premium features, but we're diligently fine-tuning the details. Here's what you can expect: ​ Insider Trading Insights: Get daily reports on major stock moves by whales and company insiders. Institutional Holders: We're adding daily reports on institutional holders, keeping you informed about their moves. Lobbying Activity: We're actively working on daily updates about lobbying activities, so you can stay informed. Government Contracts Data: We'll provide a quarterly snapshot of government contract values for the companies you're tracking. US Congress Stock Activity: Keep an eye on daily trading actions of House and Senate members. Daily Summaries & Signal Alerts: We're currently hard at work on this feature. Soon, receive daily email summaries covering signals, watchlist updates, and key news. Personalized Risk Management: Tailor signals to match your unique risk management strategy. Your investments, your way. AI Assistant: Our LLM integration is almost ready, allowing you to ask it straightforward questions about particular securities in plain English. It will provide you with real-time context on fundamentals, news, and all the metrics and data points we monitor.

What are Boilerplates?
reddit
LLM Vibe Score0
Human Vibe Score1
Inner_Lengthiness697This week

What are Boilerplates?

What are Boilerplates? Boilerplate originally referred to the rolled steel used to make boilers for steam engines in the 19th century. Over time, the term evolved to describe any standardized piece of text or code that can be reused without significant changes. Interest in SaaS has been on the rise, and many more people now want to build products. However, building products from scratch takes a lot of time, and it can be extremely frustrating. Enter SaaS Boilerplates With the standardization of stacks and basic systems that govern SaaS tools, it has become evident that there was a need, and the time was ripe for SaaS Boilerplates. SaaS Boilerplates come with landing pages, website components, authentication modules, payment modules, and various other standard features that can save developers a significant amount of time and cost. The market is flooded with Boilerplates for various tech stacks, such as NextJS, Laravel, Swift, NuxtJS, and so forth. Pros and Cons of Boilerplates Pros Save a significant amount of time and money Reduce frustration for developers as the redundant tasks are taken care of Boilerplates often follow best practices For anywhere between $49 and $299, they provide terrific value for those looking to build something very quickly Most importantly, Boilerplates also enable aspiring founders and builders with limited technical resources or abilities to ship their products faster and more cheaply. They are beacons of hope for non-technical founders looking to build a product quickly. Cons Limited flexibility May become outdated fairly quickly Setting them up still requires time Similar landing pages and design themes can make the product look like a clone Marc Lou’s Shipfast For most of us, Marc Lou popularized the idea of SaaS Boilerplate. Marc Lou launched Shipfast in August 2023. He had built 27 projects prior to this and Shipfast was nothing but all his basic code organised properly. At that time, there were no solid NextJS boilerplates, and Shipfast just took off. He got traction via Product Hunt, Twitter and Hacker News and soon Shipfast went viral. Shipfast now generates $130K/mo, just 9 months after its launch. Marc has been building Shipfast in public, which has led to a lot of interest in SaaS Boilerplates. The market is now flooded with boilerplates for every major tech stack. Marc reaped the benefits of the first mover’s advantage as well as the social proof via his Shipfast community. I don’t think any other boilerplates are as successful as Shipfast, but there are quite a few good ones out there. Shipixen* has grossed over $20K in the 5 months Makerkit* does \~$3500/mo Moreover, there are many open-source boilerplates available for popular stacks such as NextJS. The Evolution of Boilerplates Boilerplates are quickly turning into no-code/low-code code generation tools. For instance, Shipixen allows you to generate custom code for landing pages, waitlist pages and blogs using a simple User Interface. Boilerplates are perfectly posied to sit between code and no-code. Allow the flexibility of code with the interface of a no-code tool — that will be the core value proposition of SaaS boilerplates. Should you build a Boilerplate? Well, the market is flooded, but I believe there’s still an opportunity to leverage boilerplates. You can build boilerplates for certain types of apps or tools, such as Chrome extensions Boilerplates can act as a great lead funnel for building out a great productized services business No-code/low-code code generation boilerplates can become a big thing if you can help build complex tools Niche tech stack boilerplates may still be lucrative Known strategies for successfully building a boilerplate 👇🏻 Shipfast thrives because of social proof and community SaaSRock generates most of its traffic from its Gumroad listings and blogs Usenextbase and Shipixen are being built in public Many boilerplates start with waitlists They have a very clear value proposition around saving time and cost Design & No-Code Boilerplates Here is the corrected version with improved grammar and clarity: While SaaS (code) boilerplates have become fairly popular, other types of boilerplates are emerging in the market, such as design boilerplates and no-code boilerplates. To be honest, design boilerplates have been around for a while. You will find numerous landing page packs, component libraries, and so forth. Makers are now building kits that leverage standard libraries and technologies such as Tailwind CSS, Daisy UI, and more. Nick Buzz from the famous baked.design has this *50 Landing Page Design Kit* in Tailwind CSS & Figma which is wildly popular. Lastly, there is a trend of no-code boilerplates as well. Mohit is building a Bubble Boilerplate for the popular no-code platform — Bubble. All in all, I think that people want to build products and build them fast. Boilerplates help them save a significant amount of time and cost. More importantly, boilerplates are impulse purchases for people who have not shipped but who want to ship. Introducing BuilderKit.ai We have been building AI SaaS tools for quite a while now. 10+ products across text, image, speech, RAG — we have built em all. We figured that it seems easy but actually building these so called AI Wrappers can be time consuming and frustrating — there is a lot of nuance to it. So we built BuidlerKit.ai — a NextJS SaaS Boilerpalte It takes care of everything from landing pages, authentication, dashboarding, emails, SEO to payments — everything that you need to build your tool. It also comes with 8+ production-ready apps. Moreover, the BuilderKit community is an exclusive community of AI SaaS builders (Pro Only Access) The Pre Orders are now live at https://www.builderkit.ai (First 100 Customers get $100 Off — I think we have already done \~20 odd orders since the announcement yesterday, Grab your seat asap!) Starter Plan $49, Pro Plan @ $99

How I Built a $6k/mo Business with Cold Email
reddit
LLM Vibe Score0
Human Vibe Score1
Afraid-Astronomer130This week

How I Built a $6k/mo Business with Cold Email

I scaled my SaaS to a $6k/mo business in under 6 months completely using cold email. However, the biggest takeaway for me is not a business that’s potentially worth 6-figure. It’s having a glance at the power of cold emails in the age of AI. It’s a rapidly evolving yet highly-effective channel, but no one talks about how to do it properly. Below is the what I needed 3 years ago, when I was stuck with 40 free users on my first app. An app I spent 2 years building into the void. Entrepreneurship is lonely. Especially when you are just starting out. Launching a startup feel like shouting into the dark. You pour your heart out. You think you have the next big idea, but no one cares. You write tweets, write blogs, build features, add tests. You talk to some lukewarm leads on Twitter. You do your big launch on Product Hunt. You might even get your first few sales. But after that, crickets... Then, you try every distribution channel out there. SEO Influencers Facebook ads Affiliates Newsletters Social media PPC Tiktok Press releases The reality is, none of them are that effective for early-stage startups. Because, let's face it, when you're just getting started, you have no clue what your customers truly desire. Without understanding their needs, you cannot create a product that resonates with them. It's as simple as that. So what’s the best distribution channel when you are doing a cold start? Cold emails. I know what you're thinking, but give me 10 seconds to change your mind: When I first heard about cold emailing I was like: “Hell no! I’m a developer, ain’t no way I’m talking to strangers.” That all changed on Jan 1st 2024, when I actually started sending cold emails to grow. Over the period of 6 months, I got over 1,700 users to sign up for my SaaS and grew it to a $6k/mo rapidly growing business. All from cold emails. Mastering Cold Emails = Your Superpower I might not recommend cold emails 3 years ago, but in 2024, I'd go all in with it. It used to be an expensive marketing channel bootstrapped startups can’t afford. You need to hire many assistants, build a list, research the leads, find emails, manage the mailboxes, email the leads, reply to emails, do meetings. follow up, get rejected... You had to hire at least 5 people just to get the ball rolling. The problem? Managing people sucks, and it doesn’t scale. That all changed with AI. Today, GPT-4 outperforms most human assistants. You can build an army of intelligent agents to help you complete tasks that’d previously be impossible without human input. Things that’d take a team of 10 assistants a week can now be done in 30 minutes with AI, at far superior quality with less headaches. You can throw 5000 names with website url at this pipeline and you’ll automatically have 5000 personalized emails ready to fire in 30 minutes. How amazing is that? Beyond being extremely accessible to developers who are already proficient in AI, cold email's got 3 superpowers that no other distribution channels can offer. Superpower 1/3 : You start a conversation with every single user. Every. Single. User. Let that sink in. This is incredibly powerful in the early stages, as it helps you establish rapport, bounce ideas off one another, offer 1:1 support, understand their needs, build personal relationships, and ultimately convert users into long-term fans of your product. From talking to 1000 users at the early stage, I had 20 users asking me to get on a call every week. If they are ready to buy, I do a sales call. If they are not sure, I do a user research call. At one point I even had to limit the number of calls I took to avoid burnout. The depth of the understanding of my customers’ needs is unparalleled. Using this insight, I refined the product to precisely cater to their requirements. Superpower 2/3 : You choose exactly who you talk to Unlike other distribution channels where you at best pick what someone's searching for, with cold emails, you have 100% control over who you talk to. Their company Job title Seniority level Number of employees Technology stack Growth rate Funding stage Product offerings Competitive landscape Social activity (Marital status - well, technically you can, but maybe not this one…) You can dial in this targeting to match your ICP exactly. The result is super low CAC and ultra high conversion rate. For example, My competitors are paying $10 per click for the keyword "HARO agency". I pay $0.19 per email sent, and $1.92 per signup At around $500 LTV, you can see how the first means a non-viable business. And the second means a cash-generating engine. Superpower 3/3 : Complete stealth mode Unlike other channels where competitors can easily reverse engineer or even abuse your marketing strategies, cold email operates in complete stealth mode. Every aspect is concealed from end to end: Your target audience Lead generation methods Number of leads targeted Email content Sales funnel This secrecy explains why there isn't much discussion about it online. Everyone is too focused on keeping their strategies close and reaping the rewards. That's precisely why I've chosen to share my insights on leveraging cold email to grow a successful SaaS business. More founders need to harness this channel to its fullest potential. In addition, I've more or less reached every user within my Total Addressable Market (TAM). So, if any competitor is reading this, don't bother trying to replicate it. The majority of potential users for this AI product are already onboard. To recap, the three superpowers of cold emails: You start a conversation with every single user → Accelerate to PMF You choose exactly who you talk to → Super-low CAC Complete stealth mode → Doesn’t attract competition By combining the three superpowers I helped my SaaS reach product-marketing-fit quickly and scale it to $6k per month while staying fully bootstrapped. I don't believe this was a coincidence. It's a replicable strategy for any startup. The blueprint is actually straightforward: Engage with a handful of customers Validate the idea Engage with numerous customers Scale to $5k/mo and beyond More early-stage founders should leverage cold emails for validation, and as their first distribution channel. And what would it do for you? Update: lots of DM asking about more specifics so I wrote about it here. https://coldstartblueprint.com/p/ai-agent-email-list-building

Things I did to promote my product, and how they turned out
reddit
LLM Vibe Score0
Human Vibe Score1
laike9mThis week

Things I did to promote my product, and how they turned out

(I will share more updates in the future, you can find me on Twitter and/or Mastodon) Ask any ten indie developers about the toughest part of their job, and nine will likely say "marketing." I recently got a taste of this firsthand when I launched Xylect. Here's a rundown of my promotional attempts - hopefully, my experiences can help fellow developers out there. Podcast Community (✅ Success) I kicked things off by promoting Xylect in my podcast listener group. It wasn't a blockbuster, but I managed to sell a few copies and got some invaluable feedback from friends. Shoutout to those early supporters! Reddit r/macapps (✅ Success) Having had some luck promoting open-source projects on Reddit before, I decided to make r/macapps my first stop in the English-speaking world. I made an app to help you automate boring tasks with one click This post turned out to be a hit! I sold about ten copies and got a ton of useful feedback. Users pointed out compatibility issues with PopClip and suggested improvements for the website. One Italian user even requested localization, which I happily added. https://preview.redd.it/y4fuwh6hleqd1.png?width=959&format=png&auto=webp&s=7bb1b68cbf8a4f94998999e0832b9b7bd85bac67 https://preview.redd.it/8uu4cmyhleqd1.png?width=683&format=png&auto=webp&s=8f1744636aee8074b0e7491a334ef06076b143b0 I also got an intriguing email from a French user - more on that later. More Reddit Posts (❌ Failure) Riding high on my r/macapps success, I branched out to r/SideProject, r/Entrepreneur, and r/indiehackers. These subreddits frown upon direct self-promotion, so I took a softer approach with an article: The unexpected emotional cost of being an indiehacker While the article was heartfelt, it fell flat. Across all three posts, I got a grand total of three comments - two of which were complaints about the font size on mobile. Needless to say, I didn't sell a single copy. Hacker News (❌ Failure) As one of the tech world's major forums, I had to give Hacker News a shot. I wasn't too optimistic, given my past experiences there. Posting on HN feels like a mix of luck and dark magic. As expected, my post vanished without a trace - no comments, no sales. I might give it another go someday. If you're curious, you can check out my previous HN submissions. Tools Directory Websites (❌ Failure) These sites have a simple premise: you list your app, they display it. Seemed like an easy way to get some backlinks, right? Well, I learned the hard way that it's not that simple. I stumbled upon a Reddit post where someone claimed to have made a killing with their directory site in just a few days. The catch? Each listing cost $19. The site had a handful of apps listed, so I thought, "Why not? Early bird gets the worm." I paid up and listed Xylect. Spoiler alert: all I got was $19 poorer 🥲 Lesson learned: These directory sites won't magically sell your product. At best, they're just glorified backlinks. There might be some value in paid promotions on these platforms, but I can't speak to that from experience. V2EX (❌ Failure) After striking out in the English-speaking world, I turned my attention to the Chinese market, starting with V2EX (think of it as China's hybrid of HN and Reddit). This turned out to be my most unexpected flop. Here's the post: [\[Launch Discount\] Mac's most powerful AI search (Perplexity + Wikipedia + Google), boost your efficiency tenfold with one click. No API key required, no prompt needed, no token limit 🔥 - V2EX](https://www.v2ex.com/t/1064930?p=1#reply36) I'd seen decent engagement on other promo posts, so I had high hopes. I posted late at night (US time) and went to bed dreaming of waking up to a flood of comments. Reality check: The next morning, I had exactly one reply - from Kilerd, a loyal podcast listener showing some love. I was baffled. After re-reading my post, I realized I'd missed a crucial element: promo codes. A quick scan of popular posts confirmed my suspicion. Nearly every successful promo post was offering codes, and most comments were just base64-encoded email addresses. Talk about a facepalm moment. I scrambled to add a note about an upcoming free trial and invited users to drop their emails. This got the ball rolling with some code requests, but by then, the damage was done. The post fizzled out, and I didn't sell a single copy 🫠 A French Friend's Newsletter (✅ Success) At this point, my promotional efforts were looking pretty grim. My sales chart had a depressing stretch of flatline. But then, a glimmer of hope appeared in my inbox. Remember that French user I mentioned earlier? He ran a newsletter called vvmac and offered to feature Xylect if I added French support and sent him a free license. It was an offer I couldn't refuse. What followed was a crash course in French localization (thank you, Claude!) and the start of an incredible partnership. This guy was the most thorough beta tester I've ever encountered. We exchanged over sixty emails, covering everything from translations to UI tweaks to bug fixes. His response time was lightning-fast - I'd fix a bug, and five minutes later, he'd confirm it was sorted. The result? A much-improved Xylect and a glowing feature in his newsletter. https://preview.redd.it/ylcq2wxoleqd1.png?width=991&format=png&auto=webp&s=ee395110f50417d5c7f61318f27bf3dc30247809 I'm still in awe of his dedication. He single-handedly transformed Xylect from a buggy mess into a polished product. I'll be forever grateful for his help. The newsletter feature led to a few more sales, but honestly, that felt like a bonus at that point. Influencers (❌ Failure) I knew from the start that to really make waves, I'd need influencer backing. So, I added a note offering free licenses to content creators willing to collaborate. https://preview.redd.it/tyb2m1rqleqd1.png?width=799&format=png&auto=webp&s=56eabf126e772515322595613c546e6ba69fb431 I did get one taker: Hey, I'll be honest, I am not a huge content creator but I think I put a lot of effort in evaluating and figuring out which apps work... So I was wondering if I could get a license in case you are willing to share it. Thank you for considering. Have a great weekend. But I knew I needed to aim higher. With the new French localization, I thought I'd try my luck with some French-speaking Mac YouTubers. I crafted emails highlighting how Xylect could help their French audience with English content. https://preview.redd.it/07oqzemrleqd1.png?width=542&format=png&auto=webp&s=3d160c1d149f28e9029816a277c6ab2496fcd57e After days of silence, I got one reply. It was... not what I was hoping for: Hi, Thank you for your proposal. I can help you to promote your service on Tiktok, Instagram et YouTube, with unique short video. Price for this project is 3500€. Unless I've completely lost my marbles, there's no way I'm dropping 3500€ on promotion. Sure, given their follower count (YouTube: 348K, TikTok: 2.7M, Instagram: 400K), it's not an outrageous ask. For some products, it might even be worth it. But for Xylect? No way. I also reached out to a Chinese influencer on Xiaohongshu, but they weren't interested. Back to the drawing board. Conclusion If you've made it this far, you've probably realized this isn't exactly a success story. My search for effective promotional channels came up largely empty-handed. I'd naively thought that my success with open-source projects would translate seamlessly to the indie dev world. Boy, was I wrong. As I mentioned in my previous article, open-source projects create a dynamic where users feel indebted to developers for their free labor. But in the commercial world of indie development, that dynamic completely flips. While this experience was often frustrating, it was also enlightening - which was kind of the point. As my first foray into indie development, my main goal was to learn the ropes and understand the process. Making money would've been nice, sure, but it wasn't my primary focus. Thanks for sticking with me through this post. I will share more updates in the future, you can follow me on  Twitter and/or Mastodon.

I’ve built a gaming recommendation and exploration platform called Which Game Next
reddit
LLM Vibe Score0
Human Vibe Score0.714
kasperooThis week

I’ve built a gaming recommendation and exploration platform called Which Game Next

Hello there! Me and a few of my best friends are software engineers, and we’ve been working part-time on developing a side project for the past 12 months. It’s called www.whichgamenext.com, and we’ve recently launched into open beta for everyone to check out. Your feedback would be invaluable to us! Our aim has been to build a gaming recommendation engine, alongside providing market oversight for where you can legally and officially purchase or obtain modern games from multiple stores and/or subscriptions. It’s often difficult to figure out what you have access to if you only have a single specific subscription, like Game Pass PC, or if you’re only interested in games on GOG/Nintendo (what a mix!). We started by identifying the available digital stores and subscriptions and slowly compiling our database using multiple automated services to gather data on these games. Think JustWatch, but for games! One major service we’ve partnered with is IGDB, which has been supplying us with JSON data dumps that served as the initial seed for our game data. A massive thank you to them for their continued support! With the data in place, we’ve been focusing on exploring new features. So far, this has included private and public user-generated lists, personal backlog tracking, and the ability to like or dislike games. We’re now improving our recommendation engine, tackling the complexities that come with it, and having a lot of fun along the way. We’re utilising modern AI strategies and solving fascinating problems related to large-scale data aggregation. We truly can’t wait to share this fantastic work! In addition to this, you can soon expect curated collections, articles about games, and supporting links to help you make informed, unbiased purchasing decisions. Your shared data will drive the recommendations. But it doesn’t stop there—we have plenty of other features on our radar, such as importing games from your favourite stores, syncing your gameplay time, surfacing data like “How Long to Beat,” and creating new and exciting ways to interact with this growing community! This is a passion project created by a group of gamers who want to spend their time and money wisely, without purchasing biases. Since it’s a side project, we mostly work on it at night, but we’re excited to grow the community, share our vision, and, who knows, maybe one day make it our full-time job! Let’s dive into the technical details: • Monorepo architecture: This speeds up development by sharing libraries, living style guides, configs, etc. Nx.js has been brilliant, enabling us to create a dependency graph of changes and only build/deploy what’s modified in a PR. • AWS: We’re using the free tier (with a few exceptions where we pay for smaller services). Achieving self-sufficiency is critical for us. Additionally, we applied to the AWS Startup Foundation programme and received $1,000 in AWS credits, which has been incredibly helpful! • Infrastructure: Fully deployed as code with Terraform. • Backends: Built using Express and Nest.js, split into around 40 projects and counting! Each project plays a unique role in gathering and syncing game data. • Scalability: Designed from the ground up, utilising AWS Lambdas with auto-scaling and load balancing. • Databases: We use Postgres with RDS and DynamoDB for storing various data. • Frontend stack: Built with React, Next.js, Tailwind, Zustand, TanStack Query, Jest, and Storybook. • CI/CD: Managed with GitHub Actions and Amplify hooks for deploying the frontends. • Admin portal: We’ve built a bespoke CMS to control the main website. It synchronises with external services, tracks game data changes, and allows us to selectively apply ‘patches’ from sites like IGDB. The system also includes data override and rollback capabilities, ensuring we maintain control over game data. • Automation: Partially automated, so manual intervention is rarely needed. • Scraping tools: Fully integrated into the admin portal with log trail capabilities. • Cloudflare: Used for on-the-fly image transformations; we’re considering moving to it full-time as our CDN for free WebP conversions. • Authentication: Handled by Cognito, with a custom frontend built from scratch. Key learnings so far: • AWS cold starts: Not ideal! While the platform is still new, we ping endpoints to keep them responsive. This won’t be an issue once traffic increases. • Lambda memory matters: We learned the hard way that low-memory configurations can delay responses by 2-3 seconds. • DynamoDB partition keys: If not designed correctly from the start, you might have to start over (yes, we’ve been there!). • GitHub Actions: Setting up node\_modules cache reuse takes time, but it’s worth it—don’t give up! We don’t know where this project will take us yet, but it’s been a fantastic journey so far. We’ve learned a lot, explored technologies we don’t typically use in our day jobs, and built something we’re genuinely passionate about. Your feedback would mean the world to us. What do you think of what we’ve done so far? What would you like to see added? Is this a service you’d use? Do you see the value in it as we do? Thanks for reading, and we hope to see you in the comments! (or our newly created /r/whichgamenext

Built an AI to stop me from procrastinating on Reddit, it actually spies on my browser tabs & it's kinda freaking me out (but it works)
reddit
LLM Vibe Score0
Human Vibe Score-0.2
sameed_aThis week

Built an AI to stop me from procrastinating on Reddit, it actually spies on my browser tabs & it's kinda freaking me out (but it works)

hey guys, So, I have a problem. A major procrastination problem. You know the type? I start all good, like, "ok, I'm gonna spend the next 2 hrs REALLY researching this specific Reddit thread about optimizing workflow automation for small businesses." (That's literally what I'm supposed to be doing rn, lol) And then... BAM. Suddenly I'm 15 posts deep into r/aww looking at baby sloths, or somehow I've ended up on Wikipedia reading about competitive hot dog eating. It's like my brain has a mind of its own, seriously. I've tried everything. Cold Turkey, Freedom, all those blocker apps. And honestly? They kinda suck. They're so... blunt. Like, "NO REDDIT FOR U!!" But I need Reddit for my actual research! It is my research, ffs. The problem is those apps just see a URL and block it. They don't understand context. They're just digital bouncers, and terrible ones. Total roadblocks, and a complete pain. That's why I got desperate. I even spent, like, 3 solid hrs one night just chatting with an AI cuz I was too embarrassed to admit to my friends how bad I was at staying on track. Pathetic, I know. But that's when it hit me. I needed something that understood what I was supposed to be doing, and then actively, intelligently, stopped me when I got sidetracked. Something that, like, gets that this is what I meant to use, so it blocks other posts or subs. So, I built it. It's a Chrome extension, and it's basically like having a tiny, hyper-observant AI therapist/drill sergeant living in my browser. Here's the freaky part: it actually watches what I'm doing. Like, it learns my specific task. If I tell it I'm researching on Reddit, it lets me use Reddit, but only for that specific research. If I try to sneak off to r/funny or check my notifs, it knows. It's not just blocking URLs; it's analyzing the content of the pages I'm on and comparing it to what I'm supposed to be doing. It even has these lil "achievement" things, which sound cheesy, but seeing "Focused for 90 mins straight!" pop up is weirdly motivating. And it has this brutal feature that shows u, in plain numbers, how much time you've wasted. Ouch. It's been working, which is amazing, and scary at the same time! Like, the scary part is, it feels weird sharing my own edge over procrastination. I mean, if u use my lil tool too! It also kinda gives off that creepy, AI overlord watching my thoughts vibe? Why I'm even posting this: I'm looking for a few (maybe 5?) people who are as desperate as I was. People who've tried every productivity hack, app, and technique, and are still staring at the ceiling at 3 am, filled with regret. If this sounds familiar, DM "DM me". Tell me your worst procrastination story. The winner (loser?) gets a copy. I need honest, brutally honest, feedback. Does this actually work for anyone else, or am I just fooling myself? Edit: shared the extension with some of you, and for others you can give it a spin here i made it live to the chrome store: https://getfocusai.com/

My humble analysis on how @levelsio grew PhotoAI to 155K/m
reddit
LLM Vibe Score0
Human Vibe Score1
PodcastSummaryGuyThis week

My humble analysis on how @levelsio grew PhotoAI to 155K/m

In 2023,@levelsiol aunched PhotoAi and nailed perfect timing with the AI-generated photo trend He saw the rise of AI Image Generators as early as Jan 2023—when search volume was just 246K. Fast forward to now, that’s exploded to 1M+ searches/month! @levelsio didn’t guess—he tracked this exponential growth, built fast, and launched a product that people needed before they even realized it themselves. That’s how you go from idea to$155K/month in revenue. But why does this work? Demand Explosion: The need for AI tools like PhotoAI surged as more creators, marketers, and businesses searched for automated ways to generate images (graph 1 shows this insane growth curve). Timing is everything: When trends move this fast, you have to act quicker. He launched at the right time, capturing the early adopters and setting the foundation for sustainable revenue (graph 2 showcases the clear spike). Build based on data, not assumptions: The growth of AI Character Generators also hit major traction (graph 3). Both trends signaled an opportunity, and he jumped in before the market was saturated. Lesson: Don’t guess what people want—watch the trends and build fast. Trends show you what’s working, before the mainstream even knows it. This is how you launch products that solve real problems. Check out the graphs for context: AI Image Generator Growth \(graph 1\) Search volume spiking \(graph 2\) AI Character Generators \(graph 3\) Use data, not just intuition. Track the trends and execute. Simple but powerful. I hope you liked this thread.

[P] [R] sANNd: A New Neural Network Framework Using Trainable Iterators
reddit
LLM Vibe Score0
Human Vibe Score1
JackRipperVAThis week

[P] [R] sANNd: A New Neural Network Framework Using Trainable Iterators

sANNd sANNd is a lightweight, modular neural network library designed as a sandbox for experimenting with new ideas in artificial intelligence. The Mould Class: A Pythonic Building Block The Mould class is a core component of sANNd. It provides a Pythonic way to apply functions to data that’s bundled inside objects: Encapsulated Variables: Each Mould object holds a set of variables (for example, weights or parameters) inside it. This means related data is kept together in one place (the object), making the code organized and intuitive. Static Functions: A Mould class defines its operation as a static method – essentially a function that isn’t tied to a specific instance. This static function takes in inputs (and possibly other Mould objects’ variables) and produces an output. In simple terms, the Mould’s static method describes how to transform input data using the Mould’s internal variables. Pythonic Usage: Using static methods in this way is a clean, Pythonic design. You call the Mould’s function through the class, but it applies to the data in the object. This approach lets you clearly separate what the operation is (the logic in the static function) from which data it uses (the variables inside the Mould instance). Example: Imagine a Mould class called LinearMould that has a static function to compute a linear transformation (like y = W*x + b). An instance of LinearMould would hold specific W and b values, and you’d use the static method to apply that linear formula to an input. This gives you the convenience of object-oriented design (encapsulating W and b) with the clarity of a standalone function defining the math. Chaining Moulds for Complex Computations Moulds become even more powerful when you chain them together. You can connect multiple Moulds so that the output of one becomes the input of the next: Sequential Operations: Just like stacking layers in a neural network, you can place Moulds in sequence. For example, you might take the output from LinearMouldA and feed it into LinearMouldB. In code, this might look as simple as using the output of one call as the argument to the next. The design of sANNd makes this straightforward – the static function of each Mould knows how to handle the data coming in. Building Pipelines: By chaining Moulds, you create a pipeline of transformations. Each Mould handles one step of computation, and together they produce a final result. This could represent a multi-layer neural network, a data processing pipeline, or any custom sequence of operations you need. There’s no strict limit to how you can chain them; you have the freedom to combine Moulds in any order that makes sense for your experiment. Clarity and Modularity: Because each Mould is a self-contained piece (with its variables and function), chaining them doesn’t turn your code into a black box. You can inspect or modify any part of the chain easily. This modular design means you can insert, remove, or replace Moulds to see how it affects the overall computation, which is great for experimentation. Implicit Backward Path (Automatic Backpropagation) One major benefit of using chained Moulds is that they implicitly define the backward path for training with gradient descent (backpropagation): Automatic Gradient Flow: When you connect Moulds in a sequence for a forward pass (input → Mould A → Mould B → output), you’ve essentially defined a computation graph. sANNd uses this graph to handle the reverse computation automatically. In other words, if you calculate an error or loss based on the final output, sANNd can propagate that error backwards through each Mould in the chain. No Manual Backprop: You do not need to manually code how gradients flow through each Mould. The way you set up the Moulds’ static functions already determines how outputs depend on inputs and internal variables. sANNd leverages that to perform backpropagation. This is similar in spirit to how libraries like PyTorch/TF do “autograd,” but here it’s a natural result of the Mould chain architecture. Gradient Descent Ready: Because the backward path is established by the forward connections, you can apply gradient descent optimizations out of the box. For instance, you can adjust the weights inside each Mould based on the computed gradients to minimize your loss. The design ensures that each Mould’s contribution to the final error is tracked, so all parts of your model learn appropriately during training. In short, defining your model with Moulds means you get training capability for free. You focus on describing the forward computations, and sANNd handles the math behind learning from errors. Comparing sANNd to Traditional Frameworks sANNd’s approach is quite different from traditional Python-based neural network frameworks. Here’s how it stacks up against frameworks like TensorFlow, PyTorch, or Keras in terms of approach, flexibility, and intended use: Design Approach: Traditional frameworks use predefined layer classes and often build a computation graph behind the scenes. For example, Keras might have a Dense layer class, and TensorFlow might construct a static graph (in TF1) or use eager execution (in TF2). sANNd takes a simpler approach – it uses plain Python classes and static functions (Moulds) to define computations. There’s no need to learn a new graph syntax or decorators; if you know Python functions and classes, you can read and write sANNd models. This makes the internal workings more transparent and easier to follow. Flexibility: While frameworks like PyTorch and TensorFlow are very powerful, they can introduce a lot of boilerplate and assume you’re building typical architectures. sANNd is extremely modular and flexible. You aren’t limited to the layers someone else defined – you can create any operation you want as a Mould. Want to experiment with a novel activation function or a custom recurrent connection? Just define it in a Mould. There’s less magic and abstraction obscuring your code, so unconventional model structures are easier to implement. (Of course, major frameworks can also be extended, but sANNd makes this feel more natural by staying within standard Python paradigms.) Intended Use: sANNd is intended for experimentation and research. It’s like a toolkit for tinkering. You get fine-grained control over every part of the network, which is ideal for trying out bold new ideas that don’t fit the mold of common deep learning models. In contrast, TensorFlow/PyTorch shine in production environments and large-scale training – they are optimized (GPU support, highly efficient tensor operations) and come with many utilities for things like data loading, distributed training, etc. sANNd doesn’t aim to replace them for those heavy-lifting tasks. Instead, it’s meant for when you need a lighter, more interpretable setup to prototype concepts. You might use sANNd to prove out a concept or test a hypothesis in AI research, and later switch to a bigger framework if you need to scale it up. Simplicity vs. Complexity: By design, sANNd keeps things simple. The trade-off is that it might not have the raw performance optimizations of the large frameworks. However, this simplicity is a feature – it means the code is easier to understand and modify. For many research scenarios, being able to quickly tweak an idea is more important than squeezing out maximum speed. Traditional frameworks, with their complexity, can sometimes be harder to adapt for radically different ideas (you might find yourself fighting the framework). With sANNd, the framework gets out of your way as much as possible. Modular and Experimental by Nature One of the driving philosophies of sANNd is to be modular and experimental, to further ML research: Modularity: sANNd is built from small, composable pieces. The Mould class is one such piece, and you can imagine building additional components in a similar spirit. This modular design means you can re-use components, mix and match them, or replace one implementation with another without affecting the rest of your system. It’s like having a box of building blocks for neural networks – you can assemble them in standard ways or in completely novel configurations. Experimentation Friendly: Because it avoids heavy abstraction, sANNd lets you directly see and control what’s happening at each step. This is great for research, where you might need to observe intermediate results, inject custom behavior, or adjust the learning process on the fly. sANNd’s straightforward structure (Python objects and functions) makes such interventions possible. You’re not constrained to a fixed training loop or forced to use certain layer types. True Intelligence Research: Achieving “True Intelligence” (often related to artificial general intelligence or other forms of broader AI) may require going beyond the usual neural network designs. sANNd aims to be a playground for these ideas. Its flexibility allows researchers to integrate unconventional elements — be it new memory structures, dynamic connection patterns, or hybrid models that combine symbolic and neural approaches. You can use sANNd to prototype these offbeat ideas quickly. In essence, it’s easier to test “what if we try this?” scenarios with sANNd than with more rigid frameworks. In summary, sANNd’s unique Mould class and design philosophy offer a fresh take on building neural networks. It emphasizes clarity, composability, and flexibility, allowing you to focus on creativity and understanding. Whether you’re stacking simple Moulds into a deep model, or inventing a completely new form of network, sANNd provides a friendly foundation. It’s not here to dethrone TensorFlow or PyTorch in industry applications – instead, it’s here to give researchers and enthusiasts a more malleable tool for exploring the frontiers of AI. Enjoy using sANNd as your neural network sandbox, and happy experimenting!

[D] Misuse of Deep Learning in Nature Journal’s Earthquake Aftershock Paper
reddit
LLM Vibe Score0
Human Vibe Score0.333
milaworldThis week

[D] Misuse of Deep Learning in Nature Journal’s Earthquake Aftershock Paper

Recently, I saw a post by Rajiv Shah, Chicago-based data-scientist, regarding an article published in Nature last year called Deep learning of aftershock patterns following large earthquakes, written by scientists at Harvard in collaboration with Google. Below is the article: Stand Up for Best Practices: Misuse of Deep Learning in Nature’s Earthquake Aftershock Paper The Dangers of Machine Learning Hype Practitioners of AI, machine learning, predictive modeling, and data science have grown enormously over the last few years. What was once a niche field defined by its blend of knowledge is becoming a rapidly growing profession. As the excitement around AI continues to grow, the new wave of ML augmentation, automation, and GUI tools will lead to even more growth in the number of people trying to build predictive models. But here’s the rub: While it becomes easier to use the tools of predictive modeling, predictive modeling knowledge is not yet a widespread commodity. Errors can be counterintuitive and subtle, and they can easily lead you to the wrong conclusions if you’re not careful. I’m a data scientist who works with dozens of expert data science teams for a living. In my day job, I see these teams striving to build high-quality models. The best teams work together to review their models to detect problems. There are many hard-to-detect-ways that lead to problematic models (say, by allowing target leakage into their training data). Identifying issues is not fun. This requires admitting that exciting results are “too good to be true” or that their methods were not the right approach. In other words, it’s less about the sexy data science hype that gets headlines and more about a rigorous scientific discipline. Bad Methods Create Bad Results Almost a year ago, I read an article in Nature that claimed unprecedented accuracy in predicting earthquake aftershocks by using deep learning. Reading the article, my internal radar became deeply suspicious of their results. Their methods simply didn’t carry many of the hallmarks of careful predicting modeling. I started to dig deeper. In the meantime, this article blew up and became widely recognized! It was even included in the release notes for Tensorflow as an example of what deep learning could do. However, in my digging, I found major flaws in the paper. Namely, data leakage which leads to unrealistic accuracy scores and a lack of attention to model selection (you don’t build a 6 layer neural network when a simpler model provides the same level of accuracy). To my earlier point: these are subtle, but incredibly basic predictive modeling errors that can invalidate the entire results of an experiment. Data scientists are trained to recognize and avoid these issues in their work. I assumed that this was simply overlooked by the author, so I contacted her and let her know so that she could improve her analysis. Although we had previously communicated, she did not respond to my email over concerns with the paper. Falling On Deaf Ears So, what was I to do? My coworkers told me to just tweet it and let it go, but I wanted to stand up for good modeling practices. I thought reason and best practices would prevail, so I started a 6-month process of writing up my results and shared them with Nature. Upon sharing my results, I received a note from Nature in January 2019 that despite serious concerns about data leakage and model selection that invalidate their experiment, they saw no need to correct the errors, because “Devries et al. are concerned primarily with using machine learning as [a] tool to extract insight into the natural world, and not with details of the algorithm design.” The authors provided a much harsher response. You can read the entire exchange on my github. It’s not enough to say that I was disappointed. This was a major paper (it’s Nature!) that bought into AI hype and published a paper despite it using flawed methods. Then, just this week, I ran across articles by Arnaud Mignan and Marco Broccardo on shortcomings that they found in the aftershocks article. Here are two more data scientists with expertise in earthquake analysis who also noticed flaws in the paper. I also have placed my analysis and reproducible code on github. Standing Up For Predictive Modeling Methods I want to make it clear: my goal is not to villainize the authors of the aftershocks paper. I don’t believe that they were malicious, and I think that they would argue their goal was to just show how machine learning could be applied to aftershocks. Devries is an accomplished earthquake scientist who wanted to use the latest methods for her field of study and found exciting results from it. But here’s the problem: their insights and results were based on fundamentally flawed methods. It’s not enough to say, “This isn’t a machine learning paper, it’s an earthquake paper.” If you use predictive modeling, then the quality of your results are determined by the quality of your modeling. Your work becomes data science work, and you are on the hook for your scientific rigor. There is a huge appetite for papers that use the latest technologies and approaches. It becomes very difficult to push back on these papers. But if we allow papers or projects with fundamental issues to advance, it hurts all of us. It undermines the field of predictive modeling. Please push back on bad data science. Report bad findings to papers. And if they don’t take action, go to twitter, post about it, share your results and make noise. This type of collective action worked to raise awareness of p-values and combat the epidemic of p-hacking. We need good machine learning practices if we want our field to continue to grow and maintain credibility. Link to Rajiv's Article Original Nature Publication (note: paywalled) GitHub repo contains an attempt to reproduce Nature's paper Confrontational correspondence with authors

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.
reddit
LLM Vibe Score0
Human Vibe Score0.6
AlexSnakeKingThis week

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.

TD;LR: At Company A, Team X does advanced analytics using on-prem ERP tools and older programming languages. Their tools work very well and are designed based on very deep business and domain expertise. Team Y is a new and ambitious Data Science team that thinks they can replace Team X's tools with a bunch of R scripts and a custom built ML platform. Their models are simplistic, but more "fashionable" compared to the econometric models used by Team X, and team Y benefits from the ML/DS moniker so leadership is allowing Team Y to start a large scale overhaul of the analytics platform in question. Team Y doesn't have the experience for such a larger scale transformation, and is refusing to collaborate with team X. This project is very likely going to fail, and cause serious harm to the company as a whole financially and from a people perspective. I argue that this is not just because of bad leadership, but also because of various trends and mindsets in the DS community at large. Update (Jump to below the line for the original story): Several people in the comments are pointing out that this just a management failure, not something due to ML/DS, and that you can replace DS with any buzz tech and the story will still be relevant. My response: Of course, any failure at an organization level is ultimately a management failure one way or the other. Moreover, it is also the case that ML/DS when done correctly, will always improve a company's bottom line. There is no scenario where the proper ML solution, delivered at a reasonable cost and in a timely fashion, will somehow hurt the company's bottom line. My point is that in this case management is failing because of certain trends and practices that are specific to the ML/DS community, namely: The idea that DS teams should operate independently of tech and business orgs -- too much autonomy for DS teams The disregard for domain knowledge that seems prevalent nowadays thanks to the ML hype, that DS can be generalists and someone with good enough ML chops can solve any business problem. That wasn't the case when I first left academia for the industry in 2009 (back then nobody would even bother with a phone screen if you didn't have the right domain knowledge). Over reliance on resources who check all the ML hype related boxes (knows Python, R, Tensorflow, Shiny, etc..., has the right Coursera certifications, has blogged on the topic, etc...), but are lacking in depth of experience. DS interviews nowadays all seem to be: Can you tell me what a p-value is? What is elastic net regression? Show me how to fit a model in sklearn? How do you impute NAs in an R dataframe? Any smart person can look those up on Stackoverflow or Cross-Validated,.....Instead teams should be asking stuff like: why does portfolio optimization use QP not LP? How does a forecast influence a customer service level? When should a recommendation engine be content based and when should it use collaborative filtering? etc... (This is a true story, happening to the company I currently work for. Names, domains, algorithms, and roles have been shuffled around to protect my anonymity)  Company A has been around for several decades. It is not the biggest name in its domain, but it is a well respected one. Risk analysis and portfolio optimization have been a core of Company A's business since the 90s. They have a large team of 30 or so analysts who perform those tasks on a daily basis. These analysts use ERP solutions implemented for them by one the big ERP companies (SAP, Teradata, Oracle, JD Edwards,...) or one of the major tech consulting companies (Deloitte, Accenture, PWC, Capgemini, etc...) in collaboration with their own in house engineering team. The tools used are embarrassingly old school: Classic RDBMS running on on-prem servers or maybe even on mainframes, code written in COBOL, Fortran, weird proprietary stuff like ABAP or SPSS.....you get the picture. But the models and analytic functions were pretty sophisticated, and surprisingly cutting edge compared to the published academic literature. Most of all, they fit well with the company's enterprise ecosystem, and were honed based on years of deep domain knowledge.  They have a tech team of several engineers (poached from the aforementioned software and consulting companies) and product managers (who came from the experienced pools of analysts and managers who use the software, or poached from business rivals) maintaining and running this software. Their technology might be old school, but collectively, they know the domain and the company's overall architecture very, very well. They've guided the company through several large scale upgrades and migrations and they have a track record of delivering on time, without too much overhead. The few times they've stumbled, they knew how to pick themselves up very quickly. In fact within their industry niche, they have a reputation for their expertise, and have very good relations with the various vendors they've had to deal with. They were the launching pad of several successful ERP consulting careers.  Interestingly, despite dealing on a daily basis with statistical modeling and optimization algorithms, none of the analysts, engineers, or product managers involved describe themselves as data scientists or machine learning experts. It is mostly a cultural thing: Their expertise predates the Data Science/ML hype that started circa 2010, and they got most of their chops using proprietary enterprise tools instead of the open source tools popular nowadays. A few of them have formal statistical training, but most of them came from engineering or domain backgrounds and learned stats on the fly while doing their job. Call this team "Team X".  Sometime around the mid 2010s, Company A started having some serious anxiety issues: Although still doing very well for a company its size, overall economic and demographic trends were shrinking its customer base, and a couple of so called disruptors came up with a new app and business model that started seriously eating into their revenue. A suitable reaction to appease shareholders and Wall Street was necessary. The company already had a decent website and a pretty snazzy app, what more could be done? Leadership decided that it was high time that AI and ML become a core part of the company's business. An ambitious Manager, with no science or engineering background, but who had very briefly toyed with a recommender system a couple of years back, was chosen to build a data science team, call it team "Y" (he had a bachelor's in history from the local state college and worked for several years in the company's marketing org). Team "Y" consists mostly of internal hires who decided they wanted to be data scientists and completed a Coursera certification or a Galvanize boot camp, before being brought on to the team, along with a few of fresh Ph.D or M.Sc holders who didn't like academia and wanted to try their hand at an industry role. All of them were very bright people, they could write great Medium blog posts and give inspiring TED talks, but collectively they had very little real world industry experience. As is the fashion nowadays, this group was made part of a data science org that reported directly to the CEO and Board, bypassing the CIO and any tech or business VPs, since Company A wanted to claim the monikers "data driven" and "AI powered" in their upcoming shareholder meetings. In 3 or 4 years of existence, team Y produced a few Python and R scripts. Their architectural experience  consisted almost entirely in connecting Flask to S3 buckets or Redshift tables, with a couple of the more resourceful ones learning how to plug their models into Tableau or how to spin up a Kuberneties pod.  But they needn't worry: The aforementioned manager, who was now a director (and was also doing an online Masters to make up for his qualifications gap and bolster his chances of becoming VP soon - at least he now understands what L1 regularization is), was a master at playing corporate politics and self-promotion. No matter how few actionable insights team Y produced or how little code they deployed to production, he always had their back and made sure they had ample funding. In fact he now had grandiose plans for setting up an all-purpose machine learning platform that can be used to solve all of the company's data problems.  A couple of sharp minded members of team Y, upon googling their industry name along with the word "data science", realized that risk analysis was a prime candidate for being solved with Bayesian models, and there was already a nifty R package for doing just that, whose tutorial they went through on R-Bloggers.com. One of them had even submitted a Bayesian classifier Kernel for a competition on Kaggle (he was 203rd on the leaderboard), and was eager to put his new-found expertise to use on a real world problem. They pitched the idea to their director, who saw a perfect use case for his upcoming ML platform. They started work on it immediately, without bothering to check whether anybody at Company A was already doing risk analysis. Since their org was independent, they didn't really need to check with anybody else before they got funding for their initiative. Although it was basically a Naive Bayes classifier, the term ML was added to the project tile, to impress the board.  As they progressed with their work however, tensions started to build. They had asked the data warehousing and CA analytics teams to build pipelines for them, and word eventually got out to team X about their project. Team X was initially thrilled: They offered to collaborate whole heartedly, and would have loved to add an ML based feather to their already impressive cap. The product owners and analysts were totally onboard as well: They saw a chance to get in on the whole Data Science hype that they kept hearing about. But through some weird mix of arrogance and insecurity, team Y refused to collaborate with them or share any of their long term goals with them, even as they went to other parts of the company giving brown bag presentations and tutorials on the new model they created.  Team X got resentful: from what they saw of team Y's model, their approach was hopelessly naive and had little chances of scaling or being sustainable in production, and they knew exactly how to help with that. Deploying the model to production would have taken them a few days, given how comfortable they were with DevOps and continuous delivery (team Y had taken several months to figure out how to deploy a simple R script to production). And despite how old school their own tech was, team X were crafty enough to be able to plug it in to their existing architecture. Moreover, the output of the model was such that it didn't take into account how the business will consume it or how it was going to be fed to downstream systems, and the product owners could have gone a long way in making the model more amenable to adoption by the business stakeholders. But team Y wouldn't listen, and their leads brushed off any attempts at communication, let alone collaboration. The vibe that team Y was giving off was "We are the cutting edge ML team, you guys are the legacy server grunts. We don't need your opinion.", and they seemed to have a complete disregard for domain knowledge, or worse, they thought that all that domain knowledge consisted of was being able to grasp the definitions of a few business metrics.  Team X got frustrated and tried to express their concerns to leadership. But despite owning a vital link in Company A's business process, they were only \~50 people in a large 1000 strong technology and operations org, and they were several layers removed from the C-suite, so it was impossible for them to get their voices heard.  Meanwhile, the unstoppable director was doing what he did best: Playing corporate politics. Despite how little his team had actually delivered, he had convinced the board that all analysis and optimization tasks should now be migrated to his yet to be delivered ML platform. Since most leaders now knew that there was overlap between team Y and team X's objectives, his pitch was no longer that team Y was going to create a new insight, but that they were going to replace (or modernize) the legacy statistics based on-prem tools with more accurate cloud based ML tools. Never mind that there was no support in the academic literature for the idea that Naive Bayes works better than the Econometric approaches used by team X, let alone the additional wacky idea that Bayesian Optimization would definitely outperform the QP solvers that were running in production.  Unbeknownst to team X, the original Bayesian risk analysis project has now grown into a multimillion dollar major overhaul initiative, which included the eventual replacement of all of the tools and functions supported by team X along with the necessary migration to the cloud. The CIO and a couple of business VPs are on now board, and tech leadership is treating it as a done deal. An outside vendor, a startup who nobody had heard of, was contracted to help build the platform, since team Y has no engineering skills. The choice was deliberate, as calling on any of the established consulting or software companies would have eventually led leadership to the conclusion that team X was better suited for a transformation on this scale than team Y.  Team Y has no experience with any major ERP deployments, and no domain knowledge, yet they are being tasked with fundamentally changing the business process that is at the core of Company A's business. Their models actually perform worse than those deployed by team X, and their architecture is hopelessly simplistic, compared to what is necessary for running such a solution in production.  Ironically, using Bayesian thinking and based on all the evidence, the likelihood that team Y succeeds is close to 0%. At best, the project is going to end up being a write off of 50 million dollars or more. Once the !@#$!@hits the fan, a couple of executive heads are going to role, and dozens of people will get laid off. At worst, given how vital risk analysis and portfolio optimization is to Company A's revenue stream, the failure will eventually sink the whole company. It probably won't go bankrupt, but it will lose a significant portion of its business and work force. Failed ERP implementations can and do sink large companies: Just see what happened to National Grid US, SuperValu or Target Canada.  One might argue that this is more about corporate disfunction and bad leadership than about data science and AI. But I disagree. I think the core driver of this debacle is indeed the blind faith in Data Scientists, ML models and the promise of AI, and the overall culture of hype and self promotion that is very common among the ML crowd.  We haven't seen the end of this story: I sincerely hope that this ends well for the sake of my colleagues and all involved. Company A is a good company, and both its customers and its employees deserver better. But the chances of that happening are negligible given all the information available, and this failure will hit my company hard.

[N] Netflix and European Space Agency no longer working with Siraj Raval
reddit
LLM Vibe Score0
Human Vibe Score0
inarrearsThis week

[N] Netflix and European Space Agency no longer working with Siraj Raval

According to article in The Register: A Netflix spokesperson confirmed to The Register it wasn’t working with Raval, and the ESA has cancelled the whole workshop altogether. “The situation is as it is. The workshop is cancelled, and that’s all,” Guillaume Belanger, an astrophysicist and the INTEGRAL Science Operations Coordinator at the ESA, told The Register on Monday. Raval isn’t about to quit his work any time soon, however. He promised students who graduated from his course that they would be referred to recruiters at Nvidia, Intel, Google and Amazon for engineering positions, or matched with a startup co-founder or a consulting client. In an unlisted YouTube video recorded live for his students discussing week eight of his course, and seen by El Reg, he read out a question posed to him: “Will your referrals hold any value now?” “Um, yeah they’re going to hold value. I don’t see why they wouldn’t. I mean, yes, some people on Twitter were angry but that has nothing to do with… I mean… I’ve also had tons of support, you know. I’ve had tons of support from people, who, uh, you know, support me, who work at these companies. He continues to justify his actions: “Public figures called me in private to remind me that this happens. You know, people make mistakes. You just have to keep going. They’re basically just telling me to not to stop. Of course, you make mistakes but you just keep going,” he claimed. When The Register asked Raval for comment, he responded: I've hardly taken any time off to relax since I first started my YouTube channel almost four years ago. And despite the enormous amount of work it takes to release two high quality videos a week for my audience, I progressively started to take on multiple other projects simultaneously by myself – a book, a docu-series, podcasts, YouTube videos, the course, the school of AI. Basically, these past few weeks, I've been experiencing a burnout unlike anything I've felt before. As a result, all of my output has been subpar. I made the [neural qubits] video and paper in one week. I remember wishing I had three to six months to really dive into quantum machine-learning and make something awesome, but telling myself I couldn't take that long as it would hinder my other projects. I plagiarized large chunks of the paper to meet my self-imposed one-week deadline. The associated video with animations took a lot more work to make. I didn't expect the paper to be cited as serious research, I considered it an additional reading resource for people who enjoyed the associated video to learn more about quantum machine learning. If I had a second chance, I'd definitely take way more time to write the paper, and in my own words. I've given refunds to every student who's asked so far, and the majority of students are still enrolled in the course. There are many happy students, they're just not as vocal on social media. We're on week 8 of 10 of my course, fully committed to student success. “And, no, I haven't plagiarized research for any other paper,” he added. https://www.theregister.co.uk/2019/10/14/ravelaiyoutube/

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[D] AI Agents: too early, too expensive, too unreliable
reddit
LLM Vibe Score0
Human Vibe Score1
madredditscientistThis week

[D] AI Agents: too early, too expensive, too unreliable

Reference: Full blog post There has been a lot of hype about the promise of autonomous agent-based LLM workflows. By now, all major LLMs are capable of interacting with external tools and functions, letting the LLM perform sequences of tasks automatically. But reality is proving more challenging than anticipated. The WebArena leaderboard, which benchmarks LLMs agents against real-world tasks, shows that even the best-performing models have a success rate of only 35.8%. Challenges in Practice After seeing many attempts to AI agents, I believe it's too early, too expensive, too slow, too unreliable. It feels like many AI agent startups are waiting for a model breakthrough that will start the race to productize agents. Reliability: As we all know, LLMs are prone to hallucinations and inconsistencies. Chaining multiple AI steps compounds these issues, especially for tasks requiring exact outputs. Performance and costs: GPT-4o, Gemini-1.5, and Claude Opus are working quite well with tool usage/function calling, but they are still slow and expensive, particularly if you need to do loops and automatic retries. Legal concerns: Companies may be held liable for the mistakes of their agents. A recent example is Air Canada being ordered to pay a customer who was misled by the airline's chatbot. User trust: The "black box" nature of AI agents and stories like the above makes it hard for users to understand and trust their outputs. Gaining user trust for sensitive tasks involving payments or personal information will be hard (paying bills, shopping, etc.). Real-World Attempts Several startups are tackling the AI agent space, but most are still experimental or invite-only: adept.ai - $350M funding, but access is still very limited MultiOn - funding unknown, their API-first approach seems promising HypeWrite - $2.8M funding, started with an AI writing assistant and expanded into the agent space minion.ai - created some initial buzz but has gone quiet now, waitlist only Only MultiOn seems to be pursuing the "give it instructions and watch it go" approach, which is more in line with the promise of AI agents. All others are going down the record-and-replay RPA route, which may be necessary for reliability at this stage. Large players are also bringing AI capabilities to desktops and browsers, and it looks like we'll get native AI integrations on a system level: OpenAI announced their Mac desktop app that can interact with the OS screen. At Google I/O, Google demonstrated Gemini automatically processing a shopping return. Microsoft announced Copilot Studio, which will let developers build AI agent bots. Screenshot Screenshot These tech demos are impressive, but we'll see how well these agent capabilities will work when released publicly and tested against real-world scenarios instead of hand-picked demo cases. The Path Forward AI agents overhyped and it's too early. However, the underlying models continue to advance quickly, and we can expect to see more successful real-world applications. Instead of trying to have one large general purpose agent that is hard to control and test, we can use many smaller agents that basically just pick the right strategy for a specific sub-task in our workflows. These "agents" can be thought of as medium-sized LLM prompts with a) context and b) a set of functions available to call. The most promising path forward likely looks like this: Narrowly scoped, well testable automations that use AI as an augmentation tool rather than pursuing full autonomy Human-in-the-loop approaches that keep humans involved for oversight and handling edge cases Setting realistic expectations about current capabilities and limitations By combining tightly constrained agents, good evaluation data, human-in-the-loop oversight, and traditional engineering methods, we can achieve reliably good results for automating medium-complex tasks. Will AI agents automate tedious repetitive work, such as web scraping, form filling, and data entry? Yes, absolutely. Will AI agents autonomously book your vacation without your intervention? Unlikely, at least in the near future.

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[D] "Grokking" Deep Learning architectures and using them in practice
reddit
LLM Vibe Score0
Human Vibe Score1
LightGreenSquashThis week

[D] "Grokking" Deep Learning architectures and using them in practice

Hi all, I'm on the first years of my PhD in Computer Vision and obviously the vast majority of research in it is nowadays using Deep Learning techniques. I like to think that I'm far from an absolute beginner in the sense that: I've trained neural networks and more "traditional" ML models in a couple of courses, as well as for my MSc thesis, albeit almost out-of-the-box stuff. I have a decent understanding of Linear Algebra, Calculus and Probability Theory (undergrad courses from CS degree). I say "decent" because I'm of the firm opinion that the more math one knows the more impressive the things they can do in AI, so I really don't consider myself a math whiz, but judging from the math knowledge an average "How to get started with Deep Learning" blog post assumes, I'd say I'm well ahead. I'm also devoting some time every day to a more rigorous study of these areas, eventually hoping to expand to other related ones. I can get through Deep Learning papers and usually* obtain at least a basic understanding of what they're about, as well as why it works, at least according to the authors and their experiments. I do still have some trouble with more state-of-the-art works, especially ones that also use things from NLP. However, I don't really feel confident that I can actually produce useful research that investigates and/or uses this sort of methods to do something new. During undergrad, in order to actually understand most -if not all- concepts taught to me in programming and math I'd actually do things with them: solve problems, prove statements, or just code with the goal of creating some system or seeing how an idea actually works (e.g. polymorphism). I realize, however, that this has not been the case with Deep Learning, at least for me: I've never tried to actually code a CNN or ResNet, much less a word2vec model, a Transformer, or any sort of generative model. Sure, I've read about how the first layers of a CNN learn edges etc. but I've never actually "seen it with my own eyes". Transformers in particular seem to really trouble me. Although I sort-of understand the idea behind attention etc., I struggle to see what sort of features they end up using (in contrast to CNNs, where the idea of learning convolutional filters is much more intuitive to me). Which brings me to the question of what's an efficient way to go from understanding a paper to actually feeling like you really, truly, "grok" the material and could build on it, or use it in some scenario? Do you think implementing research papers from scratch or almost from scratch can be useful? Or is it way too time consuming for someone already busy with a PhD? Is it even feasible or are most papers -sadly- unreproducible if you don't use authors' code? How do you manage to stay on track with such a rapidly evolving field, on any level beyond a completely surface understanding? How do you find a good balance between learning to use tools/frameworks, reading papers and gaining the deeper sort of understanding I mention?

[R] Reinforcement Learning for Sequential Decision and Optimal Control
reddit
LLM Vibe Score0
Human Vibe Score1
isfjzzzThis week

[R] Reinforcement Learning for Sequential Decision and Optimal Control

Since early 21st century, artificial intelligence (AI) has been reshaping almost all areas of human society, which has high potential to spark the fourth industrial revolution. Notable examples can be found in the sector of road transportation, where AI has drastically changed automobile design and traffic management. Many new technologies, such as driver assistance, autonomous driving, and cloud-based cooperation, are emerging at an unbelievable speed. These new technologies have the potential to significantly improve driving ability, reduce traffic accidents, and relieve urban congestion. As one of the most important AI branches, reinforcement learning (RL) has attracted increasing attention in recent years. RL is an interdisciplinary field of trial-and-error learning and optimal control, which promises to provide optimal solutions for decision-making or control in large-scale and complex dynamic processes. One of its most conspicuous successes is AlphaGo from Google DeepMind, which has beaten the highest-level professional human player. The underlying key technology is the so-called deep reinforcement learning, which equips AlphaGo with amazing self-evolution ability and high playing intelligence. Despite a few successes, the application of RL is still in its infancy because most RL algorithms are rather difficult to comprehend and implement. RL connects deeply with statistical learning and convex optimization, and involves a wide range of new concepts and theories. As a beginner, one must undergo a long and tedious learning process to become an RL master. Without fully understanding those underlying principles, it is very difficult for new users to make proper adjustments to achieve the best application performance. ​ https://preview.redd.it/tggt6o3o481c1.jpg?width=248&format=pjpg&auto=webp&s=75e2b58ac8da9273f2511a4fe37ef508d86a6e96 Reference: Shengbo Eben Li, Reinforcement Learning for Sequential Decision and Optimal Control. Springer Verlag, Singapore, 2023 Website of e-book: https://link.springer.com/book/10.1007/978-981-19-7784-8 ​ QR code to Springer Book contents This book aims to provide a systematic introduction to fundamental RL theories, mainstream RL algorithms and typical RL applications for researchers and engineers. The main topics include Markov decision processes, Monte Carlo learning, temporal difference learning, RL with function approximation, policy gradient method, approximate dynamic programming, deep reinforcement learning, etc. Chapter 1 provides an overview of RL, including its history, famous scholars, successful examples and up-to-date challenges. Chapter 2 discusses the basis of RL, including main concepts and terminologies, Bellman’s optimality condition, and general problem formulation. Chapter 3 introduces Monte Carlo learning methods for model-free RL, including on-policy/off-policy methods and importance sampling technique. Chapter 4 introduces temporal difference learning methods for model-free RL, including Sarsa, Q-learning, and expected Sarsa. Chapter 5 introduces stochastic dynamic programming (DP), i.e., model-based RL with tabular representation, including value iteration DP, policy iteration DP and their convergence mechanisms. Chapter 6 introduces how to approximate policy and value functions in indirect RL methods as well as the associated actor-critic architecture. Chapter 7 derives different kinds of direct policy gradients, including likelihood ratio gradient, natural policy gradient and a few advanced variants. Chapter 8 introduces infinite-horizon ADP, finite-horizon ADP and its connection with model predictive control. Chapter 9 discusses how to handle state constraints and its connection with feasibility and safety, as well as the newly proposed actor-critic-scenery learning architecture. Chapter 10 is devoted to deep reinforcement learning, including how to train artificial neural networks and typical deep RL algorithms such as DQN, DDPG, TD3, TRPO, PPO, SAC, and DSAC. Chapter 11 provides various RL topics,including robust RL, POMDP, multi-agent RL, meta-RL, inverse RL, offline RL, major RL libraries and platforms. Author information: Shengbo Eben Li is currently a professor at Tsinghua University in the interdisciplinary field of autonomous driving and artificial intelligence. Before joining Tsinghua University, he has worked at Stanford University, University of Michigan, and UC Berkeley. His active research interests include intelligent vehicles and driver assistance, deep reinforcement learning, optimal control and estimation, etc. He has published more than 130 peer-reviewed papers in top-tier international journals and conferences. He is the recipient of best paper awards (finalists) of IEEE ITSC, ICCAS, IEEE ICUS, IEEE IV, L4DC, etc. He has received a number of important academic honors, including National Award for Technological Invention of China (2013), National Award for Progress in Sci & Tech of China (2018), Distinguished Young Scholar of Beijing NSF (2018), Youth Sci & Tech Innovation Leader from MOST China (2020), etc. He also serves as Board of Governor of IEEE ITS Society, Senior AE of IEEE OJ ITS, and AEs of IEEE ITSM, IEEE Trans ITS, Automotive Innovation, etc.

[D] The current and future state of AI/ML is shockingly demoralizing with little hope of redemption
reddit
LLM Vibe Score0
Human Vibe Score1
Flaky_Suit_8665This week

[D] The current and future state of AI/ML is shockingly demoralizing with little hope of redemption

I recently encountered the PaLM (Scaling Language Modeling with Pathways) paper from Google Research and it opened up a can of worms of ideas I’ve felt I’ve intuitively had for a while, but have been unable to express – and I know I can’t be the only one. Sometimes I wonder what the original pioneers of AI – Turing, Neumann, McCarthy, etc. – would think if they could see the state of AI that we’ve gotten ourselves into. 67 authors, 83 pages, 540B parameters in a model, the internals of which no one can say they comprehend with a straight face, 6144 TPUs in a commercial lab that no one has access to, on a rig that no one can afford, trained on a volume of data that a human couldn’t process in a lifetime, 1 page on ethics with the same ideas that have been rehashed over and over elsewhere with no attempt at a solution – bias, racism, malicious use, etc. – for purposes that who asked for? When I started my career as an AI/ML research engineer 2016, I was most interested in two types of tasks – 1.) those that most humans could do but that would universally be considered tedious and non-scalable. I’m talking image classification, sentiment analysis, even document summarization, etc. 2.) tasks that humans lack the capacity to perform as well as computers for various reasons – forecasting, risk analysis, game playing, and so forth. I still love my career, and I try to only work on projects in these areas, but it’s getting harder and harder. This is because, somewhere along the way, it became popular and unquestionably acceptable to push AI into domains that were originally uniquely human, those areas that sit at the top of Maslows’s hierarchy of needs in terms of self-actualization – art, music, writing, singing, programming, and so forth. These areas of endeavor have negative logarithmic ability curves – the vast majority of people cannot do them well at all, about 10% can do them decently, and 1% or less can do them extraordinarily. The little discussed problem with AI-generation is that, without extreme deterrence, we will sacrifice human achievement at the top percentile in the name of lowering the bar for a larger volume of people, until the AI ability range is the norm. This is because relative to humans, AI is cheap, fast, and infinite, to the extent that investments in human achievement will be watered down at the societal, educational, and individual level with each passing year. And unlike AI gameplay which superseded humans decades ago, we won’t be able to just disqualify the machines and continue to play as if they didn’t exist. Almost everywhere I go, even this forum, I encounter almost universal deference given to current SOTA AI generation systems like GPT-3, CODEX, DALL-E, etc., with almost no one extending their implications to its logical conclusion, which is long-term convergence to the mean, to mediocrity, in the fields they claim to address or even enhance. If you’re an artist or writer and you’re using DALL-E or GPT-3 to “enhance” your work, or if you’re a programmer saying, “GitHub Co-Pilot makes me a better programmer?”, then how could you possibly know? You’ve disrupted and bypassed your own creative process, which is thoughts -> (optionally words) -> actions -> feedback -> repeat, and instead seeded your canvas with ideas from a machine, the provenance of which you can’t understand, nor can the machine reliably explain. And the more you do this, the more you make your creative processes dependent on said machine, until you must question whether or not you could work at the same level without it. When I was a college student, I often dabbled with weed, LSD, and mushrooms, and for a while, I thought the ideas I was having while under the influence were revolutionary and groundbreaking – that is until took it upon myself to actually start writing down those ideas and then reviewing them while sober, when I realized they weren’t that special at all. What I eventually determined is that, under the influence, it was impossible for me to accurately evaluate the drug-induced ideas I was having because the influencing agent the generates the ideas themselves was disrupting the same frame of reference that is responsible evaluating said ideas. This is the same principle of – if you took a pill and it made you stupider, would even know it? I believe that, especially over the long-term timeframe that crosses generations, there’s significant risk that current AI-generation developments produces a similar effect on humanity, and we mostly won’t even realize it has happened, much like a frog in boiling water. If you have children like I do, how can you be aware of the the current SOTA in these areas, project that 20 to 30 years, and then and tell them with a straight face that it is worth them pursuing their talent in art, writing, or music? How can you be honest and still say that widespread implementation of auto-correction hasn’t made you and others worse and worse at spelling over the years (a task that even I believe most would agree is tedious and worth automating). Furthermore, I’ve yet to set anyone discuss the train – generate – train - generate feedback loop that long-term application of AI-generation systems imply. The first generations of these models were trained on wide swaths of web data generated by humans, but if these systems are permitted to continually spit out content without restriction or verification, especially to the extent that it reduces or eliminates development and investment in human talent over the long term, then what happens to the 4th or 5th generation of models? Eventually we encounter this situation where the AI is being trained almost exclusively on AI-generated content, and therefore with each generation, it settles more and more into the mean and mediocrity with no way out using current methods. By the time that happens, what will we have lost in terms of the creative capacity of people, and will we be able to get it back? By relentlessly pursuing this direction so enthusiastically, I’m convinced that we as AI/ML developers, companies, and nations are past the point of no return, and it mostly comes down the investments in time and money that we’ve made, as well as a prisoner’s dilemma with our competitors. As a society though, this direction we’ve chosen for short-term gains will almost certainly make humanity worse off, mostly for those who are powerless to do anything about it – our children, our grandchildren, and generations to come. If you’re an AI researcher or a data scientist like myself, how do you turn things back for yourself when you’ve spent years on years building your career in this direction? You’re likely making near or north of $200k annually TC and have a family to support, and so it’s too late, no matter how you feel about the direction the field has gone. If you’re a company, how do you standby and let your competitors aggressively push their AutoML solutions into more and more markets without putting out your own? Moreover, if you’re a manager or thought leader in this field like Jeff Dean how do you justify to your own boss and your shareholders your team’s billions of dollars in AI investment while simultaneously balancing ethical concerns? You can’t – the only answer is bigger and bigger models, more and more applications, more and more data, and more and more automation, and then automating that even further. If you’re a country like the US, how do responsibly develop AI while your competitors like China single-mindedly push full steam ahead without an iota of ethical concern to replace you in numerous areas in global power dynamics? Once again, failing to compete would be pre-emptively admitting defeat. Even assuming that none of what I’ve described here happens to such an extent, how are so few people not taking this seriously and discounting this possibility? If everything I’m saying is fear-mongering and non-sense, then I’d be interested in hearing what you think human-AI co-existence looks like in 20 to 30 years and why it isn’t as demoralizing as I’ve made it out to be. ​ EDIT: Day after posting this -- this post took off way more than I expected. Even if I received 20 - 25 comments, I would have considered that a success, but this went much further. Thank you to each one of you that has read this post, even more so if you left a comment, and triply so for those who gave awards! I've read almost every comment that has come in (even the troll ones), and am truly grateful for each one, including those in sharp disagreement. I've learned much more from this discussion with the sub than I could have imagined on this topic, from so many perspectives. While I will try to reply as many comments as I can, the sheer comment volume combined with limited free time between work and family unfortunately means that there are many that I likely won't be able to get to. That will invariably include some that I would love respond to under the assumption of infinite time, but I will do my best, even if the latency stretches into days. Thank you all once again!

[D] The machine learning community has a toxicity problem
reddit
LLM Vibe Score0
Human Vibe Score1
yusuf-bengioThis week

[D] The machine learning community has a toxicity problem

It is omnipresent! First of all, the peer-review process is broken. Every fourth NeurIPS submission is put on arXiv. There are DeepMind researchers publicly going after reviewers who are criticizing their ICLR submission. On top of that, papers by well-known institutes that were put on arXiv are accepted at top conferences, despite the reviewers agreeing on rejection. In contrast, vice versa, some papers with a majority of accepts are overruled by the AC. (I don't want to call any names, just have a look the openreview page of this year's ICRL). Secondly, there is a reproducibility crisis. Tuning hyperparameters on the test set seem to be the standard practice nowadays. Papers that do not beat the current state-of-the-art method have a zero chance of getting accepted at a good conference. As a result, hyperparameters get tuned and subtle tricks implemented to observe a gain in performance where there isn't any. Thirdly, there is a worshiping problem. Every paper with a Stanford or DeepMind affiliation gets praised like a breakthrough. For instance, BERT has seven times more citations than ULMfit. The Google affiliation gives so much credibility and visibility to a paper. At every ICML conference, there is a crowd of people in front of every DeepMind poster, regardless of the content of the work. The same story happened with the Zoom meetings at the virtual ICLR 2020. Moreover, NeurIPS 2020 had twice as many submissions as ICML, even though both are top-tier ML conferences. Why? Why is the name "neural" praised so much? Next, Bengio, Hinton, and LeCun are truly deep learning pioneers but calling them the "godfathers" of AI is insane. It has reached the level of a cult. Fourthly, the way Yann LeCun talked about biases and fairness topics was insensitive. However, the toxicity and backlash that he received are beyond any reasonable quantity. Getting rid of LeCun and silencing people won't solve any issue. Fifthly, machine learning, and computer science in general, have a huge diversity problem. At our CS faculty, only 30% of undergrads and 15% of the professors are women. Going on parental leave during a PhD or post-doc usually means the end of an academic career. However, this lack of diversity is often abused as an excuse to shield certain people from any form of criticism. Reducing every negative comment in a scientific discussion to race and gender creates a toxic environment. People are becoming afraid to engage in fear of being called a racist or sexist, which in turn reinforces the diversity problem. Sixthly, moral and ethics are set arbitrarily. The U.S. domestic politics dominate every discussion. At this very moment, thousands of Uyghurs are put into concentration camps based on computer vision algorithms invented by this community, and nobody seems even remotely to care. Adding a "broader impact" section at the end of every people will not make this stop. There are huge shitstorms because a researcher wasn't mentioned in an article. Meanwhile, the 1-billion+ people continent of Africa is virtually excluded from any meaningful ML discussion (besides a few Indaba workshops). Seventhly, there is a cut-throat publish-or-perish mentality. If you don't publish 5+ NeurIPS/ICML papers per year, you are a looser. Research groups have become so large that the PI does not even know the name of every PhD student anymore. Certain people submit 50+ papers per year to NeurIPS. The sole purpose of writing a paper has become to having one more NeurIPS paper in your CV. Quality is secondary; passing the peer-preview stage has become the primary objective. Finally, discussions have become disrespectful. Schmidhuber calls Hinton a thief, Gebru calls LeCun a white supremacist, Anandkumar calls Marcus a sexist, everybody is under attack, but nothing is improved. Albert Einstein was opposing the theory of quantum mechanics. Can we please stop demonizing those who do not share our exact views. We are allowed to disagree without going for the jugular. The moment we start silencing people because of their opinion is the moment scientific and societal progress dies. Best intentions, Yusuf

[D]Stuck in AI Hell: What to do in post LLM world
reddit
LLM Vibe Score0
Human Vibe Score1
Educational_News_371This week

[D]Stuck in AI Hell: What to do in post LLM world

Hey Reddit, I’ve been in an AI/ML role for a few years now, and I’m starting to feel disconnected from the work. When I started, deep learning models were getting good, and I quickly fell in love with designing architectures, training models, and fine-tuning them for specific use cases. Seeing a loss curve finally converge, experimenting with layers, and debugging training runs—it all felt like a craft, a blend of science and creativity. I enjoyed implementing research papers to see how things worked under the hood. Backprop, gradients, optimization—it was a mental workout I loved. But these days, it feels like everything has shifted. LLMs dominate the scene, and instead of building and training models, the focus is on using pre-trained APIs, crafting prompt chains, and setting up integrations. Sure, there’s engineering involved, but it feels less like creating and more like assembling. I miss the hands-on nature of experimenting with architectures and solving math-heavy problems. It’s not just the creativity I miss. The economics of this new era also feel strange to me. Back when I started, compute was a luxury. We had limited GPUs, and a lot of the work was about being resourceful—quantizing models, distilling them, removing layers, and squeezing every bit of performance out of constrained setups. Now, it feels like no one cares about cost. We’re paying by tokens. Tokens! Who would’ve thought we’d get to a point where we’re not designing efficient models but feeding pre-trained giants like they’re vending machines? I get it—abstraction has always been part of the field. TensorFlow and PyTorch abstracted tensor operations, Python abstracts C. But deep learning still left room for creation. We weren’t just abstracting away math; we were solving it. We could experiment, fail, and tweak. Working with LLMs doesn’t feel the same. It’s like fitting pieces into a pre-defined puzzle instead of building the puzzle itself. I understand that LLMs are here to stay. They’re incredible tools, and I respect their potential to revolutionize industries. Building real-world products with them is still challenging, requiring a deep understanding of engineering, prompt design, and integrating them effectively into workflows. By no means is it an “easy” task. But the work doesn’t give me the same thrill. It’s not about solving math or optimization problems—it’s about gluing together APIs, tweaking outputs, and wrestling with opaque systems. It’s like we’ve traded craftsmanship for convenience. Which brings me to my questions: Is there still room for those of us who enjoy the deep work of model design and training? Or is this the inevitable evolution of the field, where everything converges on pre-trained systems? What use cases still need traditional ML expertise? Are there industries or problems that will always require specialized models instead of general-purpose LLMs? Am I missing the bigger picture here? LLMs feel like the “kernel” of a new computing paradigm, and we don’t fully understand their second- and third-order effects. Could this shift lead to new, exciting opportunities I’m just not seeing yet? How do you stay inspired when the focus shifts? I still love AI, but I miss the feeling of building something from scratch. Is this just a matter of adapting my mindset, or should I seek out niches where traditional ML still thrives? I’m not asking this to rant (though clearly, I needed to get some of this off my chest). I want to figure out where to go next from here. If you’ve been in AI/ML long enough to see major shifts—like the move from feature engineering to deep learning—how did you navigate them? What advice would you give someone in my position? And yeah, before anyone roasts me for using an LLM to structure this post (guilty!), I just wanted to get my thoughts out in a coherent way. Guess that’s a sign of where we’re headed, huh? Thanks for reading, and I’d love to hear your thoughts! TL;DR: I entered AI during the deep learning boom, fell in love with designing and training models, and thrived on creativity, math, and optimization. Now it feels like the field is all about tweaking prompts and orchestrating APIs for pre-trained LLMs. I miss the thrill of crafting something unique. Is there still room for people who enjoy traditional ML, or is this just the inevitable evolution of the field? How do you stay inspired amidst such shifts? Update: Wow, this blew up. Thanks everyone for your comments and suggestions. I really like some of those. This thing was on my mind for a long time, glad that I put it here. Thanks again!

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.
reddit
LLM Vibe Score0
Human Vibe Score0.765
hardmaruThis week

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.

Schmidhuber interview expressing his views on the future of AI and AGI. Original source. I think the interview is of interest to r/MachineLearning, and presents an alternate view, compared to other influential leaders in AI. Juergen Schmidhuber, Renowned 'Father Of Modern AI,' Says His Life’s Work Won't Lead To Dystopia May 23, 2023. Contributed by Hessie Jones. Amid the growing concern about the impact of more advanced artificial intelligence (AI) technologies on society, there are many in the technology community who fear the implications of the advancements in Generative AI if they go unchecked. Dr. Juergen Schmidhuber, a renowned scientist, artificial intelligence researcher and widely regarded as one of the pioneers in the field, is more optimistic. He declares that many of those who suddenly warn against the dangers of AI are just seeking publicity, exploiting the media’s obsession with killer robots which has attracted more attention than “good AI” for healthcare etc. The potential to revolutionize various industries and improve our lives is clear, as are the equal dangers if bad actors leverage the technology for personal gain. Are we headed towards a dystopian future, or is there reason to be optimistic? I had a chance to sit down with Dr. Juergen Schmidhuber to understand his perspective on this seemingly fast-moving AI-train that will leap us into the future. As a teenager in the 1970s, Juergen Schmidhuber became fascinated with the idea of creating intelligent machines that could learn and improve on their own, becoming smarter than himself within his lifetime. This would ultimately lead to his groundbreaking work in the field of deep learning. In the 1980s, he studied computer science at the Technical University of Munich (TUM), where he earned his diploma in 1987. His thesis was on the ultimate self-improving machines that, not only, learn through some pre-wired human-designed learning algorithm, but also learn and improve the learning algorithm itself. Decades later, this became a hot topic. He also received his Ph.D. at TUM in 1991 for work that laid some of the foundations of modern AI. Schmidhuber is best known for his contributions to the development of recurrent neural networks (RNNs), the most powerful type of artificial neural network that can process sequential data such as speech and natural language. With his students Sepp Hochreiter, Felix Gers, Alex Graves, Daan Wierstra, and others, he published architectures and training algorithms for the long short-term memory (LSTM), a type of RNN that is widely used in natural language processing, speech recognition, video games, robotics, and other applications. LSTM has become the most cited neural network of the 20th century, and Business Week called it "arguably the most commercial AI achievement." Throughout his career, Schmidhuber has received various awards and accolades for his groundbreaking work. In 2013, he was awarded the Helmholtz Prize, which recognizes significant contributions to the field of machine learning. In 2016, he was awarded the IEEE Neural Network Pioneer Award for "pioneering contributions to deep learning and neural networks." The media have often called him the “father of modern AI,” because the most cited neural networks all build on his lab’s work. He is quick to point out, however, that AI history goes back centuries. Despite his many accomplishments, at the age of 60, he feels mounting time pressure towards building an Artificial General Intelligence within his lifetime and remains committed to pushing the boundaries of AI research and development. He is currently director of the KAUST AI Initiative, scientific director of the Swiss AI Lab IDSIA, and co-founder and chief scientist of AI company NNAISENSE, whose motto is "AI∀" which is a math-inspired way of saying "AI For All." He continues to work on cutting-edge AI technologies and applications to improve human health and extend human lives and make lives easier for everyone. The following interview has been edited for clarity. Jones: Thank you Juergen for joining me. You have signed letters warning about AI weapons. But you didn't sign the recent publication, "Pause Gigantic AI Experiments: An Open Letter"? Is there a reason? Schmidhuber: Thank you Hessie. Glad to speak with you. I have realized that many of those who warn in public against the dangers of AI are just seeking publicity. I don't think the latest letter will have any significant impact because many AI researchers, companies, and governments will ignore it completely. The proposal frequently uses the word "we" and refers to "us," the humans. But as I have pointed out many times in the past, there is no "we" that everyone can identify with. Ask 10 different people, and you will hear 10 different opinions about what is "good." Some of those opinions will be completely incompatible with each other. Don't forget the enormous amount of conflict between the many people. The letter also says, "If such a pause cannot be quickly put in place, governments should intervene and impose a moratorium." The problem is that different governments have ALSO different opinions about what is good for them and for others. Great Power A will say, if we don't do it, Great Power B will, perhaps secretly, and gain an advantage over us. The same is true for Great Powers C and D. Jones: Everyone acknowledges this fear surrounding current generative AI technology. Moreover, the existential threat of this technology has been publicly acknowledged by Sam Altman, CEO of OpenAI himself, calling for AI regulation. From your perspective, is there an existential threat? Schmidhuber: It is true that AI can be weaponized, and I have no doubt that there will be all kinds of AI arms races, but AI does not introduce a new quality of existential threat. The threat coming from AI weapons seems to pale in comparison to the much older threat from nuclear hydrogen bombs that don’t need AI at all. We should be much more afraid of half-century-old tech in the form of H-bomb rockets. The Tsar Bomba of 1961 had almost 15 times more destructive power than all weapons of WW-II combined. Despite the dramatic nuclear disarmament since the 1980s, there are still more than enough nuclear warheads to wipe out human civilization within two hours, without any AI I’m much more worried about that old existential threat than the rather harmless AI weapons. Jones: I realize that while you compare AI to the threat of nuclear bombs, there is a current danger that a current technology can be put in the hands of humans and enable them to “eventually” exact further harms to individuals of group in a very precise way, like targeted drone attacks. You are giving people a toolset that they've never had before, enabling bad actors, as some have pointed out, to be able to do a lot more than previously because they didn't have this technology. Schmidhuber: Now, all that sounds horrible in principle, but our existing laws are sufficient to deal with these new types of weapons enabled by AI. If you kill someone with a gun, you will go to jail. Same if you kill someone with one of these drones. Law enforcement will get better at understanding new threats and new weapons and will respond with better technology to combat these threats. Enabling drones to target persons from a distance in a way that requires some tracking and some intelligence to perform, which has traditionally been performed by skilled humans, to me, it seems is just an improved version of a traditional weapon, like a gun, which is, you know, a little bit smarter than the old guns. But, in principle, all of that is not a new development. For many centuries, we have had the evolution of better weaponry and deadlier poisons and so on, and law enforcement has evolved their policies to react to these threats over time. So, it's not that we suddenly have a new quality of existential threat and it's much more worrisome than what we have had for about six decades. A large nuclear warhead doesn’t need fancy face recognition to kill an individual. No, it simply wipes out an entire city with ten million inhabitants. Jones: The existential threat that’s implied is the extent to which humans have control over this technology. We see some early cases of opportunism which, as you say, tends to get more media attention than positive breakthroughs. But you’re implying that this will all balance out? Schmidhuber: Historically, we have a long tradition of technological breakthroughs that led to advancements in weapons for the purpose of defense but also for protection. From sticks, to rocks, to axes to gunpowder to cannons to rockets… and now to drones… this has had a drastic influence on human history but what has been consistent throughout history is that those who are using technology to achieve their own ends are themselves, facing the same technology because the opposing side is learning to use it against them. And that's what has been repeated in thousands of years of human history and it will continue. I don't see the new AI arms race as something that is remotely as existential a threat as the good old nuclear warheads. You said something important, in that some people prefer to talk about the downsides rather than the benefits of this technology, but that's misleading, because 95% of all AI research and AI development is about making people happier and advancing human life and health. Jones: Let’s touch on some of those beneficial advances in AI research that have been able to radically change present day methods and achieve breakthroughs. Schmidhuber: All right! For example, eleven years ago, our team with my postdoc Dan Ciresan was the first to win a medical imaging competition through deep learning. We analyzed female breast cells with the objective to determine harmless cells vs. those in the pre-cancer stage. Typically, a trained oncologist needs a long time to make these determinations. Our team, who knew nothing about cancer, were able to train an artificial neural network, which was totally dumb in the beginning, on lots of this kind of data. It was able to outperform all the other methods. Today, this is being used not only for breast cancer, but also for radiology and detecting plaque in arteries, and many other things. Some of the neural networks that we have developed in the last 3 decades are now prevalent across thousands of healthcare applications, detecting Diabetes and Covid-19 and what not. This will eventually permeate across all healthcare. The good consequences of this type of AI are much more important than the click-bait new ways of conducting crimes with AI. Jones: Adoption is a product of reinforced outcomes. The massive scale of adoption either leads us to believe that people have been led astray, or conversely, technology is having a positive effect on people’s lives. Schmidhuber: The latter is the likely case. There's intense commercial pressure towards good AI rather than bad AI because companies want to sell you something, and you are going to buy only stuff you think is going to be good for you. So already just through this simple, commercial pressure, you have a tremendous bias towards good AI rather than bad AI. However, doomsday scenarios like in Schwarzenegger movies grab more attention than documentaries on AI that improve people’s lives. Jones: I would argue that people are drawn to good stories – narratives that contain an adversary and struggle, but in the end, have happy endings. And this is consistent with your comment on human nature and how history, despite its tendency for violence and destruction of humanity, somehow tends to correct itself. Let’s take the example of a technology, which you are aware – GANs – General Adversarial Networks, which today has been used in applications for fake news and disinformation. In actuality, the purpose in the invention of GANs was far from what it is used for today. Schmidhuber: Yes, the name GANs was created in 2014 but we had the basic principle already in the early 1990s. More than 30 years ago, I called it artificial curiosity. It's a very simple way of injecting creativity into a little two network system. This creative AI is not just trying to slavishly imitate humans. Rather, it’s inventing its own goals. Let me explain: You have two networks. One network is producing outputs that could be anything, any action. Then the second network is looking at these actions and it’s trying to predict the consequences of these actions. An action could move a robot, then something happens, and the other network is just trying to predict what will happen. Now we can implement artificial curiosity by reducing the prediction error of the second network, which, at the same time, is the reward of the first network. The first network wants to maximize its reward and so it will invent actions that will lead to situations that will surprise the second network, which it has not yet learned to predict well. In the case where the outputs are fake images, the first network will try to generate images that are good enough to fool the second network, which will attempt to predict the reaction of the environment: fake or real image, and it will try to become better at it. The first network will continue to also improve at generating images whose type the second network will not be able to predict. So, they fight each other. The 2nd network will continue to reduce its prediction error, while the 1st network will attempt to maximize it. Through this zero-sum game the first network gets better and better at producing these convincing fake outputs which look almost realistic. So, once you have an interesting set of images by Vincent Van Gogh, you can generate new images that leverage his style, without the original artist having ever produced the artwork himself. Jones: I see how the Van Gogh example can be applied in an education setting and there are countless examples of artists mimicking styles from famous painters but image generation from this instance that can happen within seconds is quite another feat. And you know this is how GANs has been used. What’s more prevalent today is a socialized enablement of generating images or information to intentionally fool people. It also surfaces new harms that deal with the threat to intellectual property and copyright, where laws have yet to account for. And from your perspective this was not the intention when the model was conceived. What was your motivation in your early conception of what is now GANs? Schmidhuber: My old motivation for GANs was actually very important and it was not to create deepfakes or fake news but to enable AIs to be curious and invent their own goals, to make them explore their environment and make them creative. Suppose you have a robot that executes one action, then something happens, then it executes another action, and so on, because it wants to achieve certain goals in the environment. For example, when the battery is low, this will trigger “pain” through hunger sensors, so it wants to go to the charging station, without running into obstacles, which will trigger other pain sensors. It will seek to minimize pain (encoded through numbers). Now the robot has a friend, the second network, which is a world model ––it’s a prediction machine that learns to predict the consequences of the robot’s actions. Once the robot has a good model of the world, it can use it for planning. It can be used as a simulation of the real world. And then it can determine what is a good action sequence. If the robot imagines this sequence of actions, the model will predict a lot of pain, which it wants to avoid. If it plays this alternative action sequence in its mental model of the world, then it will predict a rewarding situation where it’s going to sit on the charging station and its battery is going to load again. So, it'll prefer to execute the latter action sequence. In the beginning, however, the model of the world knows nothing, so how can we motivate the first network to generate experiments that lead to data that helps the world model learn something it didn’t already know? That’s what artificial curiosity is about. The dueling two network systems effectively explore uncharted environments by creating experiments so that over time the curious AI gets a better sense of how the environment works. This can be applied to all kinds of environments, and has medical applications. Jones: Let’s talk about the future. You have said, “Traditional humans won’t play a significant role in spreading intelligence across the universe.” Schmidhuber: Let’s first conceptually separate two types of AIs. The first type of AI are tools directed by humans. They are trained to do specific things like accurately detect diabetes or heart disease and prevent attacks before they happen. In these cases, the goal is coming from the human. More interesting AIs are setting their own goals. They are inventing their own experiments and learning from them. Their horizons expand and eventually they become more and more general problem solvers in the real world. They are not controlled by their parents, but much of what they learn is through self-invented experiments. A robot, for example, is rotating a toy, and as it is doing this, the video coming in through the camera eyes, changes over time and it begins to learn how this video changes and learns how the 3D nature of the toy generates certain videos if you rotate it a certain way, and eventually, how gravity works, and how the physics of the world works. Like a little scientist! And I have predicted for decades that future scaled-up versions of such AI scientists will want to further expand their horizons, and eventually go where most of the physical resources are, to build more and bigger AIs. And of course, almost all of these resources are far away from earth out there in space, which is hostile to humans but friendly to appropriately designed AI-controlled robots and self-replicating robot factories. So here we are not talking any longer about our tiny biosphere; no, we are talking about the much bigger rest of the universe. Within a few tens of billions of years, curious self-improving AIs will colonize the visible cosmos in a way that’s infeasible for humans. Those who don’t won’t have an impact. Sounds like science fiction, but since the 1970s I have been unable to see a plausible alternative to this scenario, except for a global catastrophe such as an all-out nuclear war that stops this development before it takes off. Jones: How long have these AIs, which can set their own goals — how long have they existed? To what extent can they be independent of human interaction? Schmidhuber: Neural networks like that have existed for over 30 years. My first simple adversarial neural network system of this kind is the one from 1990 described above. You don’t need a teacher there; it's just a little agent running around in the world and trying to invent new experiments that surprise its own prediction machine. Once it has figured out certain parts of the world, the agent will become bored and will move on to more exciting experiments. The simple 1990 systems I mentioned have certain limitations, but in the past three decades, we have also built more sophisticated systems that are setting their own goals and such systems I think will be essential for achieving true intelligence. If you are only imitating humans, you will never go beyond them. So, you really must give AIs the freedom to explore previously unexplored regions of the world in a way that no human is really predefining. Jones: Where is this being done today? Schmidhuber: Variants of neural network-based artificial curiosity are used today for agents that learn to play video games in a human-competitive way. We have also started to use them for automatic design of experiments in fields such as materials science. I bet many other fields will be affected by it: chemistry, biology, drug design, you name it. However, at least for now, these artificial scientists, as I like to call them, cannot yet compete with human scientists. I don’t think it’s going to stay this way but, at the moment, it’s still the case. Sure, AI has made a lot of progress. Since 1997, there have been superhuman chess players, and since 2011, through the DanNet of my team, there have been superhuman visual pattern recognizers. But there are other things where humans, at the moment at least, are much better, in particular, science itself. In the lab we have many first examples of self-directed artificial scientists, but they are not yet convincing enough to appear on the radar screen of the public space, which is currently much more fascinated with simpler systems that just imitate humans and write texts based on previously seen human-written documents. Jones: You speak of these numerous instances dating back 30 years of these lab experiments where these self-driven agents are deciding and learning and moving on once they’ve learned. And I assume that that rate of learning becomes even faster over time. What kind of timeframe are we talking about when this eventually is taken outside of the lab and embedded into society? Schmidhuber: This could still take months or even years :-) Anyway, in the not-too-distant future, we will probably see artificial scientists who are good at devising experiments that allow them to discover new, previously unknown physical laws. As always, we are going to profit from the old trend that has held at least since 1941: every decade compute is getting 100 times cheaper. Jones: How does this trend affect modern AI such as ChatGPT? Schmidhuber: Perhaps you know that all the recent famous AI applications such as ChatGPT and similar models are largely based on principles of artificial neural networks invented in the previous millennium. The main reason why they works so well now is the incredible acceleration of compute per dollar. ChatGPT is driven by a neural network called “Transformer” described in 2017 by Google. I am happy about that because a quarter century earlier in 1991 I had a particular Transformer variant which is now called the “Transformer with linearized self-attention”. Back then, not much could be done with it, because the compute cost was a million times higher than today. But today, one can train such models on half the internet and achieve much more interesting results. Jones: And for how long will this acceleration continue? Schmidhuber: There's no reason to believe that in the next 30 years, we won't have another factor of 1 million and that's going to be really significant. In the near future, for the first time we will have many not-so expensive devices that can compute as much as a human brain. The physical limits of computation, however, are much further out so even if the trend of a factor of 100 every decade continues, the physical limits (of 1051 elementary instructions per second and kilogram of matter) won’t be hit until, say, the mid-next century. Even in our current century, however, we’ll probably have many machines that compute more than all 10 billion human brains collectively and you can imagine, everything will change then! Jones: That is the big question. Is everything going to change? If so, what do you say to the next generation of leaders, currently coming out of college and university. So much of this change is already impacting how they study, how they will work, or how the future of work and livelihood is defined. What is their purpose and how do we change our systems so they will adapt to this new version of intelligence? Schmidhuber: For decades, people have asked me questions like that, because you know what I'm saying now, I have basically said since the 1970s, it’s just that today, people are paying more attention because, back then, they thought this was science fiction. They didn't think that I would ever come close to achieving my crazy life goal of building a machine that learns to become smarter than myself such that I can retire. But now many have changed their minds and think it's conceivable. And now I have two daughters, 23 and 25. People ask me: what do I tell them? They know that Daddy always said, “It seems likely that within your lifetimes, you will have new types of intelligence that are probably going to be superior in many ways, and probably all kinds of interesting ways.” How should they prepare for that? And I kept telling them the obvious: Learn how to learn new things! It's not like in the previous millennium where within 20 years someone learned to be a useful member of society, and then took a job for 40 years and performed in this job until she received her pension. Now things are changing much faster and we must learn continuously just to keep up. I also told my girls that no matter how smart AIs are going to get, learn at least the basics of math and physics, because that’s the essence of our universe, and anybody who understands this will have an advantage, and learn all kinds of new things more easily. I also told them that social skills will remain important, because most future jobs for humans will continue to involve interactions with other humans, but I couldn’t teach them anything about that; they know much more about social skills than I do. You touched on the big philosophical question about people’s purpose. Can this be answered without answering the even grander question: What’s the purpose of the entire universe? We don’t know. But what’s happening right now might be connected to the unknown answer. Don’t think of humans as the crown of creation. Instead view human civilization as part of a much grander scheme, an important step (but not the last one) on the path of the universe from very simple initial conditions towards more and more unfathomable complexity. Now it seems ready to take its next step, a step comparable to the invention of life itself over 3.5 billion years ago. Alas, don’t worry, in the end, all will be good! Jones: Let’s get back to this transformation happening right now with OpenAI. There are many questioning the efficacy and accuracy of ChatGPT, and are concerned its release has been premature. In light of the rampant adoption, educators have banned its use over concerns of plagiarism and how it stifles individual development. Should large language models like ChatGPT be used in school? Schmidhuber: When the calculator was first introduced, instructors forbade students from using it in school. Today, the consensus is that kids should learn the basic methods of arithmetic, but they should also learn to use the “artificial multipliers” aka calculators, even in exams, because laziness and efficiency is a hallmark of intelligence. Any intelligent being wants to minimize its efforts to achieve things. And that's the reason why we have tools, and why our kids are learning to use these tools. The first stone tools were invented maybe 3.5 million years ago; tools just have become more sophisticated over time. In fact, humans have changed in response to the properties of their tools. Our anatomical evolution was shaped by tools such as spears and fire. So, it's going to continue this way. And there is no permanent way of preventing large language models from being used in school. Jones: And when our children, your children graduate, what does their future work look like? Schmidhuber: A single human trying to predict details of how 10 billion people and their machines will evolve in the future is like a single neuron in my brain trying to predict what the entire brain and its tens of billions of neurons will do next year. 40 years ago, before the WWW was created at CERN in Switzerland, who would have predicted all those young people making money as YouTube video bloggers? Nevertheless, let’s make a few limited job-related observations. For a long time, people have thought that desktop jobs may require more intelligence than skills trade or handicraft professions. But now, it turns out that it's much easier to replace certain aspects of desktop jobs than replacing a carpenter, for example. Because everything that works well in AI is happening behind the screen currently, but not so much in the physical world. There are now artificial systems that can read lots of documents and then make really nice summaries of these documents. That is a desktop job. Or you give them a description of an illustration that you want to have for your article and pretty good illustrations are being generated that may need some minimal fine-tuning. But you know, all these desktop jobs are much easier to facilitate than the real tough jobs in the physical world. And it's interesting that the things people thought required intelligence, like playing chess, or writing or summarizing documents, are much easier for machines than they thought. But for things like playing football or soccer, there is no physical robot that can remotely compete with the abilities of a little boy with these skills. So, AI in the physical world, interestingly, is much harder than AI behind the screen in virtual worlds. And it's really exciting, in my opinion, to see that jobs such as plumbers are much more challenging than playing chess or writing another tabloid story. Jones: The way data has been collected in these large language models does not guarantee personal information has not been excluded. Current consent laws already are outdated when it comes to these large language models (LLM). The concern, rightly so, is increasing surveillance and loss of privacy. What is your view on this? Schmidhuber: As I have indicated earlier: are surveillance and loss of privacy inevitable consequences of increasingly complex societies? Super-organisms such as cities and states and companies consist of numerous people, just like people consist of numerous cells. These cells enjoy little privacy. They are constantly monitored by specialized "police cells" and "border guard cells": Are you a cancer cell? Are you an external intruder, a pathogen? Individual cells sacrifice their freedom for the benefits of being part of a multicellular organism. Similarly, for super-organisms such as nations. Over 5000 years ago, writing enabled recorded history and thus became its inaugural and most important invention. Its initial purpose, however, was to facilitate surveillance, to track citizens and their tax payments. The more complex a super-organism, the more comprehensive its collection of information about its constituents. 200 years ago, at least, the parish priest in each village knew everything about all the village people, even about those who did not confess, because they appeared in the confessions of others. Also, everyone soon knew about the stranger who had entered the village, because some occasionally peered out of the window, and what they saw got around. Such control mechanisms were temporarily lost through anonymization in rapidly growing cities but are now returning with the help of new surveillance devices such as smartphones as part of digital nervous systems that tell companies and governments a lot about billions of users. Cameras and drones etc. are becoming increasingly tinier and more ubiquitous. More effective recognition of faces and other detection technology are becoming cheaper and cheaper, and many will use it to identify others anywhere on earth; the big wide world will not offer any more privacy than the local village. Is this good or bad? Some nations may find it easier than others to justify more complex kinds of super-organisms at the expense of the privacy rights of their constituents. Jones: So, there is no way to stop or change this process of collection, or how it continuously informs decisions over time? How do you see governance and rules responding to this, especially amid Italy’s ban on ChatGPT following suspected user data breach and the more recent news about the Meta’s record $1.3billion fine in the company’s handling of user information? Schmidhuber: Data collection has benefits and drawbacks, such as the loss of privacy. How to balance those? I have argued for addressing this through data ownership in data markets. If it is true that data is the new oil, then it should have a price, just like oil. At the moment, the major surveillance platforms such as Meta do not offer users any money for their data and the transitive loss of privacy. In the future, however, we will likely see attempts at creating efficient data markets to figure out the data's true financial value through the interplay between supply and demand. Even some of the sensitive medical data should not be priced by governmental regulators but by patients (and healthy persons) who own it and who may sell or license parts thereof as micro-entrepreneurs in a healthcare data market. Following a previous interview, I gave for one of the largest re-insurance companies , let's look at the different participants in such a data market: patients, hospitals, data companies. (1) Patients with a rare form of cancer can offer more valuable data than patients with a very common form of cancer. (2) Hospitals and their machines are needed to extract the data, e.g., through magnet spin tomography, radiology, evaluations through human doctors, and so on. (3) Companies such as Siemens, Google or IBM would like to buy annotated data to make better artificial neural networks that learn to predict pathologies and diseases and the consequences of therapies. Now the market’s invisible hand will decide about the data’s price through the interplay between demand and supply. On the demand side, you will have several companies offering something for the data, maybe through an app on the smartphone (a bit like a stock market app). On the supply side, each patient in this market should be able to profit from high prices for rare valuable types of data. Likewise, competing data extractors such as hospitals will profit from gaining recognition and trust for extracting data well at a reasonable price. The market will make the whole system efficient through incentives for all who are doing a good job. Soon there will be a flourishing ecosystem of commercial data market advisors and what not, just like the ecosystem surrounding the traditional stock market. The value of the data won’t be determined by governments or ethics committees, but by those who own the data and decide by themselves which parts thereof they want to license to others under certain conditions. At first glance, a market-based system seems to be detrimental to the interest of certain monopolistic companies, as they would have to pay for the data - some would prefer free data and keep their monopoly. However, since every healthy and sick person in the market would suddenly have an incentive to collect and share their data under self-chosen anonymity conditions, there will soon be many more useful data to evaluate all kinds of treatments. On average, people will live longer and healthier, and many companies and the entire healthcare system will benefit. Jones: Finally, what is your view on open source versus the private companies like Google and OpenAI? Is there a danger to supporting these private companies’ large language models versus trying to keep these models open source and transparent, very much like what LAION is doing? Schmidhuber: I signed this open letter by LAION because I strongly favor the open-source movement. And I think it's also something that is going to challenge whatever big tech dominance there might be at the moment. Sure, the best models today are run by big companies with huge budgets for computers, but the exciting fact is that open-source models are not so far behind, some people say maybe six to eight months only. Of course, the private company models are all based on stuff that was created in academia, often in little labs without so much funding, which publish without patenting their results and open source their code and others take it and improved it. Big tech has profited tremendously from academia; their main achievement being that they have scaled up everything greatly, sometimes even failing to credit the original inventors. So, it's very interesting to see that as soon as some big company comes up with a new scaled-up model, lots of students out there are competing, or collaborating, with each other, trying to come up with equal or better performance on smaller networks and smaller machines. And since they are open sourcing, the next guy can have another great idea to improve it, so now there’s tremendous competition also for the big companies. Because of that, and since AI is still getting exponentially cheaper all the time, I don't believe that big tech companies will dominate in the long run. They find it very hard to compete with the enormous open-source movement. As long as you can encourage the open-source community, I think you shouldn't worry too much. Now, of course, you might say if everything is open source, then the bad actors also will more easily have access to these AI tools. And there's truth to that. But as always since the invention of controlled fire, it was good that knowledge about how technology works quickly became public such that everybody could use it. And then, against any bad actor, there's almost immediately a counter actor trying to nullify his efforts. You see, I still believe in our old motto "AI∀" or "AI For All." Jones: Thank you, Juergen for sharing your perspective on this amazing time in history. It’s clear that with new technology, the enormous potential can be matched by disparate and troubling risks which we’ve yet to solve, and even those we have yet to identify. If we are to dispel the fear of a sentient system for which we have no control, humans, alone need to take steps for more responsible development and collaboration to ensure AI technology is used to ultimately benefit society. Humanity will be judged by what we do next.

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!
reddit
LLM Vibe Score0
Human Vibe Score-0.5
regalalgorithmThis week

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!

Hi there, we at Skynet Today produce a weekly newsletter summarizing each week's major AI news, which seems like it'd be of interest to this subreddit. Here's what's in our latest one: Facebook’s AI for detecting hate speech is facing its biggest challenge yet Facebook has made significant progress recently to proactively take down content that violate its community standards. For example, in the second quarter of 2020, Facebook took down 104.6 million pieces of content. While reviews are typically performed by a vast workforce of human moderators, AI-powered tools have enabled Facebook to do this work at a greater scale for textual content. However, there’s a long way to go for these systems to match or exceed the capabilities of human moderators. This is because a large proportion of hate speech and misinformation is in the form of images and memes, and reasoning about the context and language-image interplay is an extremely difficult challenge for AI. Given Facebook’s scale and the speed at which some use it to spread hate, incite violence, and share lies with millions, Facebook will have to keep running to catch up. AI Slays Top F-16 Pilot In DARPA Dogfight Simulation The Defense Advanced Research Project Agency (DARPA) recently hosted a simulated F16 dogfight competition, with different AI bots competing with each other as well as with human pilots. The top AI bot was able to beat a human pilot 5-0 in the simulated contest. DARPA started this program “as a risk-reduction effort \[…\] to flesh out how human and machine pilots share operational control of a fighter jet to maximize its chances of mission success.” Competition runners are broadly optimistic about the demonstration of AI capabilities, even if they are not close to being deployed on a real aircraft. Of concern, the program had little discussion on the ethics of AI military applications, especially with the lethal autonomous weapon systems being considered. News Advances & Business Microsoft, Energy Dept. to Develop Disaster-Response AI Tools \- The U.S. Department of Energy and Microsoft Corp. on Tuesday announced a partnership to develop artificial-intelligence tools aimed at helping first-responders better react to fast-changing natural events, such as floods and wildfires. Coronavirus: Robot CERi is a bilingual Covid-19 expert \- Ceri is bilingual, clued-up on coronavirus and can tell what mood you are in. Ceri also happens to be a robot. Moscow DOH uses AI platform to detect lung cancer symptoms \- Moscow’s department of health is using an artificial intelligence (AI) platform to detect symptoms of lung cancer in CT scans, as part of a project to implement AI technology for radiology. Scientists develop artificial intelligence system for high precision recognition of hand gestures \- The recognition of human hand gestures by AI systems has been a valuable development over the last decade and has been adopted in high-precision surgical robots, health monitoring equipment and in gaming systems. Forget credit cards - now you can pay with your face. Creepy or cool? \- A new way to pay has arrived in Los Angeles: your face. Concerns & Hype The dystopian tech that companies are selling to help schools reopen sooner \- This fall, AI could be watching students social distance and checking their masks. Thousands of schools nationwide will not be reopening this fall. NYPD Used Facial Recognition Technology In Siege Of Black Lives Matter Activist’s Apartment \- The NYPD deployed facial recognition technology in its hunt for a prominent Black Lives Matter activist, whose home was besieged by dozens of officers and police dogs last week, a spokesperson confirmed to Gothamist. Machines can spot mental health issues - if you hand over your personal data \- Digital diagnosis could transform psychiatry by mining your most intimate data for clues. But is the privacy cost worth it? Supporting Black Artists Who Are Examining AI \- Technology has a complicated relationship with racial justice. Smartphones, internet platforms, and other digital tools can be used to document and expose racism. But digital tools can also fuel racism: smart doorbells surveil Black individuals. A-level and GCSE results in England to be based on teacher assessments in U-turn \- All A-level and GCSE results in England will be based on grades assesed by teachers instead of algorithms. Analysis & Policy GPT-3 and The Question of Automation \- Automation is not an all or nothing proposition. An AI model’s automation capability is highly conjoined with the task and application it is used in. An A.I. Movie Service Could One Day Serve You a New Custom Film Every Time \- How long will it be until an A.I. can make an actual feature film on demand? Fairness, evidence, and predictive equality \- How the causal fairness principle relates to predictive equality How robotics and automation could create new jobs in the new normal \- Depending on who you ask, AI and automation will either destroy jobs or create new ones. In reality, a greater push toward automation will probably both kill and create jobs - human workers will become redundant in certain spheres, sure, but many new roles will likely crop up. Expert Opinions & Discussion within the field Too many AI researchers think real-world problems are not relevant \- The community’s hyperfocus on novel methods ignores what’s really important.

[N] TheSequence Scope: When it comes to machine learning, size matters: Microsoft's DeepSpeed framework, which can train a model with up to a trillion parameters
reddit
LLM Vibe Score0
Human Vibe Score1
KseniaseThis week

[N] TheSequence Scope: When it comes to machine learning, size matters: Microsoft's DeepSpeed framework, which can train a model with up to a trillion parameters

Hi there! Offering to your attention the latest edition of a weekly ML-newsletter that focusing on three things: impactful ML research papers, cool ML tech solutions, and ML use cases supported by investors. Please, see it below. Reddit is a new thing for me, and I've been struggling a bit with it, so please don't judge me too harsh for this promotion. This weekly digest is free and I hope you'd find the format convenient for you. Your feedback is very appreciated, and please feel free to sign up if you like it. 📝 Editorial  The recent emergence of pre-trained language models and transformer architectures pushed the creation of larger and larger machine learning models. Google’s BERT presented attention mechanism and transformer architecture possibilities as the “next big thing” in ML, and the numbers seem surreal. OpenAI’s GPT-2 set a record by processing 1.5 billion parameters, followed by Microsoft’s Turing-NLG, which processed 17 billion parameters just to see the new GPT-3 processing an astonishing 175 billion parameters. To not feel complacent, just this week Microsoft announced a new release of its DeepSpeed framework (which powers Turing-NLG), which can train a model with up to a trillion parameters. That sounds insane but it really isn’t.   What we are seeing is a consequence of several factors. First, computation power and parallelization techniques have evolved to a point where it is relatively easy to train machine learning models in large clusters of machines. Second and most importantly, in the current state of machine learning, larger models have regularly outperformed smaller and more specialized models. Knowledge reusability methods like transfer learning are still in very nascent stages. As a result, it’s really hard to build small models that can operate in uncertain environments. Furthermore, as models like GPT-3 and Turing-NLG have shown, there is some unexplainable magic that happens after models go past a certain size. Many of the immediate machine learning problems might be solved by scaling the current generation of neural network architectures. Plain and simple, when it comes to machine learning, size matters.   We would love to hear your opinions about the debate between broader-larger vs. smaller and more specialized models.   Leave a comment Now, to the most important developments in the AI industry this week 🔎 ML Research GPT-3 Falls Short in Machine Comprehension Proposed by researchers from a few major American universities, a 57-task test to measure models’ ability to reason poses challenges even for sophisticated models like GPT-3 ->read more in the original paper Better Text Summarization OpenAI published a paper showing a reinforcement learning with human feedback technique that can surpass supervised models ->read more on OpenAI blog Reinforcement Learning with Offline Datasets Researchers from the Berkeley AI Research (BAIR) Lab published a paper unveiling a method that uses offline datasets to improve reinforcement learning models->read more on BAIR blog 🤖 Cool AI Tech Releases New Version of DeepSpeed Microsoft open-sourced a new version of DeepSpeed, an open-source library for parallelizable training that can scale up to models with 1 trillion parameters->read more on Microsoft Research blog 💸 Money in AI AI-powered customer experience management platform Sprinklr has raised $200 million (kudos to our subscribers from Sprinklr 👏). Sprinklr's “AI listening processing” solution allows companies to get structured and meaningful sentiments and insights from unstructured customer data that comes from public conversations on different websites and social platforms. Xometry, an on-demand industrial parts marketplace, raises $75 million in Series E funding. The company provides a digital way of creating the right combination of buyers and manufacturers. Another example of AI implementation into matching two sides for a deal. Real estate tech company Orchard raises $69 million in its recent funding round. Orchard aims to digitize the whole real estate market, by developing a solution that combines machine learning and rapid human assistance to smooth the search, match the right deal, and simplify buying and selling relationships. Cybersecurity startup Pcysys raised $25 million in its funding round. Pcysys’ platform, which doesn’t require installation or network reconfiguration, uses algorithms to scan and “ethically” attack enterprise networks. Robotics farming company Iron Ox raised $20 million in a funding round. The system of farming robots is still semi-autonomous, the company’s goal is to become fully autonomous.  Insurtech company Descartes Underwriting raised $18.5 million. The company applies AI and machine learning technologies to climate risk predicting and insurance underwriting. Legaltech startup ThoughtRiver raised $10 million in its Series A round. Its AI solution applied to contract pre-screening aims to boost operational efficiency. Medtech startup Skin Analytics raised $5.1 million in Series A funding. Skin Analytics has developed a clinically validated AI system that can identify not only the important skin cancers but also precancerous lesions that can be treated, as well as a range of lesions that are benign. Amazon, along with several government organizations and three other industry partners, helped fund the National Science Foundation, a high-priority AI research initiative. The amount of funding is not disclosed. The content of TheSequence is written by Jesus Rodriguez, one of the most-read contributors to KDNuggets and TDS. You can check his Medium here.

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!
reddit
LLM Vibe Score0
Human Vibe Score-0.5
regalalgorithmThis week

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!

Hi there, we at Skynet Today produce a weekly newsletter summarizing each week's major AI news, which seems like it'd be of interest to this subreddit. Here's what's in our latest one: Facebook’s AI for detecting hate speech is facing its biggest challenge yet Facebook has made significant progress recently to proactively take down content that violate its community standards. For example, in the second quarter of 2020, Facebook took down 104.6 million pieces of content. While reviews are typically performed by a vast workforce of human moderators, AI-powered tools have enabled Facebook to do this work at a greater scale for textual content. However, there’s a long way to go for these systems to match or exceed the capabilities of human moderators. This is because a large proportion of hate speech and misinformation is in the form of images and memes, and reasoning about the context and language-image interplay is an extremely difficult challenge for AI. Given Facebook’s scale and the speed at which some use it to spread hate, incite violence, and share lies with millions, Facebook will have to keep running to catch up. AI Slays Top F-16 Pilot In DARPA Dogfight Simulation The Defense Advanced Research Project Agency (DARPA) recently hosted a simulated F16 dogfight competition, with different AI bots competing with each other as well as with human pilots. The top AI bot was able to beat a human pilot 5-0 in the simulated contest. DARPA started this program “as a risk-reduction effort \[…\] to flesh out how human and machine pilots share operational control of a fighter jet to maximize its chances of mission success.” Competition runners are broadly optimistic about the demonstration of AI capabilities, even if they are not close to being deployed on a real aircraft. Of concern, the program had little discussion on the ethics of AI military applications, especially with the lethal autonomous weapon systems being considered. News Advances & Business Microsoft, Energy Dept. to Develop Disaster-Response AI Tools \- The U.S. Department of Energy and Microsoft Corp. on Tuesday announced a partnership to develop artificial-intelligence tools aimed at helping first-responders better react to fast-changing natural events, such as floods and wildfires. Coronavirus: Robot CERi is a bilingual Covid-19 expert \- Ceri is bilingual, clued-up on coronavirus and can tell what mood you are in. Ceri also happens to be a robot. Moscow DOH uses AI platform to detect lung cancer symptoms \- Moscow’s department of health is using an artificial intelligence (AI) platform to detect symptoms of lung cancer in CT scans, as part of a project to implement AI technology for radiology. Scientists develop artificial intelligence system for high precision recognition of hand gestures \- The recognition of human hand gestures by AI systems has been a valuable development over the last decade and has been adopted in high-precision surgical robots, health monitoring equipment and in gaming systems. Forget credit cards - now you can pay with your face. Creepy or cool? \- A new way to pay has arrived in Los Angeles: your face. Concerns & Hype The dystopian tech that companies are selling to help schools reopen sooner \- This fall, AI could be watching students social distance and checking their masks. Thousands of schools nationwide will not be reopening this fall. NYPD Used Facial Recognition Technology In Siege Of Black Lives Matter Activist’s Apartment \- The NYPD deployed facial recognition technology in its hunt for a prominent Black Lives Matter activist, whose home was besieged by dozens of officers and police dogs last week, a spokesperson confirmed to Gothamist. Machines can spot mental health issues - if you hand over your personal data \- Digital diagnosis could transform psychiatry by mining your most intimate data for clues. But is the privacy cost worth it? Supporting Black Artists Who Are Examining AI \- Technology has a complicated relationship with racial justice. Smartphones, internet platforms, and other digital tools can be used to document and expose racism. But digital tools can also fuel racism: smart doorbells surveil Black individuals. A-level and GCSE results in England to be based on teacher assessments in U-turn \- All A-level and GCSE results in England will be based on grades assesed by teachers instead of algorithms. Analysis & Policy GPT-3 and The Question of Automation \- Automation is not an all or nothing proposition. An AI model’s automation capability is highly conjoined with the task and application it is used in. An A.I. Movie Service Could One Day Serve You a New Custom Film Every Time \- How long will it be until an A.I. can make an actual feature film on demand? Fairness, evidence, and predictive equality \- How the causal fairness principle relates to predictive equality How robotics and automation could create new jobs in the new normal \- Depending on who you ask, AI and automation will either destroy jobs or create new ones. In reality, a greater push toward automation will probably both kill and create jobs - human workers will become redundant in certain spheres, sure, but many new roles will likely crop up. Expert Opinions & Discussion within the field Too many AI researchers think real-world problems are not relevant \- The community’s hyperfocus on novel methods ignores what’s really important.

[P] Building a Code Search Engine for an AI-powered Junior Developer
reddit
LLM Vibe Score0
Human Vibe Score0
williamsweepThis week

[P] Building a Code Search Engine for an AI-powered Junior Developer

The last month building Sweep has been fun. We’ve dealt with countless formatting errors, irrelevant search results, and LLM hallucinations. Sweep is an open source AI-powered junior developer. We take your codebase and provide it as context to GPT to solve small requests related to your code. Code Search Code search is a key part of working with LLMs to automate programming. We used small language models to perform code retrieval(aka semantic search), which comes with several benefits (to be discussed in a later post!). However, one shortcoming of pure semantic search is distinguishing between two similar pieces of code in a vacuum. Example Take the following code snippets: Code Snippet A: accesstoken = os.environ.get("ACCESSTOKEN") g = Github(access_token) repo_name = "sweepai/bot-internal" issue_url = "github.com/sweepai/bot-internal/issues/28" username = "wwzeng1" repo_description = "A repo for Sweep" title = "Sweep: Use loguru.info to show the number of tokens in the anthropic call" summary = "" replies_text = "" Code Snippet B: g = getgithubclient(installation_id) if comment_id: logger.info(f"Replying to comment {comment_id}...") logger.info(f"Getting repo {repofullname}") repo = g.getrepo(repofull_name) currentissue = repo.getissue(number=issue_number) if current_issue.state == 'closed': posthog.capture(username, "issue_closed", properties=metadata) return {"success": False, "reason": "Issue is closed"} Explanation It might not be clear which file is more important, but Code Snippet A is from test\pr\diffs.py#L63-L71 (a test I wrote that’s no longer used), while B is from on\ticket.py#L87-L96 (our core logic for handling tickets). Since Code Snippet B is in an often used file, it is likely that this snippet will be more relevant as input to the LLM. Problem How can we differentiate between these two pieces of code when they’re both so similar? They both discuss issues, repositories, and some usernames. If the user asks “How can I change the username when creating an issue” it will be hard to differentiate between these two. Solution The trick is a ranking model. An important piece of ranking results is the concept of “quality”, i.e. what makes a file or snippet of code intrinsically valuable to the user. The results from our vector search model are a list of items (test\pr\diffs.py#L63-L71, on\ticket.py#L87C1-L96C63) and similarity scores (0.65, 0.63). By combining intuition and attention to the data, we can create a ranking model that is “personalized” for each repository we onboard. Ideas File Length Up to a point, longer files are generally more valuable for search. A 20-line file is probably not valuable unless the user specifically asks for it. However, 2000-line config files should not be ranked much higher either. linecountscore = min(line_count / 20, 10) Number of Commits The more commits a file has, the more valuable it is. This lets us distinguish between one off tests and core logic (which should receive the majority of commits). commitscore = numcommits + 1 Recency of changes The more recently a file was modified, the better. recencyscore = hourssincelastmodified + 1 Scoring To get the final score, we normalize and multiply these three scores together and add the similarity score. qualityscore = linecountscore * commitscore / recency_score finalscore = qualityscore/max(qualityscore) + similarityscore This solution usually worked fine, but we saw the same unexpected files showing up often. The max normalization was not enough. We fixed this by squashing the scores into percentiles, and then capping the increase at .25. In this case, the best result gets a .25 boost and the worst gets no boost. This lets us avoid fetching tests and configs which seem similar, and instead fetch business logic that actually helps Sweep write code! Sweep GitHub If this was interesting, take a look through our github repo (and give it a star!).https://github.com/sweepai/sweep

[P] I built an open SotA image tagging model to do what CLIP won't
reddit
LLM Vibe Score0
Human Vibe Score1
fpgaminerThis week

[P] I built an open SotA image tagging model to do what CLIP won't

I'm a hobbyist ML researcher and finally, after a year of work, built a state of the art machine vision model from scratch. It's ViT-B/16 based, 448x448x3 input, 91M parameters, trained for 660M samples, with multi-label classification as the target task, on over 5000 unique tags. All the big foundation vision models today were trained on heavily filtered datasets, greatly limiting the concepts they can represent, in line with arbitrary sets of rules for what is deemed "wholesome" by leading tech companies. Everything from innocuous to spicy is on the chopping block of those filters. And because CLIP pervades the industry, from StableDiffusion to LLaVA, so does OpenAI's sensibilities. My goal was to build a vision model for tagging images, mainly for labelling images for SD finetunes, but which wasn't as heavily filtered and handicapped as CLIP/BLIP/LLaVA. Something more inclusive, diverse, and sex positive. Starting from the wonderful work of SmilingWolf (https://github.com/SmilingWolf/SW-CV-ModelZoo) and the Danbooru2021 dataset, I iterated for a year on the model, training, and manually labeling a thousand images to help the model generalize beyond the danbooru domain. I'm releasing the first version of this model, dubbed JoyTag, today: https://github.com/fpgaminer/joytag It achieves a mean F1 score of 0.578 across all of its over 5000 tags and across both the anime/manga styled images of the original danbooru dataset, but also photographs and other mediums thanks to the auxiliary training data I provided to it. It was quite the struggle getting to this point, and I probably spent more time and money than any sane person should have. I learned a lot about dealing with datasets as large as danbooru2021, training models at scale, and how to keep yourself awake all night so your 8xA100 rental doesn't crash and blow all your money. In my manual testing outside of even the validation set, the model has generalized well to unseen images, so I'm quite happy with the results thus far. There's plenty more work to do expanding its dataset to improve that F1 score further, and roundout its weak points. With inclusivity and diversity being a major goal of this project, I'm disappointed by some of its remaining limitations (as documented in the GitHub README). But I'm already busy manually tagging more images using my model-augmented workflow. I'm happy to answer questions about the project, the training procedure, anything. All the training parameters are documented on GitHub, but there are so many little details that were hard won over the year. Like that damned loss multiplier. Ugh. Github: https://github.com/fpgaminer/joytag Model download: https://huggingface.co/fancyfeast/joytag/tree/main Demo: https://huggingface.co/spaces/fancyfeast/joytag

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.
reddit
LLM Vibe Score0
Human Vibe Score0.765
hardmaruThis week

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.

Schmidhuber interview expressing his views on the future of AI and AGI. Original source. I think the interview is of interest to r/MachineLearning, and presents an alternate view, compared to other influential leaders in AI. Juergen Schmidhuber, Renowned 'Father Of Modern AI,' Says His Life’s Work Won't Lead To Dystopia May 23, 2023. Contributed by Hessie Jones. Amid the growing concern about the impact of more advanced artificial intelligence (AI) technologies on society, there are many in the technology community who fear the implications of the advancements in Generative AI if they go unchecked. Dr. Juergen Schmidhuber, a renowned scientist, artificial intelligence researcher and widely regarded as one of the pioneers in the field, is more optimistic. He declares that many of those who suddenly warn against the dangers of AI are just seeking publicity, exploiting the media’s obsession with killer robots which has attracted more attention than “good AI” for healthcare etc. The potential to revolutionize various industries and improve our lives is clear, as are the equal dangers if bad actors leverage the technology for personal gain. Are we headed towards a dystopian future, or is there reason to be optimistic? I had a chance to sit down with Dr. Juergen Schmidhuber to understand his perspective on this seemingly fast-moving AI-train that will leap us into the future. As a teenager in the 1970s, Juergen Schmidhuber became fascinated with the idea of creating intelligent machines that could learn and improve on their own, becoming smarter than himself within his lifetime. This would ultimately lead to his groundbreaking work in the field of deep learning. In the 1980s, he studied computer science at the Technical University of Munich (TUM), where he earned his diploma in 1987. His thesis was on the ultimate self-improving machines that, not only, learn through some pre-wired human-designed learning algorithm, but also learn and improve the learning algorithm itself. Decades later, this became a hot topic. He also received his Ph.D. at TUM in 1991 for work that laid some of the foundations of modern AI. Schmidhuber is best known for his contributions to the development of recurrent neural networks (RNNs), the most powerful type of artificial neural network that can process sequential data such as speech and natural language. With his students Sepp Hochreiter, Felix Gers, Alex Graves, Daan Wierstra, and others, he published architectures and training algorithms for the long short-term memory (LSTM), a type of RNN that is widely used in natural language processing, speech recognition, video games, robotics, and other applications. LSTM has become the most cited neural network of the 20th century, and Business Week called it "arguably the most commercial AI achievement." Throughout his career, Schmidhuber has received various awards and accolades for his groundbreaking work. In 2013, he was awarded the Helmholtz Prize, which recognizes significant contributions to the field of machine learning. In 2016, he was awarded the IEEE Neural Network Pioneer Award for "pioneering contributions to deep learning and neural networks." The media have often called him the “father of modern AI,” because the most cited neural networks all build on his lab’s work. He is quick to point out, however, that AI history goes back centuries. Despite his many accomplishments, at the age of 60, he feels mounting time pressure towards building an Artificial General Intelligence within his lifetime and remains committed to pushing the boundaries of AI research and development. He is currently director of the KAUST AI Initiative, scientific director of the Swiss AI Lab IDSIA, and co-founder and chief scientist of AI company NNAISENSE, whose motto is "AI∀" which is a math-inspired way of saying "AI For All." He continues to work on cutting-edge AI technologies and applications to improve human health and extend human lives and make lives easier for everyone. The following interview has been edited for clarity. Jones: Thank you Juergen for joining me. You have signed letters warning about AI weapons. But you didn't sign the recent publication, "Pause Gigantic AI Experiments: An Open Letter"? Is there a reason? Schmidhuber: Thank you Hessie. Glad to speak with you. I have realized that many of those who warn in public against the dangers of AI are just seeking publicity. I don't think the latest letter will have any significant impact because many AI researchers, companies, and governments will ignore it completely. The proposal frequently uses the word "we" and refers to "us," the humans. But as I have pointed out many times in the past, there is no "we" that everyone can identify with. Ask 10 different people, and you will hear 10 different opinions about what is "good." Some of those opinions will be completely incompatible with each other. Don't forget the enormous amount of conflict between the many people. The letter also says, "If such a pause cannot be quickly put in place, governments should intervene and impose a moratorium." The problem is that different governments have ALSO different opinions about what is good for them and for others. Great Power A will say, if we don't do it, Great Power B will, perhaps secretly, and gain an advantage over us. The same is true for Great Powers C and D. Jones: Everyone acknowledges this fear surrounding current generative AI technology. Moreover, the existential threat of this technology has been publicly acknowledged by Sam Altman, CEO of OpenAI himself, calling for AI regulation. From your perspective, is there an existential threat? Schmidhuber: It is true that AI can be weaponized, and I have no doubt that there will be all kinds of AI arms races, but AI does not introduce a new quality of existential threat. The threat coming from AI weapons seems to pale in comparison to the much older threat from nuclear hydrogen bombs that don’t need AI at all. We should be much more afraid of half-century-old tech in the form of H-bomb rockets. The Tsar Bomba of 1961 had almost 15 times more destructive power than all weapons of WW-II combined. Despite the dramatic nuclear disarmament since the 1980s, there are still more than enough nuclear warheads to wipe out human civilization within two hours, without any AI I’m much more worried about that old existential threat than the rather harmless AI weapons. Jones: I realize that while you compare AI to the threat of nuclear bombs, there is a current danger that a current technology can be put in the hands of humans and enable them to “eventually” exact further harms to individuals of group in a very precise way, like targeted drone attacks. You are giving people a toolset that they've never had before, enabling bad actors, as some have pointed out, to be able to do a lot more than previously because they didn't have this technology. Schmidhuber: Now, all that sounds horrible in principle, but our existing laws are sufficient to deal with these new types of weapons enabled by AI. If you kill someone with a gun, you will go to jail. Same if you kill someone with one of these drones. Law enforcement will get better at understanding new threats and new weapons and will respond with better technology to combat these threats. Enabling drones to target persons from a distance in a way that requires some tracking and some intelligence to perform, which has traditionally been performed by skilled humans, to me, it seems is just an improved version of a traditional weapon, like a gun, which is, you know, a little bit smarter than the old guns. But, in principle, all of that is not a new development. For many centuries, we have had the evolution of better weaponry and deadlier poisons and so on, and law enforcement has evolved their policies to react to these threats over time. So, it's not that we suddenly have a new quality of existential threat and it's much more worrisome than what we have had for about six decades. A large nuclear warhead doesn’t need fancy face recognition to kill an individual. No, it simply wipes out an entire city with ten million inhabitants. Jones: The existential threat that’s implied is the extent to which humans have control over this technology. We see some early cases of opportunism which, as you say, tends to get more media attention than positive breakthroughs. But you’re implying that this will all balance out? Schmidhuber: Historically, we have a long tradition of technological breakthroughs that led to advancements in weapons for the purpose of defense but also for protection. From sticks, to rocks, to axes to gunpowder to cannons to rockets… and now to drones… this has had a drastic influence on human history but what has been consistent throughout history is that those who are using technology to achieve their own ends are themselves, facing the same technology because the opposing side is learning to use it against them. And that's what has been repeated in thousands of years of human history and it will continue. I don't see the new AI arms race as something that is remotely as existential a threat as the good old nuclear warheads. You said something important, in that some people prefer to talk about the downsides rather than the benefits of this technology, but that's misleading, because 95% of all AI research and AI development is about making people happier and advancing human life and health. Jones: Let’s touch on some of those beneficial advances in AI research that have been able to radically change present day methods and achieve breakthroughs. Schmidhuber: All right! For example, eleven years ago, our team with my postdoc Dan Ciresan was the first to win a medical imaging competition through deep learning. We analyzed female breast cells with the objective to determine harmless cells vs. those in the pre-cancer stage. Typically, a trained oncologist needs a long time to make these determinations. Our team, who knew nothing about cancer, were able to train an artificial neural network, which was totally dumb in the beginning, on lots of this kind of data. It was able to outperform all the other methods. Today, this is being used not only for breast cancer, but also for radiology and detecting plaque in arteries, and many other things. Some of the neural networks that we have developed in the last 3 decades are now prevalent across thousands of healthcare applications, detecting Diabetes and Covid-19 and what not. This will eventually permeate across all healthcare. The good consequences of this type of AI are much more important than the click-bait new ways of conducting crimes with AI. Jones: Adoption is a product of reinforced outcomes. The massive scale of adoption either leads us to believe that people have been led astray, or conversely, technology is having a positive effect on people’s lives. Schmidhuber: The latter is the likely case. There's intense commercial pressure towards good AI rather than bad AI because companies want to sell you something, and you are going to buy only stuff you think is going to be good for you. So already just through this simple, commercial pressure, you have a tremendous bias towards good AI rather than bad AI. However, doomsday scenarios like in Schwarzenegger movies grab more attention than documentaries on AI that improve people’s lives. Jones: I would argue that people are drawn to good stories – narratives that contain an adversary and struggle, but in the end, have happy endings. And this is consistent with your comment on human nature and how history, despite its tendency for violence and destruction of humanity, somehow tends to correct itself. Let’s take the example of a technology, which you are aware – GANs – General Adversarial Networks, which today has been used in applications for fake news and disinformation. In actuality, the purpose in the invention of GANs was far from what it is used for today. Schmidhuber: Yes, the name GANs was created in 2014 but we had the basic principle already in the early 1990s. More than 30 years ago, I called it artificial curiosity. It's a very simple way of injecting creativity into a little two network system. This creative AI is not just trying to slavishly imitate humans. Rather, it’s inventing its own goals. Let me explain: You have two networks. One network is producing outputs that could be anything, any action. Then the second network is looking at these actions and it’s trying to predict the consequences of these actions. An action could move a robot, then something happens, and the other network is just trying to predict what will happen. Now we can implement artificial curiosity by reducing the prediction error of the second network, which, at the same time, is the reward of the first network. The first network wants to maximize its reward and so it will invent actions that will lead to situations that will surprise the second network, which it has not yet learned to predict well. In the case where the outputs are fake images, the first network will try to generate images that are good enough to fool the second network, which will attempt to predict the reaction of the environment: fake or real image, and it will try to become better at it. The first network will continue to also improve at generating images whose type the second network will not be able to predict. So, they fight each other. The 2nd network will continue to reduce its prediction error, while the 1st network will attempt to maximize it. Through this zero-sum game the first network gets better and better at producing these convincing fake outputs which look almost realistic. So, once you have an interesting set of images by Vincent Van Gogh, you can generate new images that leverage his style, without the original artist having ever produced the artwork himself. Jones: I see how the Van Gogh example can be applied in an education setting and there are countless examples of artists mimicking styles from famous painters but image generation from this instance that can happen within seconds is quite another feat. And you know this is how GANs has been used. What’s more prevalent today is a socialized enablement of generating images or information to intentionally fool people. It also surfaces new harms that deal with the threat to intellectual property and copyright, where laws have yet to account for. And from your perspective this was not the intention when the model was conceived. What was your motivation in your early conception of what is now GANs? Schmidhuber: My old motivation for GANs was actually very important and it was not to create deepfakes or fake news but to enable AIs to be curious and invent their own goals, to make them explore their environment and make them creative. Suppose you have a robot that executes one action, then something happens, then it executes another action, and so on, because it wants to achieve certain goals in the environment. For example, when the battery is low, this will trigger “pain” through hunger sensors, so it wants to go to the charging station, without running into obstacles, which will trigger other pain sensors. It will seek to minimize pain (encoded through numbers). Now the robot has a friend, the second network, which is a world model ––it’s a prediction machine that learns to predict the consequences of the robot’s actions. Once the robot has a good model of the world, it can use it for planning. It can be used as a simulation of the real world. And then it can determine what is a good action sequence. If the robot imagines this sequence of actions, the model will predict a lot of pain, which it wants to avoid. If it plays this alternative action sequence in its mental model of the world, then it will predict a rewarding situation where it’s going to sit on the charging station and its battery is going to load again. So, it'll prefer to execute the latter action sequence. In the beginning, however, the model of the world knows nothing, so how can we motivate the first network to generate experiments that lead to data that helps the world model learn something it didn’t already know? That’s what artificial curiosity is about. The dueling two network systems effectively explore uncharted environments by creating experiments so that over time the curious AI gets a better sense of how the environment works. This can be applied to all kinds of environments, and has medical applications. Jones: Let’s talk about the future. You have said, “Traditional humans won’t play a significant role in spreading intelligence across the universe.” Schmidhuber: Let’s first conceptually separate two types of AIs. The first type of AI are tools directed by humans. They are trained to do specific things like accurately detect diabetes or heart disease and prevent attacks before they happen. In these cases, the goal is coming from the human. More interesting AIs are setting their own goals. They are inventing their own experiments and learning from them. Their horizons expand and eventually they become more and more general problem solvers in the real world. They are not controlled by their parents, but much of what they learn is through self-invented experiments. A robot, for example, is rotating a toy, and as it is doing this, the video coming in through the camera eyes, changes over time and it begins to learn how this video changes and learns how the 3D nature of the toy generates certain videos if you rotate it a certain way, and eventually, how gravity works, and how the physics of the world works. Like a little scientist! And I have predicted for decades that future scaled-up versions of such AI scientists will want to further expand their horizons, and eventually go where most of the physical resources are, to build more and bigger AIs. And of course, almost all of these resources are far away from earth out there in space, which is hostile to humans but friendly to appropriately designed AI-controlled robots and self-replicating robot factories. So here we are not talking any longer about our tiny biosphere; no, we are talking about the much bigger rest of the universe. Within a few tens of billions of years, curious self-improving AIs will colonize the visible cosmos in a way that’s infeasible for humans. Those who don’t won’t have an impact. Sounds like science fiction, but since the 1970s I have been unable to see a plausible alternative to this scenario, except for a global catastrophe such as an all-out nuclear war that stops this development before it takes off. Jones: How long have these AIs, which can set their own goals — how long have they existed? To what extent can they be independent of human interaction? Schmidhuber: Neural networks like that have existed for over 30 years. My first simple adversarial neural network system of this kind is the one from 1990 described above. You don’t need a teacher there; it's just a little agent running around in the world and trying to invent new experiments that surprise its own prediction machine. Once it has figured out certain parts of the world, the agent will become bored and will move on to more exciting experiments. The simple 1990 systems I mentioned have certain limitations, but in the past three decades, we have also built more sophisticated systems that are setting their own goals and such systems I think will be essential for achieving true intelligence. If you are only imitating humans, you will never go beyond them. So, you really must give AIs the freedom to explore previously unexplored regions of the world in a way that no human is really predefining. Jones: Where is this being done today? Schmidhuber: Variants of neural network-based artificial curiosity are used today for agents that learn to play video games in a human-competitive way. We have also started to use them for automatic design of experiments in fields such as materials science. I bet many other fields will be affected by it: chemistry, biology, drug design, you name it. However, at least for now, these artificial scientists, as I like to call them, cannot yet compete with human scientists. I don’t think it’s going to stay this way but, at the moment, it’s still the case. Sure, AI has made a lot of progress. Since 1997, there have been superhuman chess players, and since 2011, through the DanNet of my team, there have been superhuman visual pattern recognizers. But there are other things where humans, at the moment at least, are much better, in particular, science itself. In the lab we have many first examples of self-directed artificial scientists, but they are not yet convincing enough to appear on the radar screen of the public space, which is currently much more fascinated with simpler systems that just imitate humans and write texts based on previously seen human-written documents. Jones: You speak of these numerous instances dating back 30 years of these lab experiments where these self-driven agents are deciding and learning and moving on once they’ve learned. And I assume that that rate of learning becomes even faster over time. What kind of timeframe are we talking about when this eventually is taken outside of the lab and embedded into society? Schmidhuber: This could still take months or even years :-) Anyway, in the not-too-distant future, we will probably see artificial scientists who are good at devising experiments that allow them to discover new, previously unknown physical laws. As always, we are going to profit from the old trend that has held at least since 1941: every decade compute is getting 100 times cheaper. Jones: How does this trend affect modern AI such as ChatGPT? Schmidhuber: Perhaps you know that all the recent famous AI applications such as ChatGPT and similar models are largely based on principles of artificial neural networks invented in the previous millennium. The main reason why they works so well now is the incredible acceleration of compute per dollar. ChatGPT is driven by a neural network called “Transformer” described in 2017 by Google. I am happy about that because a quarter century earlier in 1991 I had a particular Transformer variant which is now called the “Transformer with linearized self-attention”. Back then, not much could be done with it, because the compute cost was a million times higher than today. But today, one can train such models on half the internet and achieve much more interesting results. Jones: And for how long will this acceleration continue? Schmidhuber: There's no reason to believe that in the next 30 years, we won't have another factor of 1 million and that's going to be really significant. In the near future, for the first time we will have many not-so expensive devices that can compute as much as a human brain. The physical limits of computation, however, are much further out so even if the trend of a factor of 100 every decade continues, the physical limits (of 1051 elementary instructions per second and kilogram of matter) won’t be hit until, say, the mid-next century. Even in our current century, however, we’ll probably have many machines that compute more than all 10 billion human brains collectively and you can imagine, everything will change then! Jones: That is the big question. Is everything going to change? If so, what do you say to the next generation of leaders, currently coming out of college and university. So much of this change is already impacting how they study, how they will work, or how the future of work and livelihood is defined. What is their purpose and how do we change our systems so they will adapt to this new version of intelligence? Schmidhuber: For decades, people have asked me questions like that, because you know what I'm saying now, I have basically said since the 1970s, it’s just that today, people are paying more attention because, back then, they thought this was science fiction. They didn't think that I would ever come close to achieving my crazy life goal of building a machine that learns to become smarter than myself such that I can retire. But now many have changed their minds and think it's conceivable. And now I have two daughters, 23 and 25. People ask me: what do I tell them? They know that Daddy always said, “It seems likely that within your lifetimes, you will have new types of intelligence that are probably going to be superior in many ways, and probably all kinds of interesting ways.” How should they prepare for that? And I kept telling them the obvious: Learn how to learn new things! It's not like in the previous millennium where within 20 years someone learned to be a useful member of society, and then took a job for 40 years and performed in this job until she received her pension. Now things are changing much faster and we must learn continuously just to keep up. I also told my girls that no matter how smart AIs are going to get, learn at least the basics of math and physics, because that’s the essence of our universe, and anybody who understands this will have an advantage, and learn all kinds of new things more easily. I also told them that social skills will remain important, because most future jobs for humans will continue to involve interactions with other humans, but I couldn’t teach them anything about that; they know much more about social skills than I do. You touched on the big philosophical question about people’s purpose. Can this be answered without answering the even grander question: What’s the purpose of the entire universe? We don’t know. But what’s happening right now might be connected to the unknown answer. Don’t think of humans as the crown of creation. Instead view human civilization as part of a much grander scheme, an important step (but not the last one) on the path of the universe from very simple initial conditions towards more and more unfathomable complexity. Now it seems ready to take its next step, a step comparable to the invention of life itself over 3.5 billion years ago. Alas, don’t worry, in the end, all will be good! Jones: Let’s get back to this transformation happening right now with OpenAI. There are many questioning the efficacy and accuracy of ChatGPT, and are concerned its release has been premature. In light of the rampant adoption, educators have banned its use over concerns of plagiarism and how it stifles individual development. Should large language models like ChatGPT be used in school? Schmidhuber: When the calculator was first introduced, instructors forbade students from using it in school. Today, the consensus is that kids should learn the basic methods of arithmetic, but they should also learn to use the “artificial multipliers” aka calculators, even in exams, because laziness and efficiency is a hallmark of intelligence. Any intelligent being wants to minimize its efforts to achieve things. And that's the reason why we have tools, and why our kids are learning to use these tools. The first stone tools were invented maybe 3.5 million years ago; tools just have become more sophisticated over time. In fact, humans have changed in response to the properties of their tools. Our anatomical evolution was shaped by tools such as spears and fire. So, it's going to continue this way. And there is no permanent way of preventing large language models from being used in school. Jones: And when our children, your children graduate, what does their future work look like? Schmidhuber: A single human trying to predict details of how 10 billion people and their machines will evolve in the future is like a single neuron in my brain trying to predict what the entire brain and its tens of billions of neurons will do next year. 40 years ago, before the WWW was created at CERN in Switzerland, who would have predicted all those young people making money as YouTube video bloggers? Nevertheless, let’s make a few limited job-related observations. For a long time, people have thought that desktop jobs may require more intelligence than skills trade or handicraft professions. But now, it turns out that it's much easier to replace certain aspects of desktop jobs than replacing a carpenter, for example. Because everything that works well in AI is happening behind the screen currently, but not so much in the physical world. There are now artificial systems that can read lots of documents and then make really nice summaries of these documents. That is a desktop job. Or you give them a description of an illustration that you want to have for your article and pretty good illustrations are being generated that may need some minimal fine-tuning. But you know, all these desktop jobs are much easier to facilitate than the real tough jobs in the physical world. And it's interesting that the things people thought required intelligence, like playing chess, or writing or summarizing documents, are much easier for machines than they thought. But for things like playing football or soccer, there is no physical robot that can remotely compete with the abilities of a little boy with these skills. So, AI in the physical world, interestingly, is much harder than AI behind the screen in virtual worlds. And it's really exciting, in my opinion, to see that jobs such as plumbers are much more challenging than playing chess or writing another tabloid story. Jones: The way data has been collected in these large language models does not guarantee personal information has not been excluded. Current consent laws already are outdated when it comes to these large language models (LLM). The concern, rightly so, is increasing surveillance and loss of privacy. What is your view on this? Schmidhuber: As I have indicated earlier: are surveillance and loss of privacy inevitable consequences of increasingly complex societies? Super-organisms such as cities and states and companies consist of numerous people, just like people consist of numerous cells. These cells enjoy little privacy. They are constantly monitored by specialized "police cells" and "border guard cells": Are you a cancer cell? Are you an external intruder, a pathogen? Individual cells sacrifice their freedom for the benefits of being part of a multicellular organism. Similarly, for super-organisms such as nations. Over 5000 years ago, writing enabled recorded history and thus became its inaugural and most important invention. Its initial purpose, however, was to facilitate surveillance, to track citizens and their tax payments. The more complex a super-organism, the more comprehensive its collection of information about its constituents. 200 years ago, at least, the parish priest in each village knew everything about all the village people, even about those who did not confess, because they appeared in the confessions of others. Also, everyone soon knew about the stranger who had entered the village, because some occasionally peered out of the window, and what they saw got around. Such control mechanisms were temporarily lost through anonymization in rapidly growing cities but are now returning with the help of new surveillance devices such as smartphones as part of digital nervous systems that tell companies and governments a lot about billions of users. Cameras and drones etc. are becoming increasingly tinier and more ubiquitous. More effective recognition of faces and other detection technology are becoming cheaper and cheaper, and many will use it to identify others anywhere on earth; the big wide world will not offer any more privacy than the local village. Is this good or bad? Some nations may find it easier than others to justify more complex kinds of super-organisms at the expense of the privacy rights of their constituents. Jones: So, there is no way to stop or change this process of collection, or how it continuously informs decisions over time? How do you see governance and rules responding to this, especially amid Italy’s ban on ChatGPT following suspected user data breach and the more recent news about the Meta’s record $1.3billion fine in the company’s handling of user information? Schmidhuber: Data collection has benefits and drawbacks, such as the loss of privacy. How to balance those? I have argued for addressing this through data ownership in data markets. If it is true that data is the new oil, then it should have a price, just like oil. At the moment, the major surveillance platforms such as Meta do not offer users any money for their data and the transitive loss of privacy. In the future, however, we will likely see attempts at creating efficient data markets to figure out the data's true financial value through the interplay between supply and demand. Even some of the sensitive medical data should not be priced by governmental regulators but by patients (and healthy persons) who own it and who may sell or license parts thereof as micro-entrepreneurs in a healthcare data market. Following a previous interview, I gave for one of the largest re-insurance companies , let's look at the different participants in such a data market: patients, hospitals, data companies. (1) Patients with a rare form of cancer can offer more valuable data than patients with a very common form of cancer. (2) Hospitals and their machines are needed to extract the data, e.g., through magnet spin tomography, radiology, evaluations through human doctors, and so on. (3) Companies such as Siemens, Google or IBM would like to buy annotated data to make better artificial neural networks that learn to predict pathologies and diseases and the consequences of therapies. Now the market’s invisible hand will decide about the data’s price through the interplay between demand and supply. On the demand side, you will have several companies offering something for the data, maybe through an app on the smartphone (a bit like a stock market app). On the supply side, each patient in this market should be able to profit from high prices for rare valuable types of data. Likewise, competing data extractors such as hospitals will profit from gaining recognition and trust for extracting data well at a reasonable price. The market will make the whole system efficient through incentives for all who are doing a good job. Soon there will be a flourishing ecosystem of commercial data market advisors and what not, just like the ecosystem surrounding the traditional stock market. The value of the data won’t be determined by governments or ethics committees, but by those who own the data and decide by themselves which parts thereof they want to license to others under certain conditions. At first glance, a market-based system seems to be detrimental to the interest of certain monopolistic companies, as they would have to pay for the data - some would prefer free data and keep their monopoly. However, since every healthy and sick person in the market would suddenly have an incentive to collect and share their data under self-chosen anonymity conditions, there will soon be many more useful data to evaluate all kinds of treatments. On average, people will live longer and healthier, and many companies and the entire healthcare system will benefit. Jones: Finally, what is your view on open source versus the private companies like Google and OpenAI? Is there a danger to supporting these private companies’ large language models versus trying to keep these models open source and transparent, very much like what LAION is doing? Schmidhuber: I signed this open letter by LAION because I strongly favor the open-source movement. And I think it's also something that is going to challenge whatever big tech dominance there might be at the moment. Sure, the best models today are run by big companies with huge budgets for computers, but the exciting fact is that open-source models are not so far behind, some people say maybe six to eight months only. Of course, the private company models are all based on stuff that was created in academia, often in little labs without so much funding, which publish without patenting their results and open source their code and others take it and improved it. Big tech has profited tremendously from academia; their main achievement being that they have scaled up everything greatly, sometimes even failing to credit the original inventors. So, it's very interesting to see that as soon as some big company comes up with a new scaled-up model, lots of students out there are competing, or collaborating, with each other, trying to come up with equal or better performance on smaller networks and smaller machines. And since they are open sourcing, the next guy can have another great idea to improve it, so now there’s tremendous competition also for the big companies. Because of that, and since AI is still getting exponentially cheaper all the time, I don't believe that big tech companies will dominate in the long run. They find it very hard to compete with the enormous open-source movement. As long as you can encourage the open-source community, I think you shouldn't worry too much. Now, of course, you might say if everything is open source, then the bad actors also will more easily have access to these AI tools. And there's truth to that. But as always since the invention of controlled fire, it was good that knowledge about how technology works quickly became public such that everybody could use it. And then, against any bad actor, there's almost immediately a counter actor trying to nullify his efforts. You see, I still believe in our old motto "AI∀" or "AI For All." Jones: Thank you, Juergen for sharing your perspective on this amazing time in history. It’s clear that with new technology, the enormous potential can be matched by disparate and troubling risks which we’ve yet to solve, and even those we have yet to identify. If we are to dispel the fear of a sentient system for which we have no control, humans, alone need to take steps for more responsible development and collaboration to ensure AI technology is used to ultimately benefit society. Humanity will be judged by what we do next.

[D] A Jobless Rant - ML is a Fool's Gold
reddit
LLM Vibe Score0
Human Vibe Score1
good_riceThis week

[D] A Jobless Rant - ML is a Fool's Gold

Aside from the clickbait title, I am earnestly looking for some advice and discussion from people who are actually employed. That being said, here's my gripe: I have been relentlessly inundated by the words "AI, ML, Big Data" throughout my undergrad from other CS majors, business and sales oriented people, media, and .ai type startups. It seems like everyone was peddling ML as the go to solution, the big money earner, and the future of the field. I've heard college freshman ask stuff like, "if I want to do CS, am I going to need to learn ML to be relevant" - if you're on this sub, I probably do not need to continue to elaborate on just how ridiculous the ML craze is. Every single university has opened up ML departments or programs and are pumping out ML graduates at an unprecedented rate. Surely, there'd be a job market to meet the incredible supply of graduates and cultural interest? Swept up in a mixture of genuine interest and hype, I decided to pursue computer vision. I majored in Math-CS at a top-10 CS university (based on at least one arbitrary ranking). I had three computer vision internships, two at startups, one at NASA JPL, in each doing non-trivial CV work; I (re)implemented and integrated CV systems from mixtures of recently published papers. I have a bunch of projects showing both CV and CS fundamentals (OS, networking, data structures, algorithms, etc) knowledge. I have taken graduate level ML coursework. I was accepted to Carnegie Mellon for an MS in Computer Vision, but I deferred to 2021 - all in all, I worked my ass off to try to simultaneously get a solid background in math AND computer science AND computer vision. That brings me to where I am now, which is unemployed and looking for jobs. Almost every single position I have seen requires a PhD and/or 5+ years of experience, and whatever I have applied for has ghosted me so far. The notion that ML is a high paying in-demand field seems to only be true if your name is Andrej Karpathy - and I'm only sort of joking. It seems like unless you have a PhD from one of the big 4 in CS and multiple publications in top tier journals you're out of luck, or at least vying for one of the few remaining positions at small companies. This seems normalized in ML, but this is not the case for quite literally every other subfield or even generalized CS positions. Getting a high paying job at a Big N company is possible as a new grad with just a bachelors and general SWE knowledge, and there are a plethora of positions elsewhere. Getting the equivalent with basically every specialization, whether operating systems, distributed systems, security, networking, etc, is also possible, and doesn't require 5 CVPR publications. TL;DR From my personal perspective, if you want to do ML because of career prospects, salaries, or job security, pick almost any other CS specialization. In ML, you'll find yourself working 2x as hard through difficult theory and math to find yourself competing with more applicants for fewer positions. I am absolutely complaining and would love to hear a more positive perspective, but in the meanwhile I'll be applying to jobs, working on more post-grad projects, and contemplating switching fields.

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.
reddit
LLM Vibe Score0
Human Vibe Score0.6
AlexSnakeKingThis week

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.

TD;LR: At Company A, Team X does advanced analytics using on-prem ERP tools and older programming languages. Their tools work very well and are designed based on very deep business and domain expertise. Team Y is a new and ambitious Data Science team that thinks they can replace Team X's tools with a bunch of R scripts and a custom built ML platform. Their models are simplistic, but more "fashionable" compared to the econometric models used by Team X, and team Y benefits from the ML/DS moniker so leadership is allowing Team Y to start a large scale overhaul of the analytics platform in question. Team Y doesn't have the experience for such a larger scale transformation, and is refusing to collaborate with team X. This project is very likely going to fail, and cause serious harm to the company as a whole financially and from a people perspective. I argue that this is not just because of bad leadership, but also because of various trends and mindsets in the DS community at large. Update (Jump to below the line for the original story): Several people in the comments are pointing out that this just a management failure, not something due to ML/DS, and that you can replace DS with any buzz tech and the story will still be relevant. My response: Of course, any failure at an organization level is ultimately a management failure one way or the other. Moreover, it is also the case that ML/DS when done correctly, will always improve a company's bottom line. There is no scenario where the proper ML solution, delivered at a reasonable cost and in a timely fashion, will somehow hurt the company's bottom line. My point is that in this case management is failing because of certain trends and practices that are specific to the ML/DS community, namely: The idea that DS teams should operate independently of tech and business orgs -- too much autonomy for DS teams The disregard for domain knowledge that seems prevalent nowadays thanks to the ML hype, that DS can be generalists and someone with good enough ML chops can solve any business problem. That wasn't the case when I first left academia for the industry in 2009 (back then nobody would even bother with a phone screen if you didn't have the right domain knowledge). Over reliance on resources who check all the ML hype related boxes (knows Python, R, Tensorflow, Shiny, etc..., has the right Coursera certifications, has blogged on the topic, etc...), but are lacking in depth of experience. DS interviews nowadays all seem to be: Can you tell me what a p-value is? What is elastic net regression? Show me how to fit a model in sklearn? How do you impute NAs in an R dataframe? Any smart person can look those up on Stackoverflow or Cross-Validated,.....Instead teams should be asking stuff like: why does portfolio optimization use QP not LP? How does a forecast influence a customer service level? When should a recommendation engine be content based and when should it use collaborative filtering? etc... (This is a true story, happening to the company I currently work for. Names, domains, algorithms, and roles have been shuffled around to protect my anonymity)  Company A has been around for several decades. It is not the biggest name in its domain, but it is a well respected one. Risk analysis and portfolio optimization have been a core of Company A's business since the 90s. They have a large team of 30 or so analysts who perform those tasks on a daily basis. These analysts use ERP solutions implemented for them by one the big ERP companies (SAP, Teradata, Oracle, JD Edwards,...) or one of the major tech consulting companies (Deloitte, Accenture, PWC, Capgemini, etc...) in collaboration with their own in house engineering team. The tools used are embarrassingly old school: Classic RDBMS running on on-prem servers or maybe even on mainframes, code written in COBOL, Fortran, weird proprietary stuff like ABAP or SPSS.....you get the picture. But the models and analytic functions were pretty sophisticated, and surprisingly cutting edge compared to the published academic literature. Most of all, they fit well with the company's enterprise ecosystem, and were honed based on years of deep domain knowledge.  They have a tech team of several engineers (poached from the aforementioned software and consulting companies) and product managers (who came from the experienced pools of analysts and managers who use the software, or poached from business rivals) maintaining and running this software. Their technology might be old school, but collectively, they know the domain and the company's overall architecture very, very well. They've guided the company through several large scale upgrades and migrations and they have a track record of delivering on time, without too much overhead. The few times they've stumbled, they knew how to pick themselves up very quickly. In fact within their industry niche, they have a reputation for their expertise, and have very good relations with the various vendors they've had to deal with. They were the launching pad of several successful ERP consulting careers.  Interestingly, despite dealing on a daily basis with statistical modeling and optimization algorithms, none of the analysts, engineers, or product managers involved describe themselves as data scientists or machine learning experts. It is mostly a cultural thing: Their expertise predates the Data Science/ML hype that started circa 2010, and they got most of their chops using proprietary enterprise tools instead of the open source tools popular nowadays. A few of them have formal statistical training, but most of them came from engineering or domain backgrounds and learned stats on the fly while doing their job. Call this team "Team X".  Sometime around the mid 2010s, Company A started having some serious anxiety issues: Although still doing very well for a company its size, overall economic and demographic trends were shrinking its customer base, and a couple of so called disruptors came up with a new app and business model that started seriously eating into their revenue. A suitable reaction to appease shareholders and Wall Street was necessary. The company already had a decent website and a pretty snazzy app, what more could be done? Leadership decided that it was high time that AI and ML become a core part of the company's business. An ambitious Manager, with no science or engineering background, but who had very briefly toyed with a recommender system a couple of years back, was chosen to build a data science team, call it team "Y" (he had a bachelor's in history from the local state college and worked for several years in the company's marketing org). Team "Y" consists mostly of internal hires who decided they wanted to be data scientists and completed a Coursera certification or a Galvanize boot camp, before being brought on to the team, along with a few of fresh Ph.D or M.Sc holders who didn't like academia and wanted to try their hand at an industry role. All of them were very bright people, they could write great Medium blog posts and give inspiring TED talks, but collectively they had very little real world industry experience. As is the fashion nowadays, this group was made part of a data science org that reported directly to the CEO and Board, bypassing the CIO and any tech or business VPs, since Company A wanted to claim the monikers "data driven" and "AI powered" in their upcoming shareholder meetings. In 3 or 4 years of existence, team Y produced a few Python and R scripts. Their architectural experience  consisted almost entirely in connecting Flask to S3 buckets or Redshift tables, with a couple of the more resourceful ones learning how to plug their models into Tableau or how to spin up a Kuberneties pod.  But they needn't worry: The aforementioned manager, who was now a director (and was also doing an online Masters to make up for his qualifications gap and bolster his chances of becoming VP soon - at least he now understands what L1 regularization is), was a master at playing corporate politics and self-promotion. No matter how few actionable insights team Y produced or how little code they deployed to production, he always had their back and made sure they had ample funding. In fact he now had grandiose plans for setting up an all-purpose machine learning platform that can be used to solve all of the company's data problems.  A couple of sharp minded members of team Y, upon googling their industry name along with the word "data science", realized that risk analysis was a prime candidate for being solved with Bayesian models, and there was already a nifty R package for doing just that, whose tutorial they went through on R-Bloggers.com. One of them had even submitted a Bayesian classifier Kernel for a competition on Kaggle (he was 203rd on the leaderboard), and was eager to put his new-found expertise to use on a real world problem. They pitched the idea to their director, who saw a perfect use case for his upcoming ML platform. They started work on it immediately, without bothering to check whether anybody at Company A was already doing risk analysis. Since their org was independent, they didn't really need to check with anybody else before they got funding for their initiative. Although it was basically a Naive Bayes classifier, the term ML was added to the project tile, to impress the board.  As they progressed with their work however, tensions started to build. They had asked the data warehousing and CA analytics teams to build pipelines for them, and word eventually got out to team X about their project. Team X was initially thrilled: They offered to collaborate whole heartedly, and would have loved to add an ML based feather to their already impressive cap. The product owners and analysts were totally onboard as well: They saw a chance to get in on the whole Data Science hype that they kept hearing about. But through some weird mix of arrogance and insecurity, team Y refused to collaborate with them or share any of their long term goals with them, even as they went to other parts of the company giving brown bag presentations and tutorials on the new model they created.  Team X got resentful: from what they saw of team Y's model, their approach was hopelessly naive and had little chances of scaling or being sustainable in production, and they knew exactly how to help with that. Deploying the model to production would have taken them a few days, given how comfortable they were with DevOps and continuous delivery (team Y had taken several months to figure out how to deploy a simple R script to production). And despite how old school their own tech was, team X were crafty enough to be able to plug it in to their existing architecture. Moreover, the output of the model was such that it didn't take into account how the business will consume it or how it was going to be fed to downstream systems, and the product owners could have gone a long way in making the model more amenable to adoption by the business stakeholders. But team Y wouldn't listen, and their leads brushed off any attempts at communication, let alone collaboration. The vibe that team Y was giving off was "We are the cutting edge ML team, you guys are the legacy server grunts. We don't need your opinion.", and they seemed to have a complete disregard for domain knowledge, or worse, they thought that all that domain knowledge consisted of was being able to grasp the definitions of a few business metrics.  Team X got frustrated and tried to express their concerns to leadership. But despite owning a vital link in Company A's business process, they were only \~50 people in a large 1000 strong technology and operations org, and they were several layers removed from the C-suite, so it was impossible for them to get their voices heard.  Meanwhile, the unstoppable director was doing what he did best: Playing corporate politics. Despite how little his team had actually delivered, he had convinced the board that all analysis and optimization tasks should now be migrated to his yet to be delivered ML platform. Since most leaders now knew that there was overlap between team Y and team X's objectives, his pitch was no longer that team Y was going to create a new insight, but that they were going to replace (or modernize) the legacy statistics based on-prem tools with more accurate cloud based ML tools. Never mind that there was no support in the academic literature for the idea that Naive Bayes works better than the Econometric approaches used by team X, let alone the additional wacky idea that Bayesian Optimization would definitely outperform the QP solvers that were running in production.  Unbeknownst to team X, the original Bayesian risk analysis project has now grown into a multimillion dollar major overhaul initiative, which included the eventual replacement of all of the tools and functions supported by team X along with the necessary migration to the cloud. The CIO and a couple of business VPs are on now board, and tech leadership is treating it as a done deal. An outside vendor, a startup who nobody had heard of, was contracted to help build the platform, since team Y has no engineering skills. The choice was deliberate, as calling on any of the established consulting or software companies would have eventually led leadership to the conclusion that team X was better suited for a transformation on this scale than team Y.  Team Y has no experience with any major ERP deployments, and no domain knowledge, yet they are being tasked with fundamentally changing the business process that is at the core of Company A's business. Their models actually perform worse than those deployed by team X, and their architecture is hopelessly simplistic, compared to what is necessary for running such a solution in production.  Ironically, using Bayesian thinking and based on all the evidence, the likelihood that team Y succeeds is close to 0%. At best, the project is going to end up being a write off of 50 million dollars or more. Once the !@#$!@hits the fan, a couple of executive heads are going to role, and dozens of people will get laid off. At worst, given how vital risk analysis and portfolio optimization is to Company A's revenue stream, the failure will eventually sink the whole company. It probably won't go bankrupt, but it will lose a significant portion of its business and work force. Failed ERP implementations can and do sink large companies: Just see what happened to National Grid US, SuperValu or Target Canada.  One might argue that this is more about corporate disfunction and bad leadership than about data science and AI. But I disagree. I think the core driver of this debacle is indeed the blind faith in Data Scientists, ML models and the promise of AI, and the overall culture of hype and self promotion that is very common among the ML crowd.  We haven't seen the end of this story: I sincerely hope that this ends well for the sake of my colleagues and all involved. Company A is a good company, and both its customers and its employees deserver better. But the chances of that happening are negligible given all the information available, and this failure will hit my company hard.

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[D] Misuse of Deep Learning in Nature Journal’s Earthquake Aftershock Paper
reddit
LLM Vibe Score0
Human Vibe Score0.333
milaworldThis week

[D] Misuse of Deep Learning in Nature Journal’s Earthquake Aftershock Paper

Recently, I saw a post by Rajiv Shah, Chicago-based data-scientist, regarding an article published in Nature last year called Deep learning of aftershock patterns following large earthquakes, written by scientists at Harvard in collaboration with Google. Below is the article: Stand Up for Best Practices: Misuse of Deep Learning in Nature’s Earthquake Aftershock Paper The Dangers of Machine Learning Hype Practitioners of AI, machine learning, predictive modeling, and data science have grown enormously over the last few years. What was once a niche field defined by its blend of knowledge is becoming a rapidly growing profession. As the excitement around AI continues to grow, the new wave of ML augmentation, automation, and GUI tools will lead to even more growth in the number of people trying to build predictive models. But here’s the rub: While it becomes easier to use the tools of predictive modeling, predictive modeling knowledge is not yet a widespread commodity. Errors can be counterintuitive and subtle, and they can easily lead you to the wrong conclusions if you’re not careful. I’m a data scientist who works with dozens of expert data science teams for a living. In my day job, I see these teams striving to build high-quality models. The best teams work together to review their models to detect problems. There are many hard-to-detect-ways that lead to problematic models (say, by allowing target leakage into their training data). Identifying issues is not fun. This requires admitting that exciting results are “too good to be true” or that their methods were not the right approach. In other words, it’s less about the sexy data science hype that gets headlines and more about a rigorous scientific discipline. Bad Methods Create Bad Results Almost a year ago, I read an article in Nature that claimed unprecedented accuracy in predicting earthquake aftershocks by using deep learning. Reading the article, my internal radar became deeply suspicious of their results. Their methods simply didn’t carry many of the hallmarks of careful predicting modeling. I started to dig deeper. In the meantime, this article blew up and became widely recognized! It was even included in the release notes for Tensorflow as an example of what deep learning could do. However, in my digging, I found major flaws in the paper. Namely, data leakage which leads to unrealistic accuracy scores and a lack of attention to model selection (you don’t build a 6 layer neural network when a simpler model provides the same level of accuracy). To my earlier point: these are subtle, but incredibly basic predictive modeling errors that can invalidate the entire results of an experiment. Data scientists are trained to recognize and avoid these issues in their work. I assumed that this was simply overlooked by the author, so I contacted her and let her know so that she could improve her analysis. Although we had previously communicated, she did not respond to my email over concerns with the paper. Falling On Deaf Ears So, what was I to do? My coworkers told me to just tweet it and let it go, but I wanted to stand up for good modeling practices. I thought reason and best practices would prevail, so I started a 6-month process of writing up my results and shared them with Nature. Upon sharing my results, I received a note from Nature in January 2019 that despite serious concerns about data leakage and model selection that invalidate their experiment, they saw no need to correct the errors, because “Devries et al. are concerned primarily with using machine learning as [a] tool to extract insight into the natural world, and not with details of the algorithm design.” The authors provided a much harsher response. You can read the entire exchange on my github. It’s not enough to say that I was disappointed. This was a major paper (it’s Nature!) that bought into AI hype and published a paper despite it using flawed methods. Then, just this week, I ran across articles by Arnaud Mignan and Marco Broccardo on shortcomings that they found in the aftershocks article. Here are two more data scientists with expertise in earthquake analysis who also noticed flaws in the paper. I also have placed my analysis and reproducible code on github. Standing Up For Predictive Modeling Methods I want to make it clear: my goal is not to villainize the authors of the aftershocks paper. I don’t believe that they were malicious, and I think that they would argue their goal was to just show how machine learning could be applied to aftershocks. Devries is an accomplished earthquake scientist who wanted to use the latest methods for her field of study and found exciting results from it. But here’s the problem: their insights and results were based on fundamentally flawed methods. It’s not enough to say, “This isn’t a machine learning paper, it’s an earthquake paper.” If you use predictive modeling, then the quality of your results are determined by the quality of your modeling. Your work becomes data science work, and you are on the hook for your scientific rigor. There is a huge appetite for papers that use the latest technologies and approaches. It becomes very difficult to push back on these papers. But if we allow papers or projects with fundamental issues to advance, it hurts all of us. It undermines the field of predictive modeling. Please push back on bad data science. Report bad findings to papers. And if they don’t take action, go to twitter, post about it, share your results and make noise. This type of collective action worked to raise awareness of p-values and combat the epidemic of p-hacking. We need good machine learning practices if we want our field to continue to grow and maintain credibility. Link to Rajiv's Article Original Nature Publication (note: paywalled) GitHub repo contains an attempt to reproduce Nature's paper Confrontational correspondence with authors

[D] AI Agents: too early, too expensive, too unreliable
reddit
LLM Vibe Score0
Human Vibe Score1
madredditscientistThis week

[D] AI Agents: too early, too expensive, too unreliable

Reference: Full blog post There has been a lot of hype about the promise of autonomous agent-based LLM workflows. By now, all major LLMs are capable of interacting with external tools and functions, letting the LLM perform sequences of tasks automatically. But reality is proving more challenging than anticipated. The WebArena leaderboard, which benchmarks LLMs agents against real-world tasks, shows that even the best-performing models have a success rate of only 35.8%. Challenges in Practice After seeing many attempts to AI agents, I believe it's too early, too expensive, too slow, too unreliable. It feels like many AI agent startups are waiting for a model breakthrough that will start the race to productize agents. Reliability: As we all know, LLMs are prone to hallucinations and inconsistencies. Chaining multiple AI steps compounds these issues, especially for tasks requiring exact outputs. Performance and costs: GPT-4o, Gemini-1.5, and Claude Opus are working quite well with tool usage/function calling, but they are still slow and expensive, particularly if you need to do loops and automatic retries. Legal concerns: Companies may be held liable for the mistakes of their agents. A recent example is Air Canada being ordered to pay a customer who was misled by the airline's chatbot. User trust: The "black box" nature of AI agents and stories like the above makes it hard for users to understand and trust their outputs. Gaining user trust for sensitive tasks involving payments or personal information will be hard (paying bills, shopping, etc.). Real-World Attempts Several startups are tackling the AI agent space, but most are still experimental or invite-only: adept.ai - $350M funding, but access is still very limited MultiOn - funding unknown, their API-first approach seems promising HypeWrite - $2.8M funding, started with an AI writing assistant and expanded into the agent space minion.ai - created some initial buzz but has gone quiet now, waitlist only Only MultiOn seems to be pursuing the "give it instructions and watch it go" approach, which is more in line with the promise of AI agents. All others are going down the record-and-replay RPA route, which may be necessary for reliability at this stage. Large players are also bringing AI capabilities to desktops and browsers, and it looks like we'll get native AI integrations on a system level: OpenAI announced their Mac desktop app that can interact with the OS screen. At Google I/O, Google demonstrated Gemini automatically processing a shopping return. Microsoft announced Copilot Studio, which will let developers build AI agent bots. Screenshot Screenshot These tech demos are impressive, but we'll see how well these agent capabilities will work when released publicly and tested against real-world scenarios instead of hand-picked demo cases. The Path Forward AI agents overhyped and it's too early. However, the underlying models continue to advance quickly, and we can expect to see more successful real-world applications. Instead of trying to have one large general purpose agent that is hard to control and test, we can use many smaller agents that basically just pick the right strategy for a specific sub-task in our workflows. These "agents" can be thought of as medium-sized LLM prompts with a) context and b) a set of functions available to call. The most promising path forward likely looks like this: Narrowly scoped, well testable automations that use AI as an augmentation tool rather than pursuing full autonomy Human-in-the-loop approaches that keep humans involved for oversight and handling edge cases Setting realistic expectations about current capabilities and limitations By combining tightly constrained agents, good evaluation data, human-in-the-loop oversight, and traditional engineering methods, we can achieve reliably good results for automating medium-complex tasks. Will AI agents automate tedious repetitive work, such as web scraping, form filling, and data entry? Yes, absolutely. Will AI agents autonomously book your vacation without your intervention? Unlikely, at least in the near future.

I tested hundreds of marketing tools in the last three years and these 50 made it to the list. I'll sum up my top 50 marketing tools with one or two sentences + give you pricings.
reddit
LLM Vibe Score0
Human Vibe Score1
SpicyCopyThis week

I tested hundreds of marketing tools in the last three years and these 50 made it to the list. I'll sum up my top 50 marketing tools with one or two sentences + give you pricings.

Hey guys, I'm working in a growth marketing agency. Marketing tools are 30% of what we do, so we use them a lot and experiment with the new ones as much as possible. There are thousands of tools and it's easy to get lost, so I wanted to share the tools we use most on a daily basis. And divide the list into 14 categories. I thought this could be handy for Entrepreneurs subreddit. Why adopt tools? I see marketing tools as tireless colleagues. If you can't hire an employee, choosing the right tool can solve your problems, because they Are super cheap. Work 7/24 for you. Don’t make mistakes. Don’t need management. (or needless management) Help you to automate the majority of your lead gen process. Onwards to the list. (With the pricings post ended up quite long, you can find a link in the end if you want to check the prices) Email marketing tools #1 ActiveCampaign is armed with the most complicated email automation features and has the most intuitive user experience. It feels like you already know how to use it. \#2 Autopilot is visual marketing automation and customer journey tool that helps you acquire, nurture based on behaviors, interest etc. #3 Mailjet: This is the tool we use to send out bulky email campaigns such as newsletters. It doesn't have sexy features like others but does its job for a cheap price. Email address finders #4 Skrapp finds email of your contacts by name and company. It also works with LinkedIn Sales Navigator and can extract thousands of emails in bulk + have a browser add-on. #5 Hunter: Similar to Skrapp but doesn't work with LinkedIn Sales Navigator directly. In addition, there are email templates and you can set up email campaigns. Prospecting and outreach tools #6 Prospect combines the personal emails, follow-up calls, other social touches and helps you create multichannel campaigns.  #7 Reply is a more intuitive version of Prospect. It is easy to learn and use; their UX makes you feel good and sufficient.  CRM tools #8 Salesflare helps you to stop managing your data and start managing your customers. Not yet popular as Hubspot and etc but the best solution for smaller B2B businesses. (we're fans) \#9 Hubspot: The most popular CRM for good reason and has a broader product range you can adopt in your next steps. Try this if you have a bulky list of customers because it is free. #10 Pardot: Pardot is by Salesforce, it's armed with features that can close the gap between marketing and sales. Sales Tools #11 Salesforce is the best sales automation and lead management software. It helps you to create complicated segmentations and run, track, analyze campaigns from the same dashboard. #12 LinkedIn Sales Navigator gives you full access to LinkedIn's user database. You can even find a kidnapped CEO if you know how to use it with other marketing automation tools like Skrapp. #13 Pipedrive is a simple tool and excels in one thing. It tracks your leads and tells you when to take the next action. It makes sales easier. #14 Qwilr creates great-looking docs, at speed. You can design perfect proposals, quotes, client updates, and more in a flash. We use it a lot to close deals, it's effective. #15 Crystalknows is an add-on that tells you anyone’s personality on LinkedIn and gives you a detailed approach specific to that person. It's eerily accurate. #16 Leadfeeder shows you the companies that visited your website. Tells how they found you and what they’re interested in. It has a free version. Communication Tools #17 Intercom is a sweet and smart host that welcomes your visitors when you’re not home. It’s one of the best chatbot tools in the market. #18 Drift is famous for its conversational marketing features and more sales-focused than Intercom. #19 Manychat is a chatbot that helps you create high converting Facebook campaigns. #20 Plann3r helps you create your personalized meeting page. You can schedule meetings witch clients, candidates, and prospects. #21 Loom is a video messaging tool, it helps you to be more expressive and create closer relationships. #22 Callpage collects your visitors’ phone number and connects you with them in seconds. No matter where you are. Landing page tools #23 Instapage is the best overall landing page builder. It has a broad range of features and even squirrel can build a compelling landing page with templates. No coding needed. #24 Unbounce can do everything that Instapage does and lets you build a great landing page without a developer. But it's less intuitive. Lead generation / marketing automation tools #25 Phantombuster is by far the most used lead generation software in our tool kit. It extracts data, emails, sends requests, customized messages, and does many things on autopilot in any platform. You can check this, this and this if you want to see it in action. #26 Duxsoup is a Google Chrome add-on and can also automate some of LinkedIn lead generation efforts like Phantombuster. But not works in the cloud. #27 Zapier is a glue that holds all the lead generation tools together. With Zapier, You can connect different marketing tools and no coding required. Conversion rate optimization tools #28 Hotjar tracks what people are doing on your website by recording sessions and capturing mouse movements. Then it gives you a heatmap. #29 UsabilityHub shows your page to a digital crowd and measures the first impressions and helps you to validate your ideas. #30 Optinmonster is a top tier conversion optimization tool. It helps you to capture leads and enables you to increase conversions rates with many features. #31 Notifia is one mega tool of widgets that arms your website with the wildest social proof and lead capturing tactics. #32 Sumo is a much simpler version of Notifia. But Sumo has everything to help you capture leads and build your email lists. Web scrapers #33 Data Miner is a Google Chrome browser extension that helps you scrape data from web pages and into a CSV file or Excel spreadsheet. #34 Webscraper does the same thing as Data Miner; however, it is capable of handling more complex tasks. SEO and Content #35 Grammarly: Your English could be your first language and your grammar could be better than Shakespeare. Grammarly still can make your writing better. #36 Hemingwayapp is a copywriting optimization tool that gives you feedback about your copy and improves your readability score, makes your writing bolder and punchier. Free. #37 Ahrefs is an all-rounder search engine optimization tool that helps you with off-page, on-page or technical SEO. #38 SurferSEO makes things easier for your on-page SEO efforts. It’s a tool that analyzes top Google results for specific keywords and gives you a content brief based on that data. Video editing and design tools #39 Canva is a graphic design platform that makes everything easy. It has thousands of templates for anything from Facebook ads, stylish presentations to business cards.  #40 Kapwing is our go-to platform for quick video edits. It works on the browser and can help you to create stylish videos, add subtitles, resize videos, create memes, or remove backgrounds. #41 Animoto can turn your photos and video clip into beautiful video slideshows. It comes handy when you want to create an advertising material but don’t have a budget. Advertising tools #42 AdEspresso lets you create and test multiple ads with few clicks. You can optimize your FB, IG, and Google ads from this tool and measure your ads with in-depth analytics. #43 AdRoll is an AI-driven platform that connects and coordinates marketing efforts across ads, email, and online stores. Other tools #44 Replug helps you to shorten, track, optimize your links with call-to-actions, branded links, and retargeting pixels #45 Draw.io = Mindmaps, schemes, and charts. With Draw.io, you can put your brain in a digital paper in an organized way. #46 Built With is a tool that finds out what websites are built with. So you can see what tools they're using and so on. #47 Typeform can turn data collection into an experience with Typeform. This tool helps you to engage your audience with conversational forms or surveys and help you to collect more data. #48 Livestorm helped us a lot, especially in COVID-19 tiles. It’s a webinar software that works on your browser, mobile, and desktop. #49 Teachable \- If you have an online course idea but hesitating because of the production process, Teachable can help you. It's easy to configure and customizable for your needs. #50 Viral Loops provides a revolutionary referral marketing solution for modern marketers. You can create and run referral campaigns in a few clicks with templates. Remember, most of these tools have a free trial or free version. Going over them one by one can teach you a lot and help you grow your business with less work power in the early stages of your business. I hope you enjoyed the read and can find some tools to make things easier! Let me know about your favorite tools in the comments, so I can try them out. \------ If you want to check the prices and see a broader explanation about the tools, you can go here.

I run an AI automation agency (AAA). My honest overview and review of this new business model
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

I run an AI automation agency (AAA). My honest overview and review of this new business model

I started an AI tools directory in February, and then branched off that to start an AI automation agency (AAA) in June. So far I've come across a lot of unsustainable "ideas" to make money with AI, but at the same time a few diamonds in the rough that aren't fully tapped into yet- especially the AAA model. Thought I'd share this post to shine light into this new business model and share some ways you could potentially start your own agency, or at the very least know who you are dealing with and how to pick and choose when you (inevitably) get bombarded with cold emails from them down the line. Foreword Running an AAA does NOT involve using AI tools directly to generate and sell content directly. That ship has sailed, and unless you are happy with $5 from Fiverr every month or so, it is not a real business model. Cry me a river but generating generic art with AI and slapping it onto a T-shirt to sell on Etsy won't make you a dime. At the same time, the AAA model will NOT require you to have a deep theoretical knowledge of AI, or any academic degree, as we are more so dealing with the practical applications of generative AI and how we can implement these into different workflows and tech-stacks, rather than building AI models from the ground up. Regardless of all that, common sense and a willingness to learn will help (a shit ton), as with anything. Keep in mind - this WILL involve work and motivation as well. The mindset that AI somehow means everything can be done for you on autopilot is not the right way to approach things. The common theme of businesses I've seen who have successfully implemented AI into their operations is the willingess to work with AI in a way that augments their existing operations, rather than flat out replace a worker or team. And this is exactly the train of thought you need when working with AI as a business model. However, as the field is relatively unsaturated and hype surrounding AI is still fresh for enterprises, right now is the prime time to start something new if generative AI interests you at all. With that being said, I'll be going over three of the most successful AI-adjacent businesses I've seen over this past year, in addition to some tips and resources to point you in the right direction. so.. WTF is an AI Automation Agency? The AI automation agency (or as some YouTubers have coined it, the AAA model) at its core involves creating custom AI solutions for businesses. I have over 1500 AI tools listed in my directory, however the feedback I've received from some enterprise users is that ready-made SaaS tools are too generic to meet their specific needs. Combine this with the fact virtually no smaller companies have the time or skills required to develop custom solutions right off the bat, and you have yourself real demand. I would say in practice, the AAA model is quite similar to Wordpress and even web dev agencies, with the major difference being all solutions you develop will incorporate key aspects of AI AND automation. Which brings me to my second point- JUST AI IS NOT ENOUGH. Rather than reducing the amount of time required to complete certain tasks, I've seen many AI agencies make the mistake of recommending and (trying to) sell solutions that more likely than not increase the workload of their clients. For example, if you were to make an internal tool that has AI answer questions based on their knowledge base, but this knowledge base has to be updated manually, this is creating unnecessary work. As such I think one of the key components of building successful AI solutions is incorporating the new (Generative AI/LLMs) with the old (programmtic automation- think Zapier, APIs, etc.). Finally, for this business model to be successful, ideally you should target a niche in which you have already worked and understand pain points and needs. Not only does this make it much easier to get calls booked with prospects, the solutions you build will have much greater value to your clients (meaning you get paid more). A mistake I've seen many AAA operators make (and I blame this on the "Get Rich Quick" YouTubers) is focusing too much on a specific productized service, rather than really understanding the needs of businesses. The former is much done via a SaaS model, but when going the agency route the only thing that makes sense is building custom solutions. This is why I always take a consultant-first approach. You can only build once you understand what they actually need and how certain solutions may impact their operations, workflows, and bottom-line. Basics of How to Get Started Pick a niche. As I mentioned previously, preferably one that you've worked in before. Niches I know of that are actively being bombarded with cold emails include real estate, e-commerce, auto-dealerships, lawyers, and medical offices. There is a reason for this, but I will tell you straight up this business model works well if you target any white-collar service business (internal tools approach) or high volume businesses (customer facing tools approach). Setup your toolbox. If you wanted to start a pressure washing business, you would need a pressure-washer. This is no different. For those without programming knowledge, I've seen two common ways AAA get setup to build- one is having a network of on-call web developers, whether its personal contacts or simply going to Upwork or any talent sourcing agency. The second is having an arsenal of no-code tools. I'll get to this more in a second, but this works beecause at its core, when we are dealing with the practical applications of AI, the code is quite simple, simply put. Start cold sales. Unless you have a network already, this is not a step you can skip. You've already picked a niche, so all you have to do is find the right message. Keep cold emails short, sweet, but enticing- and it will help a lot if you did step 1 correctly and intimately understand who your audience is. I'll be touching base later about how you can leverage AI yourself to help you with outreach and closing. The beauty of gen AI and the AAA model You don't need to be a seasoned web developer to make this business model work. The large majority of solutions that SME clients want is best done using an API for an LLM for the actual AI aspect. The value we create with the solutions we build comes with the conceptual framework and design that not only does what they need it to but integrates smoothly with their existing tech-stack and workflow. The actual implementation is quite straightforward once you understand the high level design and know which tools you are going to use. To give you a sense, even if you plan to build out these apps yourself (say in Python) the large majority of the nitty gritty technical work has already been done for you, especially if you leverage Python libraries and packages that offer high level abstraction for LLM-related functions. For instance, calling GPT can be as little as a single line of code. (And there are no-code tools where these functions are simply an icon on a GUI). Aside from understanding the capabilities and limitations of these tools and frameworks, the only thing that matters is being able to put them in a way that makes sense for what you want to build. Which is why outsourcing and no-code tools both work in our case. Okay... but how TF am I suppposed to actually build out these solutions? Now the fun part. I highly recommend getting familiar with Langchain and LlamaIndex. Both are Python libraires that help a lot with the high-level LLM abstraction I mentioned previously. The two most important aspects include being able to integrate internal data sources/knowledge bases with LLMs, and have LLMs perform autonomous actions. The two most common methods respectively are RAG and output parsing. RAG (retrieval augmented Generation) If you've ever seen a tool that seemingly "trains" GPT on your own data, and wonder how it all works- well I have an answer from you. At a high level, the user query is first being fed to what's called a vector database to run vector search. Vector search basically lets you do semantic search where you are searching data based on meaning. The vector databases then retrieves the most relevant sections of text as it relates to the user query, and this text gets APPENDED to your GPT prompt to provide extra context to the AI. Further, with prompt engineering, you can limit GPT to only generate an answer if it can be found within this extra context, greatly limiting the chance of hallucination (this is where AI makes random shit up). Aside from vector databases, we can also implement RAG with other data sources and retrieval methods, for example SQL databses (via parsing the outputs of LLM's- more on this later). Autonomous Agents via Output Parsing A common need of clients has been having AI actually perform tasks, rather than simply spitting out text. For example, with autonomous agents, we can have an e-commerce chatbot do the work of a basic customer service rep (i.e. look into orders, refunds, shipping). At a high level, what's going on is that the response of the LLM is being used programmtically to determine which API to call. Keeping on with the e-commerce example, if I wanted a chatbot to check shipping status, I could have a LLM response within my app (not shown to the user) with a prompt that outputs a random hash or string, and programmatically I can determine which API call to make based on this hash/string. And using the same fundamental concept as with RAG, I can append the the API response to a final prompt that would spit out the answer for the user. How No Code Tools Can Fit In (With some example solutions you can build) With that being said, you don't necessarily need to do all of the above by coding yourself, with Python libraries or otherwise. However, I will say that having that high level overview will help IMMENSELY when it comes to using no-code tools to do the actual work for you. Regardless, here are a few common solutions you might build for clients as well as some no-code tools you can use to build them out. Ex. Solution 1: AI Chatbots for SMEs (Small and Medium Enterprises) This involves creating chatbots that handle user queries, lead gen, and so forth with AI, and will use the principles of RAG at heart. After getting the required data from your client (i.e. product catalogues, previous support tickets, FAQ, internal documentation), you upload this into your knowledge base and write a prompt that makes sense for your use case. One no-code tool that does this well is MyAskAI. The beauty of it especially for building external chatbots is the ability to quickly ingest entire websites into your knowledge base via a sitemap, and bulk uploading files. Essentially, they've covered the entire grunt work required to do this manually. Finally, you can create a inline or chat widget on your client's website with a few lines of HTML, or altneratively integrate it with a Slack/Teams chatbot (if you are going for an internal Q&A chatbot approach). Other tools you could use include Botpress and Voiceflow, however these are less for RAG and more for building out complete chatbot flows that may or may not incorporate LLMs. Both apps are essentially GUIs that eliminate the pain and tears and trying to implement complex flows manually, and both natively incoporate AI intents and a knowledge base feature. Ex. Solution 2: Internal Apps Similar to the first example, except we go beyond making just chatbots but tools such as report generation and really any sort of internal tool or automations that may incorporate LLM's. For instance, you can have a tool that automatically generates replies to inbound emails based on your client's knowledge base. Or an automation that does the same thing but for replies to Instagram comments. Another example could be a tool that generates a description and screeenshot based on a URL (useful for directory sites, made one for my own :P). Getting into more advanced implementations of LLMs, we can have tools that can generate entire drafts of reports (think 80+ pages), based not only on data from a knowledge base but also the writing style, format, and author voice of previous reports. One good tool to create content generation panels for your clients would be MindStudio. You can train LLM's via prompt engineering in a structured way with your own data to essentially fine tune them for whatever text you need it to generate. Furthermore, it has a GUI where you can dictate the entire AI flow. You can also upload data sources via multiple formats, including PDF, CSV, and Docx. For automations that require interactions between multiple apps, I recommend the OG zapier/make.com if you want a no-code solution. For instance, for the automatic email reply generator, I can have a trigger such that when an email is received, a custom AI reply is generated by MyAskAI, and finally a draft is created in my email client. Or, for an automation where I can create a social media posts on multiple platforms based on a RSS feed (news feed), I can implement this directly in Zapier with their native GPT action (see screenshot) As for more complex LLM flows that may require multiple layers of LLMs, data sources, and APIs working together to generate a single response i.e. a long form 100 page report, I would recommend tools such as Stack AI or Flowise (open-source alternative) to build these solutions out. Essentially, you get most of the functions and features of Python packages such as Langchain and LlamaIndex in a GUI. See screenshot for an example of a flow How the hell are you supposed to find clients? With all that being said, none of this matters if you can't find anyone to sell to. You will have to do cold sales, one way or the other, especially if you are brand new to the game. And what better way to sell your AI services than with AI itself? If we want to integrate AI into the cold outreach process, first we must identify what it's good at doing, and that's obviously writing a bunch of text, in a short amount of time. Similar to the solutions that an AAA can build for its clients, we can take advantage of the same principles in our own sales processes. How to do outreach Once you've identified your niche and their pain points/opportunities for automation, you want to craft a compelling message in which you can send via cold email and cold calls to get prospects booked on demos/consultations. I won't get into too much detail in terms of exactly how to write emails or calling scripts, as there are millions of resources to help with this, but I will tell you a few key points you want to keep in mind when doing outreach for your AAA. First, you want to keep in mind that many businesses are still hesitant about AI and may not understand what it really is or how it can benefit their operations. However, we can take advantage of how mass media has been reporting on AI this past year- at the very least people are AWARE that sooner or later they may have to implement AI into their businesses to stay competitive. We want to frame our message in a way that introduces generative AI as a technology that can have a direct, tangible, and positive impact on their business. Although it may be hard to quantify, I like to include estimates of man-hours saved or costs saved at least in my final proposals to prospects. Times are TOUGH right now, and money is expensive, so you need to have a compelling reason for businesses to get on board. Once you've gotten your messaging down, you will want to create a list of prospects to contact. Tools you can use to find prospects include Apollo.io, reply.io, zoominfo (expensive af), and Linkedin Sales Navigator. What specific job titles, etc. to target will depend on your niche but for smaller companies this will tend to be the owner. For white collar niches, i.e. law, the professional that will be directly benefiting from the tool (i.e. partners) may be better to contact. And for larger organizations you may want to target business improvement and digital transformation leads/directors- these are the people directly in charge of projects like what you may be proposing. Okay- so you have your message, and your list, and now all it comes down to is getting the good word out. I won't be going into the details of how to send these out, a quick Google search will give you hundreds of resources for cold outreach methods. However, personalization is key and beyond simple dynamic variables you want to make sure you can either personalize your email campaigns directly with AI (SmartWriter.ai is an example of a tool that can do this), or at the very least have the ability to import email messages programmatically. Alternatively, ask ChatGPT to make you a Python Script that can take in a list of emails, scrape info based on their linkedin URL or website, and all pass this onto a GPT prompt that specifies your messaging to generate an email. From there, send away. How tf do I close? Once you've got some prospects booked in on your meetings, you will need to close deals with them to turn them into clients. Call #1: Consultation Tying back to when I mentioned you want to take a consultant-first appraoch, you will want to listen closely to their goals and needs and understand their pain points. This would be the first call, and typically I would provide a high level overview of different solutions we could build to tacke these. It really helps to have a presentation available, so you can graphically demonstrate key points and key technologies. I like to use Plus AI for this, it's basically a Google Slides add-on that can generate slide decks for you. I copy and paste my default company messaging, add some key points for the presentation, and it comes out with pretty decent slides. Call #2: Demo The second call would involve a demo of one of these solutions, and typically I'll quickly prototype it with boilerplate code I already have, otherwise I'll cook something up in a no-code tool. If you have a niche where one type of solution is commonly demanded, it helps to have a general demo set up to be able to handle a larger volume of calls, so you aren't burning yourself out. I'll also elaborate on how the final product would look like in comparison to the demo. Call #3 and Beyond: Once the initial consultation and demo is complete, you will want to alleviate any remaining concerns from your prospects and work with them to reach a final work proposal. It's crucial you lay out exactly what you will be building (in writing) and ensure the prospect understands this. Furthermore, be clear and transparent with timelines and communication methods for the project. In terms of pricing, you want to take this from a value-based approach. The same solution may be worth a lot more to client A than client B. Furthermore, you can create "add-ons" such as monthly maintenance/upgrade packages, training sessions for employeees, and so forth, separate from the initial setup fee you would charge. How you can incorporate AI into marketing your businesses Beyond cold sales, I highly recommend creating a funnel to capture warm leads. For instance, I do this currently with my AI tools directory, which links directly to my AI agency and has consistent branding throughout. Warm leads are much more likely to close (and honestly, much nicer to deal with). However, even without an AI-related website, at the very least you will want to create a presence on social media and the web in general. As with any agency, you will want basic a professional presence. A professional virtual address helps, in addition to a Google Business Profile (GBP) and TrustPilot. a GBP (especially for local SEO) and Trustpilot page also helps improve the looks of your search results immensely. For GBP, I recommend using ProfilePro, which is a chrome extension you can use to automate SEO work for your GBP. Aside from SEO optimzied business descriptions based on your business, it can handle Q/A answers, responses, updates, and service descriptions based on local keywords. Privacy and Legal Concerns of the AAA Model Aside from typical concerns for agencies relating to service contracts, there are a few issues (especially when using no-code tools) that will need to be addressed to run a successful AAA. Most of these surround privacy concerns when working with proprietary data. In your terms with your client, you will want to clearly define hosting providers and any third party tools you will be using to build their solution, and a DPA with these third parties listed as subprocessors if necessary. In addition, you will want to implement best practices like redacting private information from data being used for building solutions. In terms of addressing concerns directly from clients, it helps if you host your solutions on their own servers (not possible with AI tools), and address the fact only ChatGPT queries in the web app, not OpenAI API calls, will be used to train OpenAI's models (as reported by mainstream media). The key here is to be open and transparent with your clients about ALL the tools you are using, where there data will be going, and make sure to get this all in writing. have fun, and keep an open mind Before I finish this post, I just want to reiterate the fact that this is NOT an easy way to make money. Running an AI agency will require hours and hours of dedication and work, and constantly rearranging your schedule to meet prospect and client needs. However, if you are looking for a new business to run, and have a knack for understanding business operations and are genuinely interested in the pracitcal applications of generative AI, then I say go for it. The time is ticking before AAA becomes the new dropshipping or SMMA, and I've a firm believer that those who set foot first and establish themselves in this field will come out top. And remember, while 100 thousand people may read this post, only 2 may actually take initiative and start.

Raised $450k for my startup, here are the lessons I've learned along the way
reddit
LLM Vibe Score0
Human Vibe Score1
marin_smiljanicThis week

Raised $450k for my startup, here are the lessons I've learned along the way

2021 has been a pretty amazing year for Omnisearch. Having started initial work on Omnisearch at the end of 2020, we entered the new year with a working MVP yet no revenue, no significant partnerships, and no funding. Fast forward to the end of 2021, and we now have fantastic revenue growth, a partnership with a public company, and a far more powerful, complete and polished product. But one milestone really changed Omnisearch’s trajectory: our $450,000 USD pre-seed round by GoAhead Ventures. In this post I want to share the story of how it came about and offer a couple of takeaways to keep in mind when preparing for fundraising. ​ The story Contrary to most advice, my co-founder Matej and I didn’t allocate a specific time to switch to “fundraising mode” but rather talked to investors on an ongoing basis. It was a bit of a distraction from working on the product, but on the positive side we were able to constantly get feedback on the idea, pitch, go-to-market strategy and hiring, as well as hearing investors’ major concerns sooner rather than later. That being said, our six-month long fundraising efforts weren’t yielding results - we talked to about twenty investors, mostly angels or smaller funds, with no success. The feedback was generally of the “too early for us” variety (since we were still pre-revenue), with additional questions about our go-to-market strategy and ideal customer persona. The introduction to our eventual investors, California-based GoAhead Ventures, came through a friend who had pitched them previously. We wrote a simple blurb and sent our pitch deck. We then went through GoAhead’s hyper-efficient screening process, consisting of a 30-minute call, a recorded three-minute pitch, and filling out a simple Google doc. Throughout the whole process, the GoAhead team left an awesome impression thanks to their knowledge of enterprise software and their responsiveness. They ended up investing and the whole deal was closed within two weeks, which is super fast even by Silicon Valley standards. While our fundraising experience is a single data point and your case might be different, here are the key takeaways from our journey. ​ Perseverance wins: Like I said above, we talked to about twenty investors before we closed our round. Getting a series of “no”s sucks, but we took the feedback seriously and tried to prepare better for questions that caught us off guard. But we persevered, keeping in mind that from a bird’s eye perspective it’s an amazing time to be building startups and raising funds. Focus on traction: Sounds pretty obvious, right? The truth is, though, that even a small amount of revenue is infinitely better than none at all. One of the major differences between our eventual successful investor pitch and the earlier ones was that we had actual paying customers, though our MRR was low. This allows you to talk about customers in the present tense, showing there’s actual demand for your product and making the use cases more tangible. And ideally, highlight a couple of customer testimonials to boost your credibility. Have a demo ready: In Omnisearch’s case, the demo was oftentimes the best received part of the pitch or call. We’d show investors the live demo, and for bonus points even asked them to choose a video from YouTube and then try searching through it. This always had a “wow” effect on prospective investors and made the subsequent conversation more exciting and positive. Accelerators: Accelerators like Y Combinator or Techstars can add enormous value to a startup, especially in the early stages. And while it’s a great idea to apply, don’t rely on them too heavily. Applications happen only a few times a year, and you should have a foolproof fundraising plan in case you don’t get in. In our case, we just constantly looked for investors who were interested in our space (defined as enterprise SaaS more broadly), using LinkedIn, AngelList, and intros from our own network. Practice the pitch ad nauseam: Pitching is tough to get right even for seasoned pros, so it pays to practice as often as possible. We took every opportunity to perfect the pitch: attending meetups and giving the thirty-second elevator pitch to other attendees over beer and pizza, participating in startup competitions, going to conferences and exhibiting at our own booth, attending pre-accelerator programs, and pitching to friends who are in the startup world. Show an understanding of the competition: Frankly, this was one of the strongest parts of our pitch and investor conversations. If you’re in a similar space to ours, Gartner Magic Quadrants and Forrester Waves are an awesome resource, as well as sites like AlternativeTo or Capterra and G2. By thoroughly studying these resources we gained a great understanding of the industry landscape and were able to articulate our differentiation more clearly and succinctly. Presenting this visually in a coordinate system or a feature grid is, from our experience, even more effective. Remember it’s just the beginning! Getting your first round of funding is just the beginning of the journey, so it’s important to avoid euphoria and get back to building and selling the product as soon as possible. While securing funding enables you to scale the team, and is a particular relief if the founders had worked without a salary, the end goal is still to build a big, profitable, and overall awesome startup.

The delicate balance of building an online community business
reddit
LLM Vibe Score0
Human Vibe Score0.895
matthewbarbyThis week

The delicate balance of building an online community business

Hey /r/Entrepreneur 👋 Just under two years ago I launched an online community business called Traffic Think Tank with two other co-founders, Nick Eubanks and Ian Howells. As a Traffic Think Tank customer you (currently) pay $119 a month to get access to our online community, which is run through Slack. The community is focused on helping you learn various aspects of marketing, with a particular focus on search engine optimization (SEO). Alongside access to the Slack community, we publish new educational video content from outside experts every week that all customers have access to. At the time of writing, Traffic Think Tank has around 650 members spanning across 17 of the 24 different global time zones. I was on a business trip over in Sydney recently, and during my time there I met up with some of our Australia-based community members. During dinner I was asked by several of them how the idea for Traffic Think Tank came about and what steps we took to validate that the idea was worth pursuing.  This is what I told them… How it all began It all started with a personal need. Nick, an already successful entrepreneur and owner of a marketing agency, had tested out an early version Traffic Think Tank in early 2017. He offered real-time consulting for around ten customers that he ran from Slack. He would publish some educational videos and offer his advice on projects that the members were running. The initial test went well, but it was tough to maintain on his own and he had to charge a fairly high price to make it worth his time. That’s when he spoke to me and Ian about turning this idea into something much bigger. Both Ian and I offered something slightly different to Nick. We’ve both spent time in senior positions at marketing agencies, but currently hold senior director positions in 2,000+ public employee companies (HubSpot and LendingTree). Alongside this, as a trio we could really ramp up the quality and quantity of content within the community, spread out the administrative workload and just generally have more resources to throw at getting this thing off the ground. Admittedly, Nick was much more optimistic about the potential of Traffic Think Tank – something I’m very thankful for now – whereas Ian and I were in the camp of “you’re out of your mind if you think hundreds of people are going to pay us to be a part of a Slack channel”. To validate the idea at scale, we decided that we’d get an initial MVP of the community up and running with a goal of reaching 100 paying customers in the first six months. If we achieved that, we’d validated that it was a viable business and we would continue to pursue it. If not, we’d kill it. We spent the next month building out the initial tech stack that enabled us to accept payments, do basic user management to the Slack channel, and get a one-page website up and running with information on what Traffic Think Tank was all about.  After this was ready, we doubled down on getting some initial content created for members – I mean, we couldn’t have people just land in an empty Slack channel, could we? We created around ten initial videos, 20 or so articles and then some long threads full of useful information within the Slack channel so that members would have some content to pour into right from the beginning.  Then, it was time to go live. The first 100 customers Fortunately, both Nick and I had built a somewhat substantial following in the SEO space over the previous 5-10 years, so we at least had a large email list to tap into (a total of around 40,000 people). We queued up some launch emails, set an initial price of $99 per month and pressed send. [\[LINK\] The launch email I sent to my subscribers announcing Traffic Think Tank](https://mailchi.mp/matthewbarby/future-of-marketing-1128181) What we didn’t expect was to sell all of the initial 100 membership spots in the first 72 hours. “Shit. What do we do now? Are we ready for this many people? Are we providing them with enough value? What if something breaks in our tech stack? What if they don’t like the content? What if everyone hates Slack?” All of these were thoughts running through my head. This brings me to the first great decision we made: we closed down new membership intake for 3 months so that we could focus completely on adding value to the first cohort of users. The right thing at the right time SEO is somewhat of a dark art to many people that are trying to learn about it for the first time. There’s hundreds of thousands (possibly millions) of articles and videos online that talk about how to do SEO.  Some of it’s good advice; a lot of it is very bad advice.  Add to this that the barrier to entry of claiming to be an “expert” in SEO is practically non-existent and you have a recipe for disaster. This is why, for a long time, individuals involved in SEO have flocked in their masses to online communities for information and to bounce ideas off of others in the space. Forums like SEObook, Black Hat World, WickedFire, Inbound.org, /r/BigSEO, and many more have, at one time, been called home by many SEOs.  In recent times, these communities have either been closed down or just simply haven’t adapted to the changing needs of the community – one of those needs being real-time feedback on real-world problems.  The other big need that we all spotted and personally had was the ability to openly share the things that are working – and the things that aren’t – in SEO within a private forum. Not everyone wanted to share their secret sauce with the world. One of the main reasons we chose Slack as the platform to run our community on was the fact that it solved these two core needs. It gave the ability to communicate in real-time across multiple devices, and all of the information shared within it was outside of the public domain. The other problem that plagued a lot of these early communities was spam. Most of them were web-based forums that were free to access. That meant they became a breeding ground for people trying to either sell their services or promote their own content – neither of which is conducive to building a thriving community. This was our main motivation for charging a monthly fee to access Traffic Think Tank. We spent a lot of time thinking through pricing. It needed to be enough money that people would be motivated to really make use of their membership and act in a way that’s beneficial to the community, but not too much money that it became cost prohibitive to the people that would benefit from it the most. Considering that most of our members would typically spend between $200-800 per month on SEO software, $99 initially felt like the perfect balance. Growing pains The first three months of running the community went by without any major hiccups. Members were incredibly patient with us, gave us great feedback and were incredibly helpful and accommodating to other members. Messages were being posted every day, with Nick, Ian and myself seeding most of the engagement at this stage.  With everything going smoothly, we decided that it was time to open the doors to another intake of new members. At this point we’d accumulated a backlog of people on our waiting list, so we knew that simply opening our doors would result in another large intake. Adding more members to a community has a direct impact on the value that each member receives. For Traffic Think Tank in particular, the value for members comes from three areas: The ability to have your questions answered by me, Nick and Ian, as well as other members of the community. The access to a large library of exclusive content. The ability to build connections with the wider community. In the early stages of membership growth, there was a big emphasis on the first of those three points. We didn’t have an enormous content library, nor did we have a particularly large community of members, so a lot of the value came from getting a lot of one-to-one time with the community founders. [\[IMAGE\] Screenshot of engagement within the Traffic Think Tank Slack community](https://cdn.shortpixel.ai/client/qglossy,retimg,w_1322/https://www.matthewbarby.com/wp-content/uploads/2019/08/Community-Engagement-in-Traffic-Think-Tank.png) The good thing about having 100 members was that it was just about feasible to give each and every member some one-to-one time within the month, which really helped us to deliver those moments of delight that the community needed early on. Two-and-a-half months after we launched Traffic Think Tank, we opened the doors to another 250 people, taking our total number of members to 350. This is where we experienced our first growing pains.  Our original members had become used to being able to drop us direct messages and expect an almost instant response, but this wasn’t feasible anymore. There were too many people, and we needed to create a shift in behavior. We needed more value to come from the community engaging with one another or we’d never be able to scale beyond this level. We started to really pay attention to engagement metrics; how many people were logging in every day, and of those, how many were actually posting messages within public channels.  We asked members that were logging in a lot but weren’t posting (the “lurkers”) why that was the case. We also asked the members that engaged in the community the most what motivated them to post regularly. We learned a lot from doing this. We found that the large majority of highly-engaged members had much more experience in SEO, whereas most of the “lurkers” were beginners. This meant that most of the information being shared in the community was very advanced, with a lot of feedback from the beginners in the group being that they “didn’t want to ask a stupid question”.  As managers of the community, we needed to facilitate conversations that catered to all of our members, not just those at a certain level of skill. To tackle this problem, we created a number of new channels that had a much deeper focus on beginner topics so novice members had a safe place to ask questions without judgment.  We also started running live video Q&As each month where we’d answer questions submitted by the community. This gave our members one-on-one time with me, Nick and Ian, but spread the value of these conversations across the whole community rather than them being hidden within private messages. As a result of these changes, we found that the more experienced members in the community were really enjoying sharing their knowledge with those with less experience. The number of replies within each question thread was really starting to increase, and the community started to shift away from just being a bunch of threads created by me, Nick and Ian to a thriving forum of diverse topics compiled by a diverse set of individuals. This is what we’d always wanted. A true community. It was starting to happen. [\[IMAGE\] Chart showing community engagement vs individual member value](https://cdn.shortpixel.ai/client/qglossy,retimg,w_1602/https://www.matthewbarby.com/wp-content/uploads/2019/08/Community-Engagement-Balance-Graph.jpg) At the same time, we started to realize that we’ll eventually reach a tipping point where there’ll be too much content for us to manage and our members to engage with. When we reach this point, the community will be tough to follow and the quality of any given post will go down. Not only that, but the community will become increasingly difficult to moderate. We’re not there yet, but we recognize that this will come, and we’ll have to adjust our model again. Advocating advocacy As we started to feel more comfortable about the value that members were receiving, we made the decision to indefinitely open for new members. At the same time, we increased the price of membership (from $99 a month to $119) in a bid to strike the right balance between profitability as a business and to slow down the rate at which we were reaching the tipping point of community size. We also made the decision to repay all of our early adopters by grandfathering them in to the original pricing – and committing to always do this in the future. Despite the price increase, we saw a continued flow of new members come into the community. The craziest part about this was that we were doing practically no marketing activities to encourage new members– this was all coming from word of mouth. Our members were getting enough value from the community that they were recommending it to their friends, colleagues and business partners.  The scale at which this was happening really took us by surprise and it told us one thing very clearly: delivering more value to members resulted in more value being delivered to the business. This is a wonderful dynamic to have because it perfectly aligns the incentives on both sides. We’d said from the start that we wouldn’t sacrifice value to members for more revenue – this is something that all three of us felt very strongly about. First and foremost, we wanted to create a community that delivered value to its members and was run in a way that aligned with our values as people. If we could find a way to stimulate brand advocacy, while also tightening the bonds between all of our individual community members, we’d be boosting both customer retention and customer acquisition in the same motion. This became our next big focus. [\[TWEET\] Adam, one of our members wore his Traffic Think Tank t-shirt in the Sahara desert](https://twitter.com/AdamGSteele/status/1130892481099382784) We started with some simple things: We shipped out Traffic Think Tank branded T-shirts to all new members. We’d call out each of the individuals that would submit questions to our live Q&A sessions and thank them live on air. We set up a new channel that was dedicated to sharing a quick introduction to who you are, what you do and where you’re based for all new members. We’d created a jobs channel and a marketplace for selling, buying and trading services with other members. Our monthly “blind dates” calls were started where you’d be randomly grouped with 3-4 other community members so that you could hop on a call to get to know each other better. The Traffic Think Tank In Real Life (IRL)* channel was born, which enabled members to facilitate in-person meetups with each other. In particular, we saw that as members started to meet in person or via calls the community itself was feeling more and more like a family. It became much closer knit and some members started to build up a really positive reputation for being particularly helpful to other members, or for having really strong knowledge in a specific area. [\[TWEET\] Dinner with some of the Traffic Think Tank members in Brighton, UK](https://twitter.com/matthewbarby/status/1117175584080134149) Nick, Ian and I would go out of our way to try and meet with members in real life wherever we could. I was taken aback by how appreciative people were for us doing this, and it also served as an invaluable way to gain honest feedback from members. There was another trend that we’d observed that we didn’t really expect to happen. More and more members were doing business with each another. We’ve had people find new jobs through the community, sell businesses to other members, launch joint ventures together and bring members in as consultants to their business. This has probably been the most rewarding thing to watch, and it was clear that the deeper relationships that our members were forming were resulting in an increased level of trust to work with each other. We wanted to harness this and take it to a new level. This brought us to arguably the best decision we’ve made so far running Traffic Think Tank… we were going to run a big live event for our members. I have no idea what I’m doing It’s the first week of January 2019 and we’re less than three weeks away from Traffic Think Tank LIVE, our first ever in-person event hosting 150 people, most of which are Traffic Think Tank members. It's like an ongoing nightmare I can’t wake up from. That was Nick’s response in our private admin channel to myself and Ian when I asked if they were finding the run-up to the event as stressful as I was. I think that all three of us were riding on such a high from how the community was growing that we felt like we could do anything. Running an event? How hard can it be? Well, turns out it’s really hard. We had seven different speakers flying over from around the world to speak at the event, there was a pre- and after event party, and we’d planned a charity dinner where we would take ten attendees (picked at random via a raffle) out for a fancy meal. Oh, and Nick, Ian and I were hosting a live Q&A session on stage. It wasn’t until precisely 48 hours before the event that we’d realized we didn’t have any microphones, nor had a large amount of the swag we’d ordered arrived. Plus, a giant storm had hit Philly causing a TON of flight cancellations. Perfect. Just perfect. This was honestly the tip of the iceberg. We hadn’t thought about who was going to run the registration desk, who would be taking photos during the event and who would actually field questions from the audience while all three of us sat on stage for our live Q&A panel. Turns out that the answer to all of those questions were my wife, Laura, and Nick’s wife, Kelley. Thankfully, they were on hand to save our asses. The weeks running up to the event were honestly some of the most stressful of my life. We sold around 50% of our ticket allocation within the final two weeks before the event. All of the event organizers told us this would happen, but did we believe them? Hell no!  Imagine having two weeks until the big day and as it stood half of the room would be completely empty. I was ready to fly most of my extended family over just to make it look remotely busy. [\[IMAGE\] One of our speakers, Ryan Stewart, presenting at Traffic Think Tank LIVE](https://cdn.shortpixel.ai/client/qglossy,retimg,w_1920/https://www.matthewbarby.com/wp-content/uploads/2019/08/Traffic-Think-Tank-LIVE-Ryan-Presenting.jpg) Thankfully, if all came together. We managed to acquire some microphones, the swag arrived on the morning of the event, all of our speakers were able to make it on time and the weather just about held up so that our entire allocation of ticket holders was able to make it to the event. We pooled together and I’m proud to say that the event was a huge success. While we made a substantial financial loss on the event itself, January saw a huge spike in new members, which more than recouped our losses. Not only that, but we got to hang out with a load of our members all day while they said really nice things about the thing we’d built. It was both exhausting and incredibly rewarding. Bring on Traffic Think Tank LIVE 2020! (This time we’re hiring an event manager...)   The road ahead Fast forward to today (August 2019) and Traffic Think Tank has over 650 members. The biggest challenges that we’re tackling right now include making sure the most interesting conversations and best content surfaces to the top of the community, making Slack more searchable (this is ultimately one of its flaws as a platform) and giving members a quicker way to find the exclusive content that we create. You’ll notice there’s a pretty clear theme here. In the past 30 days, 4,566 messages were posted in public channels inside Traffic Think Tank. If you add on any messages posted inside private direct messages, this number rises to 21,612. That’s a lot of messages. To solve these challenges and enable further scale in the future, we’ve invested a bunch of cash and our time into building out a full learning management system (LMS) that all members will get access to alongside the Slack community. The LMS will be a web-based portal that houses all of the video content we produce. It will also  provide an account admin section where users can update or change their billing information (they have to email us to do this right now, which isn’t ideal), a list of membership perks and discounts with our partners, and a list of links to some of the best threads within Slack – when clicked, these will drop you directly into Slack. [\[IMAGE\] Designs for the new learning management system (LMS)](https://cdn.shortpixel.ai/client/qglossy,retimg,w_2378/https://www.matthewbarby.com/wp-content/uploads/2019/08/Traffic-Think-Tank-LMS.png) It’s not been easy, but we’re 95% of the way through this and I’m certain that it will have a hugely positive impact on the experience for our members. Alongside this we hired a community manager, Liz, who supports with any questions that our members have, coordinates with external experts to arrange webinars for the community, helps with new member onboarding, and has tightened up some of our processes around billing and general accounts admin. This was a great decision. Finally, we’ve started planning next year’s live event, which we plan to more than double in size to 350 attendees, and we decided to pick a slightly warmer location in Miami this time out. Stay tuned for me to have a complete meltdown 3 weeks from the event. Final thoughts When I look back on the journey we’ve had so far building Traffic Think Tank, there’s one very important piece to this puzzle that’s made all of this work that I’ve failed to mention so far: co-founder alignment. Building a community is a balancing act that relies heavily on those in charge being completely aligned. Nick, Ian and I completely trust each other and more importantly, are philosophically aligned on how we want to run and grow the community. If we didn’t have this, the friction between us could tear apart the entire community. Picking the right people to work with is important in any company, but when your business is literally about bringing people together, there’s no margin for error here.  While I’m sure there will be many more challenges ahead, knowing that we all trust each other to make decisions that fall in line with each of our core values makes these challenges dramatically easier to overcome. Finally, I’d like to thank all of our members for making the community what it is today – it’d be nothing without you and I promise that we’ll never take that for granted. ​ I originally posted this on my blog here. Welcoming all of your thoughts, comments, questions and I'll do my best to answer them :)

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security

Uzair Javaid, a Ph.D. with a passion for data privacy, co-founded Betterdata to tackle one of AI's most pressing challenges: protecting privacy while enabling innovation. Recently, Betterdata secured a lucrative contract with the US Department of Homeland Security, 1 of only 4 companies worldwide to do so and the only one in Asia. Here's how he did it: The Story So what's your story? I grew up in Peshawar, Pakistan, excelling in coding despite studying electrical engineering. Inspired by my professors, I set my sights on studying abroad and eventually earned a Ph.D. scholarship at NUS Singapore, specializing in data security and privacy. During my research, I ethically hacked Ethereum and published 15 papers—three times the requirement. While wrapping up my Ph.D., I explored startup ideas and joined Entrepreneur First, where I met Kevin Yee. With his expertise in generative models and mine in privacy, we founded Betterdata. Now, nearly three years in, we’ve secured a major contract with the U.S. Department of Homeland Security—one of only four companies globally and the only one from Asia. The Startup In a nutshell, what does your startup do? Betterdata is a startup that uses AI and synthetic data generation to address two major challenges: data privacy and the scarcity of high-quality data for training AI models. By leveraging generative models and privacy-enhancing technologies, Betterdata enables businesses, such as banks, to use customer data without breaching privacy regulations. The platform trains AI on real data, learns its patterns, and generates synthetic data that mimics the real thing without containing any personal or sensitive information. This allows companies to innovate and develop AI solutions safely and ethically, all while tackling the growing need for diverse, high-quality data in AI development. How did you conduct ideation and validation for your startup? The initial idea for Betterdata came from personal experience. During my Ph.D., I ethically hacked Ethereum’s blockchain, exposing flaws in encryption-based data sharing. This led me to explore AI-driven deep synthesis technology—similar to deepfakes but for structured data privacy. With GDPR impacting 28M+ businesses, I saw a massive opportunity to help enterprises securely share data while staying compliant. To validate the idea, I spoke to 50 potential customers—a number that strikes the right balance. Some say 100, but that’s impractical for early-stage founders. At 50, patterns emerge: if 3 out of 10 mention the same problem, and this repeats across 50, you have 10–15 strong signals, making it a solid foundation for an MVP. Instead of outbound sales, which I dislike, we used three key methods: Account-Based Marketing (ABM)—targeting technically savvy users with solutions for niche problems, like scaling synthetic data for banks. Targeted Content Marketing—regular customer conversations shaped our thought leadership and outreach. Raising Awareness Through Partnerships—collaborating with NUS, Singapore’s PDPC, and Plug and Play to build credibility and educate the market. These strategies attracted serious customers willing to pay, guiding Betterdata’s product development and market fit. How did you approach the initial building and ongoing product development? In the early stages, we built synthetic data generation algorithms and a basic UI for proof-of-concept, using open-source datasets to engage with banks. We quickly learned that banks wouldn't share actual customer data due to privacy concerns, so we had to conduct on-site installations and gather feedback to refine our MVP. Through continuous consultation with customers, we discovered real enterprise data posed challenges, such as missing values, which led us to adapt our prototype accordingly. This iterative approach of listening to customer feedback and observing their usage allowed us to improve our product, enhance UX, and address unmet needs while building trust and loyalty. Working closely with our customers also gives us a data advantage. Our solution’s effectiveness depends on customer data, which we can't fully access, but bridging this knowledge gap gives us a competitive edge. The more customers we test on, the more our algorithms adapt to diverse use cases, making it harder for competitors to replicate our insights. My approach to iteration is simple: focus solely on customer feedback and ignore external noise like trends or advice. The key question for the team is: which customer is asking for this feature or solution? As long as there's a clear answer, we move forward. External influences, such as AI hype, often bring more confusion than clarity. True long-term success comes from solving real customer problems, not chasing trends. Customers may not always know exactly what they want, but they understand their problems. Our job is to identify these problems and solve them in innovative ways. While customers may suggest specific features, we stay focused on solving the core issue rather than just fulfilling their exact requests. The idea aligns with the quote often attributed to Henry Ford: "If I asked people what they wanted, they would have said faster horses." The key is understanding their problems, not just taking requests at face value. How do you assess product-market fit? To assess product-market fit, we track two key metrics: Customers' Willingness to Pay: We measure both the quantity and quality of meetings with potential customers. A high number of meetings with key decision-makers signals genuine interest. At Betterdata, we focused on getting meetings with people in banks and large enterprises to gauge our product's resonance with the target market. How Much Customers Are Willing to Pay: We monitor the price customers are willing to pay, especially in the early stages. For us, large enterprises, like banks, were willing to pay a premium for our synthetic data platform due to the growing need for privacy tech. This feedback guided our product refinement and scaling strategy. By focusing on these metrics, we refined our product and positioned it for scaling. What is your business model? We employ a structured, phase-driven approach for out business model, as a B2B startup. I initially struggled with focusing on the core value proposition in sales, often becoming overly educational. Eventually, we developed a product roadmap with models that allowed us to match customer needs to specific offerings and justify our pricing. Our pricing structure includes project-based pilots and annual contracts for successful deployments. At Betterdata, our customer engagement unfolds across three phases: Phase 1: Trial and Benchmarking \- We start with outreach and use open-source datasets to showcase results, offering customers a trial period to evaluate the solution. Phase 2: Pilot or PoC \- After positive trial results, we conduct a PoC or pilot using the customer’s private data, with the understanding that successful pilots lead to an annual contract. Phase 3: Multi-Year Contracts \- Following a successful pilot, we transition to long-term commercial contracts, focusing on multi-year agreements to ensure stability and ongoing partnerships. How do you do marketing for your brand? We take a non-conventional approach to marketing, focusing on answering one key question: Which customers are willing to pay, and how much? This drives our messaging to show how our solution meets their needs. Our strategy centers around two main components: Building a network of lead magnets \- These are influential figures like senior advisors, thought leaders, and strategic partners. Engaging with institutions like IMDA, SUTD, and investors like Plug and Play helps us gain access to the right people and foster warm introductions, which shorten our sales cycle and ensure we’re reaching the right audience. Thought leadership \- We build our brand through customer traction, technology evidence, and regulatory guidelines. This helps us establish credibility in the market and position ourselves as trusted leaders in our field. This holistic approach has enabled us to navigate diverse market conditions in Asia and grow our B2B relationships. By focusing on these areas, we drive business growth and establish strong trust with stakeholders. What's your advice for fundraising? Here are my key takeaways for other founders when it comes to fundraising: Fundraise When You Don’t Need To We closed our seed round in April 2023, a time when we weren't actively raising. Founders should always be in fundraising mode, even when they're not immediately in need of capital. Don’t wait until you have only a few months of runway left. Keep the pipeline open and build relationships. When the timing is right, execution becomes much easier. For us, our investment came through a combination of referrals and inbound interest. Even our lead investor initially rejected us, but after re-engaging, things eventually fell into place. It’s crucial to stay humble, treat everyone with respect, and maintain those relationships for when the time is right. Be Mindful of How You Present Information When fundraising, how you present information matters a lot. We created a comprehensive, easily digestible investment memo, hosted on Notion, which included everything an investor might need—problem, solution, market, team, risks, opportunities, and data. The goal was for investors to be able to get the full picture within 30 minutes without chasing down extra details. We also focused on making our financial model clear and meaningful, even though a 5-year forecast might be overkill at the seed stage. The key was clarity and conciseness, and making it as easy as possible for investors to understand the opportunity. I learned that brevity and simplicity are often the best ways to make a memorable impact. For the pitch itself, keep it simple and focus on 4 things: problem, solution, team, and market. If you can summarize each of these clearly and concisely, you’ll have a compelling pitch. Later on, you can expand into market segments, traction, and other metrics, but for seed-stage, focus on those four areas, and make sure you’re strong in at least three of them. If you do, you'll have a compelling case. How do you run things day-to-day? i.e what's your operational workflow and team structure? Here's an overview of our team structure and process: Internally: Our team is divided into two main areas: backend (internal team) and frontend (market-facing team). There's no formal hierarchy within the backend team. We all operate as equals, defining our goals based on what needs to be developed, assigning tasks, and meeting weekly to share updates and review progress. The focus is on full ownership of tasks and accountability for getting things done. I also contribute to product development, identifying challenges and clearing obstacles to help the team move forward. Backend Team: We approach tasks based on the scope defined by customers, with no blame or hierarchy. It's like a sports team—sometimes someone excels, and other times they struggle, but we support each other and move forward together. Everyone has the creative freedom to work in the way that suits them best, but we establish regular meetings and check-ins to ensure alignment and progress. Frontend Team: For the market-facing side, we implement a hierarchy because the market expects this structure. If I present myself as "CEO," it signals authority and credibility. This distinction affects how we communicate with the market and how we build our brand. The frontend team is split into four main areas: Business Product (Software Engineering) Machine Learning Engineering R&D The C-suite sits at the top, followed by team leads, and then the executors. We distill market expectations into actionable tasks, ensuring that everyone is clear on their role and responsibilities. Process: We start by receiving market expectations and defining tasks based on them. Tasks are assigned to relevant teams, and execution happens with no communication barriers between team members. This ensures seamless collaboration and focused execution. The main goal is always effectiveness—getting things done efficiently while maintaining flexibility in how individuals approach their work. In both teams, there's an emphasis on accountability, collaboration, and clear communication, but the structure varies according to the nature of the work and external expectations.

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies)
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies)

AI Palette is an AI-driven platform that helps food and beverage companies predict emerging product trends. I had the opportunity recently to sit down with the founder to get his advice on building an AI-first startup, which he'll be going through in this post. About AI Palette: Co-founders: >!2 (Somsubhra GanChoudhuri, Himanshu Upreti)!!100+!!$12.7M USD!!AI-powered predictive analytics for the CPG (Consumer Packaged Goods) industry!!Signed first paying customer in the first year!!65+ global brands, including Cargill, Diageo, Ajinomoto, Symrise, Mondelez, and L’Oréal, use AI Palette!!Every new product launched has secured a paying client within months!!Expanded into Beauty & Personal Care (BPC), onboarding one of India’s largest BPC companies within weeks!!Launched multiple new product lines in the last two years, creating a unified suite for brand innovation!Identify the pain points in your industry for ideas* When I was working in the flavour and fragrance industry, I noticed a major issue CPG companies faced: launching a product took at least one to two years. For instance, if a company decided today to launch a new juice, it wouldn’t hit the market until 2027. This long timeline made it difficult to stay relevant and on top of trends. Another big problem I noticed was that companies relied heavily on market research to determine what products to launch. While this might work for current consumer preferences, it was highly inefficient since the product wouldn’t actually reach the market for several years. By the time the product launched, the consumer trends had already shifted, making that research outdated. That’s where AI can play a crucial role. Instead of looking at what consumers like today, we realised that companies should use AI to predict what they will want next. This allows businesses to create products that are ahead of the curve. Right now, the failure rate for new product launches is alarmingly high, with 8 out of 10 products failing. By leveraging AI, companies can avoid wasting resources on products that won’t succeed, leading to better, more successful launches. Start by talking to as many industry experts as possible to identify the real problems When we first had the idea for AI Palette, it was just a hunch, a gut feeling—we had no idea whether people would actually pay for it. To validate the idea, we reached out to as many people as we could within the industry. Since our focus area was all about consumer insights, we spoke to professionals in the CPG sector, particularly those in the insights departments of CPG companies. Through these early conversations, we began to see a common pattern emerge and identified the exact problem we wanted to solve. Don’t tell people what you’re building—listen to their frustrations and challenges first. Going into these early customer conversations, our goal was to listen and understand their challenges without telling them what we were trying to build. This is crucial as it ensures that you can gather as much data about the problem to truly understand it and that you aren't biasing their answers by showing your solution. This process helped us in two key ways: First, it validated that there was a real problem in the industry through the number of people who spoke about experiencing the same problem. Second, it allowed us to understand the exact scale and depth of the problem—e.g., how much money companies were spending on consumer research, what kind of tools they were currently using, etc. Narrow down your focus to a small, actionable area to solve initially. Once we were certain that there was a clear problem worth solving, we didn’t try to tackle everything at once. As a small team of two people, we started by focusing on a specific area of the problem—something big enough to matter but small enough for us to handle. Then, we approached customers with a potential solution and asked them for feedback. We learnt that our solution seemed promising, but we wanted to validate it further. If customers are willing to pay you for the solution, it’s a strong validation signal for market demand. One of our early customer interviewees even asked us to deliver the solution, which we did manually at first. We used machine learning models to analyse the data and presented the results in a slide deck. They paid us for the work, which was a critical moment. It meant we had something with real potential, and we had customers willing to pay us before we had even built the full product. This was the key validation that we needed. By the time we were ready to build the product, we had already gathered crucial insights from our early customers. We understood the specific information they wanted and how they wanted the results to be presented. This input was invaluable in shaping the development of our final product. Building & Product Development Start with a simple concept/design to validate with customers before building When we realised the problem and solution, we began by designing the product, but not by jumping straight into coding. Instead, we created wireframes and user interfaces using tools like InVision and Figma. This allowed us to visually represent the product without the need for backend or frontend development at first. The goal was to showcase how the product would look and feel, helping potential customers understand its value before we even started building. We showed these designs to potential customers and asked for feedback. Would they want to buy this product? Would they pay for it? We didn’t dive into actual development until we found a customer willing to pay a significant amount for the solution. This approach helped us ensure we were on the right track and didn’t waste time or resources building something customers didn’t actually want. Deliver your solution using a manual consulting approach before developing an automated product Initially, we solved problems for customers in a more "consulting" manner, delivering insights manually. Recall how I mentioned that when one of our early customer interviewees asked us to deliver the solution, we initially did it manually by using machine learning models to analyse the data and presenting the results to them in a slide deck. This works for the initial stages of validating your solution, as you don't want to invest too much time into building a full-blown MVP before understanding the exact features and functionalities that your users want. However, after confirming that customers were willing to pay for what we provided, we moved forward with actual product development. This shift from a manual service to product development was key to scaling in a sustainable manner, as our building was guided by real-world feedback and insights rather than intuition. Let ongoing customer feedback drive iteration and the product roadmap Once we built the first version of the product, it was basic, solving only one problem. But as we worked closely with customers, they requested additional features and functionalities to make it more useful. As a result, we continued to evolve the product to handle more complex use cases, gradually developing new modules based on customer feedback. Product development is a continuous process. Our early customers pushed us to expand features and modules, from solving just 20% of their problems to tackling 50–60% of their needs. These demands shaped our product roadmap and guided the development of new features, ultimately resulting in a more complete solution. Revenue and user numbers are key metrics for assessing product-market fit. However, critical mass varies across industries Product-market fit (PMF) can often be gauged by looking at the size of your revenue and the number of customers you're serving. Once you've reached a certain critical mass of customers, you can usually tell that you're starting to hit product-market fit. However, this critical mass varies by industry and the type of customers you're targeting. For example, if you're building an app for a broad consumer market, you may need thousands of users. But for enterprise software, product-market fit may be reached with just a few dozen key customers. Compare customer engagement and retention with other available solutions on the market for product-market fit Revenue and the number of customers alone isn't always enough to determine if you're reaching product-market fit. The type of customer and the use case for your product also matter. The level of engagement with your product—how much time users are spending on the platform—is also an important metric to track. The more time they spend, the more likely it is that your product is meeting a crucial need. Another way to evaluate product-market fit is by assessing retention, i.e whether users are returning to your platform and relying on it consistently, as compared to other solutions available. That's another key indication that your solution is gaining traction in the market. Business Model & Monetisation Prioritise scalability Initially, we started with a consulting-type model where we tailor-made specific solutions for each customer use-case we encountered and delivered the CPG insights manually, but we soon realized that this wasn't scalable. The problem with consulting is that you need to do the same work repeatedly for every new project, which requires a large team to handle the workload. That is not how you sustain a high-growth startup. To solve this, we focused on building a product that would address the most common problems faced by our customers. Once built, this product could be sold to thousands of customers without significant overheads, making the business scalable. With this in mind, we decided on a SaaS (Software as a Service) business model. The benefit of SaaS is that once you create the software, you can sell it to many customers without adding extra overhead. This results in a business with higher margins, where the same product can serve many customers simultaneously, making it much more efficient than the consulting model. Adopt a predictable, simplistic business model for efficiency. Look to industry practices for guidance When it came to monetisation, we considered the needs of our CPG customers, who I knew from experience were already accustomed to paying annual subscriptions for sales databases and other software services. We decided to adopt the same model and charge our customers an annual upfront fee. This model worked well for our target market, aligning with industry standards and ensuring stable, recurring revenue. Moreover, our target CPG customers were already used to this business model and didn't have to choose from a huge variety of payment options, making closing sales a straightforward and efficient process. Marketing & Sales Educate the market to position yourself as a thought leader When we started, AI was not widely understood, especially in the CPG industry. We had to create awareness around both AI and its potential value. Our strategy focused on educating potential users and customers about AI, its relevance, and why they should invest in it. This education was crucial to the success of our marketing efforts. To establish credibility, we adopted a thought leadership approach. We wrote blogs on the importance of AI and how it could solve problems for CPG companies. We also participated in events and conferences to demonstrate our expertise in applying AI to the industry. This helped us build our brand and reputation as leaders in the AI space for CPG, and word-of-mouth spread as customers recognized us as the go-to company for AI solutions. It’s tempting for startups to offer products for free in the hopes of gaining early traction with customers, but this approach doesn't work in the long run. Free offerings don’t establish the value of your product, and customers may not take them seriously. You should always charge for pilots, even if the fee is minimal, to ensure that the customer is serious about potentially working with you, and that they are committed and engaged with the product. Pilots/POCs/Demos should aim to give a "flavour" of what you can deliver A paid pilot/POC trial also gives you the opportunity to provide a “flavour” of what your product can deliver, helping to build confidence and trust with the client. It allows customers to experience a detailed preview of what your product can do, which builds anticipation and desire for the full functionality. During this phase, ensure your product is built to give them a taste of the value you can provide, which sets the stage for a broader, more impactful adoption down the line. Fundraising & Financial Management Leverage PR to generate inbound interest from VCs When it comes to fundraising, our approach was fairly traditional—we reached out to VCs and used connections from existing investors to make introductions. However, looking back, one thing that really helped us build momentum during our fundraising process was getting featured in Tech in Asia. This wasn’t planned; it just so happened that Tech in Asia was doing a series on AI startups in Southeast Asia and they reached out to us for an article. During the interview, they asked if we were fundraising, and we mentioned that we were. As a result, several VCs we hadn’t yet contacted reached out to us. This inbound interest was incredibly valuable, and we found it far more effective than our outbound efforts. So, if you can, try to generate some PR attention—it can help create inbound interest from VCs, and that interest is typically much stronger and more promising than any outbound strategies because they've gone out of their way to reach out to you. Be well-prepared and deliberate about fundraising. Keep trying and don't lose heart When pitching to VCs, it’s crucial to be thoroughly prepared, as you typically only get one shot at making an impression. If you mess up, it’s unlikely they’ll give you a second chance. You need to have key metrics at your fingertips, especially if you're running a SaaS company. Be ready to answer questions like: What’s your retention rate? What are your projections for the year? How much will you close? What’s your average contract value? These numbers should be at the top of your mind. Additionally, fundraising should be treated as a structured process, not something you do on the side while juggling other tasks. When you start, create a clear plan: identify 20 VCs to reach out to each week. By planning ahead, you’ll maintain momentum and speed up the process. Fundraising can be exhausting and disheartening, especially when you face multiple rejections. Remember, you just need one investor to say yes to make it all worthwhile. When using funds, prioritise profitability and grow only when necessary. Don't rely on funding to survive. In the past, the common advice for startups was to raise money, burn through it quickly, and use it to boost revenue numbers, even if that meant operating at a loss. The idea was that profitability wasn’t the main focus, and the goal was to show rapid growth for the next funding round. However, times have changed, especially with the shift from “funding summer” to “funding winter.” My advice now is to aim for profitability as soon as possible and grow only when it's truly needed. For example, it’s tempting to hire a large team when you have substantial funds in the bank, but ask yourself: Do you really need 10 new hires, or could you get by with just four? Growing too quickly can lead to unnecessary expenses, so focus on reaching profitability as soon as possible, rather than just inflating your team or burn rate. The key takeaway is to spend your funds wisely and only when absolutely necessary to reach profitability. You want to avoid becoming dependent on future VC investments to keep your company afloat. Instead, prioritize reaching break-even as quickly as you can, so you're not reliant on external funding to survive in the long run. Team-Building & Leadership Look for complementary skill sets in co-founders When choosing a co-founder, it’s important to find someone with a complementary skill set, not just someone you’re close to. For example, I come from a business and commercial background, so I needed someone with technical expertise. That’s when I found my co-founder, Himanshu, who had experience in machine learning and AI. He was a great match because his technical knowledge complemented my business skills, and together we formed a strong team. It might seem natural to choose your best friend as your co-founder, but this can often lead to conflict. Chances are, you and your best friend share similar interests, skills, and backgrounds, which doesn’t bring diversity to the table. If both of you come from the same industry or have the same strengths, you may end up butting heads on how things should be done. Having diverse skill sets helps avoid this and fosters a more collaborative working relationship. Himanshu (left) and Somsubhra (right) co-founded AI Palette in 2018 Define roles clearly to prevent co-founder conflict To avoid conflict, it’s essential that your roles as co-founders are clearly defined from the beginning. If your co-founder and you have distinct responsibilities, there is no room for overlap or disagreement. This ensures that both of you can work without stepping on each other's toes, and there’s mutual respect for each other’s expertise. This is another reason as to why it helps to have a co-founder with a complementary skillset to yours. Not only is having similar industry backgrounds and skillsets not particularly useful when building out your startup, it's also more likely to lead to conflicts since you both have similar subject expertise. On the other hand, if your co-founder is an expert in something that you're not, you're less likely to argue with them about their decisions regarding that aspect of the business and vice versa when it comes to your decisions. Look for employees who are driven by your mission, not salary For early-stage startups, the first hires are crucial. These employees need to be highly motivated and excited about the mission. Since the salary will likely be low and the work demanding, they must be driven by something beyond just the paycheck. The right employees are the swash-buckling pirates and romantics, i.e those who are genuinely passionate about the startup’s vision and want to be part of something impactful beyond material gains. When employees are motivated by the mission, they are more likely to stick around and help take the startup to greater heights. A litmus test for hiring: Would you be excited to work with them on a Sunday? One of the most important rounds in the hiring process is the culture fit round. This is where you assess whether a candidate shares the same values as you and your team. A key question to ask yourself is: "Would I be excited to work with this person on a Sunday?" If there’s any doubt about your answer, it’s likely not a good fit. The idea is that you want employees who align with the company's culture and values and who you would enjoy collaborating with even outside of regular work hours. How we structure the team at AI Palette We have three broad functions in our organization. The first two are the big ones: Technical Team – This is the core of our product and technology. This team is responsible for product development and incorporating customer feedback into improving the technology Commercial Team – This includes sales, marketing, customer service, account managers, and so on, handling everything related to business growth and customer relations. General and Administrative Team – This smaller team supports functions like finance, HR, and administration. As with almost all businesses, we have teams that address the two core tasks of building (technical team) and selling (commercial team), but given the size we're at now, having the administrative team helps smoothen operations. Set broad goals but let your teams decide on execution What I've done is recruit highly skilled people who don't need me to micromanage them on a day-to-day basis. They're experts in their roles, and as Steve Jobs said, when you hire the right person, you don't have to tell them what to do—they understand the purpose and tell you what to do. So, my job as the CEO is to set the broader goals for them, review the plans they have to achieve those goals, and periodically check in on progress. For example, if our broad goal is to meet a certain revenue target, I break it down across teams: For the sales team, I’ll look at how they plan to hit that target—how many customers they need to sell to, how many salespeople they need, and what tactics and strategies they plan to use. For the technical team, I’ll evaluate our product offerings—whether they think we need to build new products to attract more customers, and whether they think it's scalable for the number of customers we plan to serve. This way, the entire organization's tasks are cascaded in alignment with our overarching goals, with me setting the direction and leaving the details of execution to the skilled team members that I hire.

[Ultimate List] A list of Marketing Tools That I’ve tested over the years and found helpful to do better marketing with less work. More than 50 Tools To Help you with Marketing, Copywriting & Sales!
reddit
LLM Vibe Score0
Human Vibe Score0.973
lazymentorsThis week

[Ultimate List] A list of Marketing Tools That I’ve tested over the years and found helpful to do better marketing with less work. More than 50 Tools To Help you with Marketing, Copywriting & Sales!

Starting to focus on marketing for your business, You will come across the same tools mentioned over and over by marketers. I would like to mention here tools that you might haven’t seen going viral in the community but actually will help you grow faster and efficiently. Starting off with My favourite Marketing Channel! #Email Marketing For SMBs Convertkit / Mailerlite / Mailchimp - These 3 Platforms are the best options for SMBs and entrepreneurs just starting out with email marketing. All 3 have free plans up to 1,000 subscribers. Scribe - Email Signature Tool, Create Great Email signatures for your emails. Liramail - Most Email marketing platforms don’t offer great email templates. This tool will help you build great email templates with drag and drop. Quick mail Auto-Warmer - Most Businesses at the beginning don’t know what to do when open rate drops. You need to use an email warmer like this to keep it up. #Email Marketing For Big Businesses SendGrid - Overall Email Marketing Tools, this tool is best for brands that have huge email lists and email marketing is the key marketing channel. Braze - This tool is leading in email marketing for large Email senders. When I was working for agencies, this was one of the best email marketing tools I had used. NeoCertified - Protect your emails for spammers and threats. To keep your email list healthy, this is a must have! Sparkloop - Referral Marketing For Email Campaigns. Email can generate great huge amount of referrals for you and Sparkloop makes it easier. #Cold Emails & Lead Generation Hunter - A Great Tool to scrape emails from domain names. The tool comes with a green free plan but Pro plan is worth the amount of features it provides. Icyleads - It’s better than Hunter as it’s heavily focused on the sales and prospecting to help you derive great results from your campaigns. Mailshake - Beginner Friend Cold Email Tool with Great features like email list warming. #Communication Tools Twilio - One do the best customer engagement platform used by Companies like Stripe and mine too. Chatlio - Use Live chat feature on your website with slack integration. My favourite easier to catch up on conversations through slack integration. Intercom - Used by Most Marketers, Industry Leading customer communication platform. Great for beginners! Chatwoot - Another Amazing Communication Tool but the best part is they have a great free plan useful for new businesses. Loom - Communicate with your audience through Videos. Loom is great for SaaS and to show human interaction to close new visitors effectively. #CRM Outseta - This tool provides great CRM and their billing system is better than other tools out their which makes it stands out! Hubspot - I don’t think this tool needs an introduction because Hubspot’s CRM is the best in industry. Salesflare - This CRM is a great alternative to hubspot as it’s beginner friendly and helpful for SMBs. #SEO Tools Ahrefs - One of the best SEO tool in the industry. They also just launched a bunch of free tools to help SEO beginners. Screaming frog - The only website crawler I have used since I bought my first domain. It’s the best! Ubersuggest- The Tool by Neil Patel is the best SEO tool for you. (I’m Joking, it’s the worst) Contentking - This tool is good at Real-time SEO Auditing, they do a lot of Marketing work through Newsletters. If you are subscribed to any SEO newsletter. You may have seen this tool. SEOquake & Semrush - SEOquake is a great tool to conduct on-page analysis, SERP, and much more. Great tool but it’s owned by Semrush. You should go for Semrush because that tool will cover all SEO aspects for you. #Content Marketing Buzzsumo - This tool is great for content research and but you may find the regular emails pretty annoying sometimes. Contentrow - Analyse Your Content and find it’s strength. Highly recommended who are weak at content structuring like me. Grammarly - If you are not a native English speaker like me, you might think you need it or not. You need it for sure for grammar corrections. #Graphic Design Tools Visme - At agencies, Infographics can be more effective than usual postscript. Visme is a graphic design tool focused on infographics and designs related to B2B and B2C. It’s great for agencies! Glorify - A Graphic Design Tool focused on E-commerce, filled with Designs useful for E-commerce store owners. Canva - All-in-one Industry leading Graphic Design Tool that everyone knows and every template is overused now. Adobe Creative Cloud ( previously Sparkpost) - It’s a great alternative to Canva filled with Amazing Stock images to use in your visuals but the only backlash is the exports in this tool are not high quality. Snaps - A Canva Alternative that might not have overused templates for your Social Accounts. #Advertising Tools Plai - It’s a great PPC tool to create Ads for Instagram and Tiktok. Wordstream - It’s an industry leading PPC Tool, great for Ad Grading and auditing. AdEspresso - This Is a tool by Hootsuite. They have a lot of Data sourced at the backend, which helps in Ad optimisation through this tool. That’s the reason I recommend this tool. #Video Editing Tools Veed Studio - I have been using Veed from last year. It’s one of the best Video Marketing Tool Optimized for Instagram & Tiktok. Synthesia - It’s a new AI video generation platform. From last few months, if you have seen marketing agencies including Videos in Emails. The chances are that’s not a Agency member taking but AI generated Human. Motionbox - It’s also a great video editing tool focused on video editing for Digital Marketers. Jitter Video - It’s a great motion design tool. Comes with great templates, the only place where other tools I mentioned lacks. It’s great and beginner friendly. #Copywriting Jasper AI - Google’s John Mueller says AI generated content is banned on Search but I think with Jasper AI you can generate SEO optimised Content but you have to put in some efforts like at least give 30 minutes for editing the Copy by yourself. Copy AI - Another AI tool to help you write better copy. This one is more focused on helping you write copy suitable for Ads and Social media campaigns. Hemingway App - To help you write more clearly and Bold. This tool is better than Grammarly if you look for writing perspective and it’s free. #Social Media Management App I’ve used a Lot of SMM Tools and that’s why going to mention all of them with a short review. Sprout social - The Best with deep insights coverage. Hootsuite - Great Scheduling tool just under sprout social. Later - Heavily Focused on Instagram from beginning and Now Tiktok too. SkedSocial - It’s like a Later alternative with great addition features like link-in-bio. Facebook’s Business Manager- Great but sometimes bugs can make a huge issue for you and customer support is like dead. Tweet Hunter & Hypefury- Both are Twitter Scheduling tools growing very fast on platform and are great for growth. Buffer - It’s a great tool but I haven’t seen any new updates to help with management. Zoho Social - It’s a great SMM tool and if you use other marketing solutions from Zoho. It’s a must have! #Market Research Tool • SparkToro - That’s the only one I have ever used. It’s great for audience research and comes with great customer service. Founded by Rand Fishkin, it’s one of the best research tool. #Influencer Marketing & UGC InfluenceGrid - A free search engine To find Tiktok & Instagram Influencers for your campaigns. Tiktok Creative Center- TikTok’s in-built tool called “Creative Center” is the best to find content trends, audience demographics and much more. Archive - Find Instagram Stories and Posts mentioning Your brands and use them as Ads for your business Marketing. #Landing Page Builders Leadpages - Its a great landing page builder because the integration and drag-and-drop features makes it easier to work with! Cardd co - A Great Landing page builder with easy step up but it lacks the copywriting and tracking features. Instapage - It’s one of the best out and I think the overall product is effective enough to help you stand out with your landing page. Unbounce - It’s a great alternative to Instapage due its well polished landing page templates that might be helpful for you. #Community Building Mighty Networks - A Great Community building platform, and you can also sell courses within the platform. Circle so - A great alternative to Mighty networks focused on Communities specifically. We are currently using for small community Of ours. #Sales Tools Drift - You can get much more out of Drift than just sales tools but The Sales solutions provided in Drift are one of the best. Salesforce - It’s the industry Sales solution provider. A go-to and have various pricing plans making it suitable for majority of SMBs. #Social Proof Tools People don’t have enough time to search across internet to decide to trust you after seeing your Ad first time. That’s what you might be facing too. Here are two tools I absolutely love for social proof! Use Proof - Show Recent Activities occurring on your website and build the trust of your visitors. Testimonial to - Gather Testimonials across Social Media platforms related to your business with this tool. Capture tweets and comments mentioning your brands and mention them. #Analytics Tools Plausible Analytics- A privacy friendly Analytics alternative to Google Analytics if you hate Analytics 4 like me. Mixpanel - Product Analytics and funnel reports better than Google Analytics. #Reddit Marketing Gummysearch- This tool will help To find your target audience on Reddit and interact with them with its help and close your new customers. Howitzer- It’s another pretty similar tool to Gummysearch focused on Reddit cold outreach to get clients and new customers. Both are great but Gummysearch provides better customer support while Howtizer is helpful on a large scale Reddit Marketing. #Text Marketing Klaviyo - It’s an email + SMS marketing tool, it’s taking up space in marketing industry very quickly as an industry leader due to its great integrations but you need to learn the platform usage to maximise the outcome. Cartloop - This tool provides great text marketing solutions with integration with Spotify and other e-commerce marketing tools. Attentive Mobile - This is my favourite Text marketing tool due to the interactive dashboard + they have a library of Text marketing examples to help you out with your campaigns. #Other Tools I have used throughout my journey! Triple Whale - It’s a great E-commerce marketing tools with Triple pixel to help you track your campaigns more efficiently. Fastory - To create well optimized Instagram & Tiktok Stories for your business. Jotform - Online Form Builder with integrations with leading marketing tools. Gated - As an entrepreneur and marketer, you may receive a bunch of unwanted emails. Use Gated to get rid of them and receive useful mails only! ClickUp- The main Tool for Project Management, one of the best and highly recommended. Riverside - Forget Zoom or Google Meet, For your Podcast Interviews and Marketing conferences. You need riverside with great video quality and recording features. Manychat- Automate your Instagram DMs and interact with your followers more efficiently + sell out your products/ services when you are offline. Calendy - To schedule meetings with your ideal clients. ServiceProviderPro - It’s a client portal for SEO & Growing Agencies, very helpful in scaling agencies. SendCheckit - Compare your Email Subject Lines with 100,000+ others in the database for free. Otter AI - Using AI track your meetings more effectively, you can easily edit, annotate and share notes from the meetings. Ryte - Optimise your website User experience with this tool focused on UX aspects + SEO too. PhantomBuster - Scrape LinkedIn Profile and Data from Facebook/LinkedIn groups. I clearly love this tool! #Honourable Mentions Zapier - The Only tool you need to integrate your favourite tool with a new effective tool. Elementor - That’s what I use for web design and it’s great! Marketer Hire - To hire world class marketers to work with you. InShot & Capcut - I create Instagram Reels and TikTok’s and life without these tools isn’t possible. Nira - It’s a great tool to Manage your workspace and this tool has launched many marketing templates in-built helpful for marketers and also entrepreneurs. X - The tool you love that wasn’t mentioned here is valuable and I honour that tool and share that if you would like to! I mean thanks for reading what I have curated all over my life as a marketer. I share 5 Marketing Tools, 5 Marketing Resources and 1 Free Resourceevery week in my newsletter, you can subscribe here to receive that for free. Also, You can read an expanded list of email marketing tools in this Reddit post!

In 2018, I started an AI chatbot company...today, we have over 4000 paying customers and ChatGPT is changing EVERYTHING
reddit
LLM Vibe Score0
Human Vibe Score1
Millionaire_This week

In 2018, I started an AI chatbot company...today, we have over 4000 paying customers and ChatGPT is changing EVERYTHING

Intro: 5 years ago, my co-founders and I ventured into the space of AI chatbots and started our first truly successful company. Never in a million years did I see myself in this business and we truly stumbled upon the opportunity by chance. Prior to that, we ran a successful lead generation business and questioned whether a simple ai chat product would increase our online conversions. Of the 3 co-founders, I was skeptical that it would, but the data was clear that we had something that really worked. We built a really simple MVP version of the product and gave it to some of our top lead buyers who saw even better conversion improvements on their own websites. In just a matter of weeks, a new business opportunity was born and a major pivot away from our lead generation business started. Our growth story: Startup growth is really interesting and in most cases, founders aren't really educated on what a typical growth curve looks like. While we hear about "hockey stick" growth curves, it's really atypical to actually see or experience this. From my experience, growth curves take place in a "stair curve". For example, you can scrap your way to a $100k run rate without much process or tracking. You can even get to $1 million ARR being super disorganized. As you start going beyond $1M ARR, things start to break and growth can flatten out while you put new processes and systems in place. Eventually you'll get to $2M or 3M with your new strategy and then things start breaking again. I've seen the process repeat itself and as you increase your ARR, the processes and systems become more difficult to work through...mainly because more people get involved and the product becomes more complex. When you do end up cracking the code in each step, the growth accelerates faster and faster before things start to break down and flatten out again. Without getting too much into the numbers, here were some of our initial levers for growth: Our first "stair" step was to leverage our existing customer base from our prior lead generation business. Having prior business relationships and a proven track record made it really simple to have conversations with people who already trusted us to try something new that we had to offer. Stair #2 was to build out a partner channel. Since our chat product involved a web developer or agency installing the chat on client sites, we partnered with these developers and agencies to leverage their already existing customer bases. We essentially piggy-backed off of their relationships and gave them a cut of the revenue. We built an internal partner tracking portal which took 6+ months, but it was well worth it. Stair #3 was our most expensive step, biggest headache, but added the most revenue. After COVID, we had and SDR/Account Executive sales team of roughly 30 people. It added revenue fast, but the payback periods were 12+ months so we had to cut back on this strategy after exhausting our universe of clients. Stair #4 involves a variety of paid advertisement strategies with product changes and the introduction of new onboarding features. We're in the middle of this stair and hope it's multiple years before things breakdown again. Don't give up I know it sounds really cliché, but the #1 indicator of success is doing the really boring stuff day in and day out and making incremental improvements. As the weeks, months, and years pass by, you will slowly gain domain expertise and start to see the gaps in the market that can set you apart from your competition. It's so hard for founders to stay focused and not get distracted so I would say it's equally as important to have co-founders who hold each other accountable on what your collective goals are. How GPT is changing everything I could write pages and pages about how GPT is going to change how the world operates, but I'll keep it specific to our business and chatbots. In 2021, we built an industry specific AI model that did a great job of classifying intents which allowed us to train future actions during a chat. It was a great advancement in our customer's industry at the time. With GPT integrated into our system, that training process that would take an employee hours to do, can be done in 5 minutes. The model is also cheaper than our own and more accurate. Because of these training improvements, we have been able to conduct research that is allowing us to leverage GPT models like no one else in the industry. This is both in the realm of chat and also training during onboarding. I really want to refrain from sharing our company, but if you are interested in seeing a model trained for your specific company or website, just PM me your link and I'll send you a free testing link with a model fully trained for your site to play around with. Where we are headed and the dangers of AI The level of advancement in AI is not terribly dangerous in its current state. I'm sure you've heard it before, but those who leverage the technology today will be the ones who get ahead. In the coming years, AI will inevitably replace a large percentage of human labor. This will be great for overall value creation and productivity for the world, but the argument that humans have always adapted and new jobs will be created is sadly not going to be as relevant in this case. As the possibility of AGI becomes a reality in the coming years or decades, productivity through AI will be off the charts. There is a major risk that human innovation and creative thinking will be completely stalled...human potential as we know it will be capped off and there will need to be major economic reform for displaced workers. This may not happen in the next 5 or 10 years, but you would be naïve not to believe the world we live in today will not be completely different in 20 to 30 years. Using AI to create deepfakes, fake voice agents, scam the unsuspecting, or exploit technical vulnerabilities are just a few other examples I could write about, but don't want to go into to much detail for obvious reasons. Concluding If you found the post interesting or you have any questions, please don't hesitate to ask. I'll do my best to answer whatever questions come from this! ​ \*EDIT: Wasn't expecting this sort of response. I posted this right before I went to sleep so I'll get to responding soon.

Follow Along as I Flip this Website - Case Study
reddit
LLM Vibe Score0
Human Vibe Score1
jshogren10This week

Follow Along as I Flip this Website - Case Study

I am starting a new case study where I will be documenting my attempt to flip a website that I just purchased from Flippa. However, unlike most case studies where people hide certain parts and details from the public I will instead be sharing everything. That means you will know the exact URL of the site that I purchased and I will share everything with you all as I progress.I know that case studies are lot more interesting and you can learn better when you can see real examples of what I am talking about. Enough of the chatting, let's jump straight into this new case study and I will explain what this is all about. Before you get into the case study I want to give you the option of reading this one my website where all of the images can be seen within the post and it is easier to read. I also want to say that I have nothing to sell you or anything close to it. So if you want to read it there you can do so here ##Introductory Video I have put together a video that talks about many of the things that I cover in this article. So if you would rather watch a video you can watch that here - https://www.youtube.com/watch?v=EE3SxtNnqts However, I go into more detail in the actual article FYI. Also, I plan on using Youtube very frequently in this case study so be on the lookout for new videos.There is going to be a video that will accompany every single case study post because I like having it being presented in two different mediums. ##The Website I Just Bought Around a week ago I made a new website purchase from Flippa and you can view the website's Flippa listing here - https://flippa.com/6439965-hvactraining101-com Screenshot of the Homepage - http://imgur.com/T6Iv1QN I paid $1,250 for the site and you will soon see that I got a really good deal. As you might be able to tell from the URL, this site is focused around training and education for becoming a HVAC technician. This is a lucrative niche to be in and Adsense pays very well. I do not have control of the site yet due to the transfer process not being completed. However, I am hoping within a few days everything will be finalized and I will take full control of the site. In the meantime, I figured it would be a good time to put together the introduction post for this new case study! ##Why I Bought this Website Now that you have a general idea of the website that I purchased, I now want to explain the reasoning behind the purchase. There are 3 major reasons for this purchase and I will explain each one of them below. GREAT Price As I mentioned earlier, I bought this website for $1,250. However, that doesn't mean a whole lot unless you know how much the site is making each month. Screenshot of the earnings for the last 12 months - http://imgur.com/NptxCHy Average Monthly Profits: 3 Month = $126 6 Month = $128 12 Month = $229.50 Let's use the 6 month average of $128/month as our baseline average. Since it is making on average $128/month and it was sold for $1,250 then that means I bought this site at a multiple of 9.76x! Most sites in today's market go for 20x-30x multiples. As you can see, I got a great deal on this site. Although the great price was the biggest reason for me buying this site there are other factors that persuaded me as well. You need to remember that just because you can get a website for a good price it doesn't mean it is a good deal. There are other factors that you need to look at as well. Extremely Under Optimized This site is currently being monetized mainly by Adsense and a very small amount from Quinstreet. From my experience with testing and optimizing Adsense layouts for my site in my Website Investing case study I know the common ad layouts that work best for maximizing Adsense revenue. With that being said, I can quickly determine if a website is being under optimized in terms of the ad layout. One of the first things I did when analyzing this site was examine the ad layout it was using. Screenshot of the website with the ad layout the previous owner was using - http://imgur.com/wqleLVA There is only ONE ad per page being used, that's it. Google allows up to 6 total ads to be used per page and you can imagine how much money is being left on the table because of this. I am estimating that I can probably double the earnings for the site practically overnight once I add more ads to the site. Adding more ads in combination with my favorite Adsense plugin, AmpedSense, I will be able to easily boost the earnings for this site quickly. It is also worth mentioning how lucrative this niche is and how much advertisers are willing to spend on a per click basis. The average CPC for the top keywords this site is currently ranking for in Google - http://imgur.com/ifxiy8B Look at those average CPC numbers, they are insanely high! I could be making up to $25 per click for some of those keywords, which is so absurd to me. Combine these extremely high CPC with the fact that the site currently only has one ad per page and you can start to understand just how under optimized this site truly is. I also plan on utilizing other ad networks such as Quinstreet and Campus Explorer more as well. These two networks are targeted at the education niche which works very well with my site. I will be testing to see if these convert better than normal Adsense ads. Goldmine of Untapped Keywords One of the biggest opportunities I see for growing this site is to target local keywords related to HVAC training. As of right now, the site has only scratched the surface when it comes to trying to rank for state/city keywords. Currently there are only two pages on the entire website which go after local keywords, those two pages target Texas and Florida HVAC search terms. These two pages are two of the more popular pages in terms of total amount of traffic. See the screenshot of the Google Analytics - http://imgur.com/NB0xJ4G Two out of the top five most popular pages for the entire website are focused on local search terms. However, these are the ONLY two pages that target local search terms on the whole site! There are 48 other states, although there may not be search volume for all states, and countless cities that are not being targeted. Why do I think this is such a good opportunity? For a few reasons: Local keywords are a lot easier to rank for in Google than more general keywords This site has been able to rank for two states successfully already and it proves it is possible Traffic going to these local pages is WAY more targeted and will convert at a much higher rate, which means more commissions for me There are so many more states and cities that get a good amount of searches that I can target To give you an idea of the type of keywords these local pages rank for, you can see the top keywords that the Florida page is ranking for in Google: Top ranking keywords for the Florida page - http://imgur.com/j7uKzl2 As you can see these keywords don't get a ton of searches each month, but ranking 1st for a keyword getting 90 searches a month is better than being ranked 10th for a keyword getting 1,000 searches a month. I have started to do some keyword research for other states and I am liking what I am finding so far. Keywords that I have found which I will be targeting with future articles - http://imgur.com/8CCCCWU I will go into more detail about my keyword research in future articles, but I wanted to give you an idea of what my strategy will be! I also wanted to share why I am super excited about the future potential to grow this site by targeting local keywords. ##Risks Yes, there are many good things about this website, but there are always risks involved no matter what the investment is. The same thing goes for this site. Below are some of the risks that I currently see. HTML Site This website is a HTML site and I will need to transfer it to Wordpress ASAP. I have been doing some research on this process and it shouldn't be too hard to get this over to Wordpress. In doing so it will make adding content, managing the back end and just about everything else easier. Also, I am hoping that when I transfer it to Wordpress that it will become more optimized for Google which will increase keyword rankings. Declining Earnings Looking at the last 12 months of earnings you will notice a drop off from last year till now. Earnings from the last 12 months - http://imgur.com/WsotZsj In May of 2015 it looks like the site earned right around $500, which is much higher than the $128 that it is earning now. However, the last 7 or so months have been consistent which is a good sign. Even though the earnings are much lower now then they were a year ago it is good to know that this site has the potential to earn $500/month because it has done it before. Slightly Declining Traffic In the last 12 months the site's traffic has declined, however, it looks like it is picking back up. Traffic from the last 12 months - http://imgur.com/aiYZW9W The decline is nothing serious, but there is a drop on traffic. Let's take a look at the complete history of this site's traffic so we can get a better idea of what is going on here: Complete traffic history - http://imgur.com/tYmboVn The above screenshot is from 2012 all the way up to right now. In the grand scheme of things you can see that the traffic is still doing well and it looks like it is on the upswing now. Those three risks mentioned above are the three biggest risks with this site at this point. It is always good to note the risks and do everything you can to prevent them from causing a problem. ##My Growth Strategy Whenever I purchase a new site I always create an outline or plan on how I will grow the site. Right now, I have some basic ideas on how I will grow this site, but as I go on I will continue to change and optimize my strategies to be more effective. Below I have outlined my current plans to grow: Add more Adsense Ads The very first thing I will do once I get control of the site is add more ads per page. I am predicting that by just adding a few more ads per page I will be able to more than likely double the earnings. I will touch on exactly how I will be optimizing the ad layouts in future posts. Test other Ad Networks I will be doing a lot of testing and experimenting when it comes to the ad networks. I plan on trying out Adsense, Media.net, Quinstreet, Campus Explorer and finding the combination of those 4 which produces the most revenue. The Adsense and Media.net ads will perform well on the more general pages while Quinstreet and Campus Explorer ads will be geared towards the local search terms. There will probably be other ad networks I will try out but these are the four which I will be using right away. If you are aware of any other ad networks out there which are geared towards the education niche please let me know in the comments below! Target Local Keywords with new Content I have already touched on this, but I will starting to produce content targeting these local keywords ASAP. The sooner I add the content to the site the sooner it will start to rank and bring in traffic. I will not be writing my own content and instead I will be outsourcing all of it via Upwork. I will show you all how I go about outsourcing content production and you can see my process for doing that. ##Goals for this Website My goal for the website is to have it valued at $10,000+ within 12 months. Let's break down this larger goal into smaller chunks which will make achieving it easier and more attainable. Earnings - $500/month To get the site valued at $10,000 the site will need to be making $500/month using a 20x monthly multiple. Right now, the site is making around $130/month so it has a ways to before it reaches the $500 a month mark. However, after doing some Adsense optimization I think we could push the earnings to around $300/month without much work. From there, it will come down to trying to bring in more traffic! Traffic - 5,000 Visitors per Month Why 5,000 visitors? Because that is how much traffic it is going to take to get to the $500/month goal. Let me explain how I came to this conclusion: The average RPM for this site is currently $50, which means for every 1,000 page views the site earns $50. After I optimize the Adsense layout for the site and add more ads per page I think I will be able to double the RPM to $100. Using the RPM of $100 the site will need to have 5,000 monthly visitors to earn $500. So 5,000 monthly visitors is the traffic goal I have set and aiming for! The site is currently getting around 3,000 visitors per month so I will need to add an extra 2,000 visitors to get to this goal. ##Want to Follow this Case Study? I will be using Youtube a lot in this case study so make sure to follow my Youtube channel here - www.youtube.com/c/joshshogren Other than that, I think that is going to bring us to the end of the introductory post for this new case study. I hope that you enjoyed reading and that you are excited to follow along! If you have any suggestions to make this case study better PLEASE let me know in the comment below. I want to make this case study the best one I have done yet. Talk to you all in the comment section.

I run an AI automation agency (AAA). My honest overview and review of this new business model
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

I run an AI automation agency (AAA). My honest overview and review of this new business model

I started an AI tools directory in February, and then branched off that to start an AI automation agency (AAA) in June. So far I've come across a lot of unsustainable "ideas" to make money with AI, but at the same time a few diamonds in the rough that aren't fully tapped into yet- especially the AAA model. Thought I'd share this post to shine light into this new business model and share some ways you could potentially start your own agency, or at the very least know who you are dealing with and how to pick and choose when you (inevitably) get bombarded with cold emails from them down the line. Foreword Running an AAA does NOT involve using AI tools directly to generate and sell content directly. That ship has sailed, and unless you are happy with $5 from Fiverr every month or so, it is not a real business model. Cry me a river but generating generic art with AI and slapping it onto a T-shirt to sell on Etsy won't make you a dime. At the same time, the AAA model will NOT require you to have a deep theoretical knowledge of AI, or any academic degree, as we are more so dealing with the practical applications of generative AI and how we can implement these into different workflows and tech-stacks, rather than building AI models from the ground up. Regardless of all that, common sense and a willingness to learn will help (a shit ton), as with anything. Keep in mind - this WILL involve work and motivation as well. The mindset that AI somehow means everything can be done for you on autopilot is not the right way to approach things. The common theme of businesses I've seen who have successfully implemented AI into their operations is the willingess to work with AI in a way that augments their existing operations, rather than flat out replace a worker or team. And this is exactly the train of thought you need when working with AI as a business model. However, as the field is relatively unsaturated and hype surrounding AI is still fresh for enterprises, right now is the prime time to start something new if generative AI interests you at all. With that being said, I'll be going over three of the most successful AI-adjacent businesses I've seen over this past year, in addition to some tips and resources to point you in the right direction. so.. WTF is an AI Automation Agency? The AI automation agency (or as some YouTubers have coined it, the AAA model) at its core involves creating custom AI solutions for businesses. I have over 1500 AI tools listed in my directory, however the feedback I've received from some enterprise users is that ready-made SaaS tools are too generic to meet their specific needs. Combine this with the fact virtually no smaller companies have the time or skills required to develop custom solutions right off the bat, and you have yourself real demand. I would say in practice, the AAA model is quite similar to Wordpress and even web dev agencies, with the major difference being all solutions you develop will incorporate key aspects of AI AND automation. Which brings me to my second point- JUST AI IS NOT ENOUGH. Rather than reducing the amount of time required to complete certain tasks, I've seen many AI agencies make the mistake of recommending and (trying to) sell solutions that more likely than not increase the workload of their clients. For example, if you were to make an internal tool that has AI answer questions based on their knowledge base, but this knowledge base has to be updated manually, this is creating unnecessary work. As such I think one of the key components of building successful AI solutions is incorporating the new (Generative AI/LLMs) with the old (programmtic automation- think Zapier, APIs, etc.). Finally, for this business model to be successful, ideally you should target a niche in which you have already worked and understand pain points and needs. Not only does this make it much easier to get calls booked with prospects, the solutions you build will have much greater value to your clients (meaning you get paid more). A mistake I've seen many AAA operators make (and I blame this on the "Get Rich Quick" YouTubers) is focusing too much on a specific productized service, rather than really understanding the needs of businesses. The former is much done via a SaaS model, but when going the agency route the only thing that makes sense is building custom solutions. This is why I always take a consultant-first approach. You can only build once you understand what they actually need and how certain solutions may impact their operations, workflows, and bottom-line. Basics of How to Get Started Pick a niche. As I mentioned previously, preferably one that you've worked in before. Niches I know of that are actively being bombarded with cold emails include real estate, e-commerce, auto-dealerships, lawyers, and medical offices. There is a reason for this, but I will tell you straight up this business model works well if you target any white-collar service business (internal tools approach) or high volume businesses (customer facing tools approach). Setup your toolbox. If you wanted to start a pressure washing business, you would need a pressure-washer. This is no different. For those without programming knowledge, I've seen two common ways AAA get setup to build- one is having a network of on-call web developers, whether its personal contacts or simply going to Upwork or any talent sourcing agency. The second is having an arsenal of no-code tools. I'll get to this more in a second, but this works beecause at its core, when we are dealing with the practical applications of AI, the code is quite simple, simply put. Start cold sales. Unless you have a network already, this is not a step you can skip. You've already picked a niche, so all you have to do is find the right message. Keep cold emails short, sweet, but enticing- and it will help a lot if you did step 1 correctly and intimately understand who your audience is. I'll be touching base later about how you can leverage AI yourself to help you with outreach and closing. The beauty of gen AI and the AAA model You don't need to be a seasoned web developer to make this business model work. The large majority of solutions that SME clients want is best done using an API for an LLM for the actual AI aspect. The value we create with the solutions we build comes with the conceptual framework and design that not only does what they need it to but integrates smoothly with their existing tech-stack and workflow. The actual implementation is quite straightforward once you understand the high level design and know which tools you are going to use. To give you a sense, even if you plan to build out these apps yourself (say in Python) the large majority of the nitty gritty technical work has already been done for you, especially if you leverage Python libraries and packages that offer high level abstraction for LLM-related functions. For instance, calling GPT can be as little as a single line of code. (And there are no-code tools where these functions are simply an icon on a GUI). Aside from understanding the capabilities and limitations of these tools and frameworks, the only thing that matters is being able to put them in a way that makes sense for what you want to build. Which is why outsourcing and no-code tools both work in our case. Okay... but how TF am I suppposed to actually build out these solutions? Now the fun part. I highly recommend getting familiar with Langchain and LlamaIndex. Both are Python libraires that help a lot with the high-level LLM abstraction I mentioned previously. The two most important aspects include being able to integrate internal data sources/knowledge bases with LLMs, and have LLMs perform autonomous actions. The two most common methods respectively are RAG and output parsing. RAG (retrieval augmented Generation) If you've ever seen a tool that seemingly "trains" GPT on your own data, and wonder how it all works- well I have an answer from you. At a high level, the user query is first being fed to what's called a vector database to run vector search. Vector search basically lets you do semantic search where you are searching data based on meaning. The vector databases then retrieves the most relevant sections of text as it relates to the user query, and this text gets APPENDED to your GPT prompt to provide extra context to the AI. Further, with prompt engineering, you can limit GPT to only generate an answer if it can be found within this extra context, greatly limiting the chance of hallucination (this is where AI makes random shit up). Aside from vector databases, we can also implement RAG with other data sources and retrieval methods, for example SQL databses (via parsing the outputs of LLM's- more on this later). Autonomous Agents via Output Parsing A common need of clients has been having AI actually perform tasks, rather than simply spitting out text. For example, with autonomous agents, we can have an e-commerce chatbot do the work of a basic customer service rep (i.e. look into orders, refunds, shipping). At a high level, what's going on is that the response of the LLM is being used programmtically to determine which API to call. Keeping on with the e-commerce example, if I wanted a chatbot to check shipping status, I could have a LLM response within my app (not shown to the user) with a prompt that outputs a random hash or string, and programmatically I can determine which API call to make based on this hash/string. And using the same fundamental concept as with RAG, I can append the the API response to a final prompt that would spit out the answer for the user. How No Code Tools Can Fit In (With some example solutions you can build) With that being said, you don't necessarily need to do all of the above by coding yourself, with Python libraries or otherwise. However, I will say that having that high level overview will help IMMENSELY when it comes to using no-code tools to do the actual work for you. Regardless, here are a few common solutions you might build for clients as well as some no-code tools you can use to build them out. Ex. Solution 1: AI Chatbots for SMEs (Small and Medium Enterprises) This involves creating chatbots that handle user queries, lead gen, and so forth with AI, and will use the principles of RAG at heart. After getting the required data from your client (i.e. product catalogues, previous support tickets, FAQ, internal documentation), you upload this into your knowledge base and write a prompt that makes sense for your use case. One no-code tool that does this well is MyAskAI. The beauty of it especially for building external chatbots is the ability to quickly ingest entire websites into your knowledge base via a sitemap, and bulk uploading files. Essentially, they've covered the entire grunt work required to do this manually. Finally, you can create a inline or chat widget on your client's website with a few lines of HTML, or altneratively integrate it with a Slack/Teams chatbot (if you are going for an internal Q&A chatbot approach). Other tools you could use include Botpress and Voiceflow, however these are less for RAG and more for building out complete chatbot flows that may or may not incorporate LLMs. Both apps are essentially GUIs that eliminate the pain and tears and trying to implement complex flows manually, and both natively incoporate AI intents and a knowledge base feature. Ex. Solution 2: Internal Apps Similar to the first example, except we go beyond making just chatbots but tools such as report generation and really any sort of internal tool or automations that may incorporate LLM's. For instance, you can have a tool that automatically generates replies to inbound emails based on your client's knowledge base. Or an automation that does the same thing but for replies to Instagram comments. Another example could be a tool that generates a description and screeenshot based on a URL (useful for directory sites, made one for my own :P). Getting into more advanced implementations of LLMs, we can have tools that can generate entire drafts of reports (think 80+ pages), based not only on data from a knowledge base but also the writing style, format, and author voice of previous reports. One good tool to create content generation panels for your clients would be MindStudio. You can train LLM's via prompt engineering in a structured way with your own data to essentially fine tune them for whatever text you need it to generate. Furthermore, it has a GUI where you can dictate the entire AI flow. You can also upload data sources via multiple formats, including PDF, CSV, and Docx. For automations that require interactions between multiple apps, I recommend the OG zapier/make.com if you want a no-code solution. For instance, for the automatic email reply generator, I can have a trigger such that when an email is received, a custom AI reply is generated by MyAskAI, and finally a draft is created in my email client. Or, for an automation where I can create a social media posts on multiple platforms based on a RSS feed (news feed), I can implement this directly in Zapier with their native GPT action (see screenshot) As for more complex LLM flows that may require multiple layers of LLMs, data sources, and APIs working together to generate a single response i.e. a long form 100 page report, I would recommend tools such as Stack AI or Flowise (open-source alternative) to build these solutions out. Essentially, you get most of the functions and features of Python packages such as Langchain and LlamaIndex in a GUI. See screenshot for an example of a flow How the hell are you supposed to find clients? With all that being said, none of this matters if you can't find anyone to sell to. You will have to do cold sales, one way or the other, especially if you are brand new to the game. And what better way to sell your AI services than with AI itself? If we want to integrate AI into the cold outreach process, first we must identify what it's good at doing, and that's obviously writing a bunch of text, in a short amount of time. Similar to the solutions that an AAA can build for its clients, we can take advantage of the same principles in our own sales processes. How to do outreach Once you've identified your niche and their pain points/opportunities for automation, you want to craft a compelling message in which you can send via cold email and cold calls to get prospects booked on demos/consultations. I won't get into too much detail in terms of exactly how to write emails or calling scripts, as there are millions of resources to help with this, but I will tell you a few key points you want to keep in mind when doing outreach for your AAA. First, you want to keep in mind that many businesses are still hesitant about AI and may not understand what it really is or how it can benefit their operations. However, we can take advantage of how mass media has been reporting on AI this past year- at the very least people are AWARE that sooner or later they may have to implement AI into their businesses to stay competitive. We want to frame our message in a way that introduces generative AI as a technology that can have a direct, tangible, and positive impact on their business. Although it may be hard to quantify, I like to include estimates of man-hours saved or costs saved at least in my final proposals to prospects. Times are TOUGH right now, and money is expensive, so you need to have a compelling reason for businesses to get on board. Once you've gotten your messaging down, you will want to create a list of prospects to contact. Tools you can use to find prospects include Apollo.io, reply.io, zoominfo (expensive af), and Linkedin Sales Navigator. What specific job titles, etc. to target will depend on your niche but for smaller companies this will tend to be the owner. For white collar niches, i.e. law, the professional that will be directly benefiting from the tool (i.e. partners) may be better to contact. And for larger organizations you may want to target business improvement and digital transformation leads/directors- these are the people directly in charge of projects like what you may be proposing. Okay- so you have your message, and your list, and now all it comes down to is getting the good word out. I won't be going into the details of how to send these out, a quick Google search will give you hundreds of resources for cold outreach methods. However, personalization is key and beyond simple dynamic variables you want to make sure you can either personalize your email campaigns directly with AI (SmartWriter.ai is an example of a tool that can do this), or at the very least have the ability to import email messages programmatically. Alternatively, ask ChatGPT to make you a Python Script that can take in a list of emails, scrape info based on their linkedin URL or website, and all pass this onto a GPT prompt that specifies your messaging to generate an email. From there, send away. How tf do I close? Once you've got some prospects booked in on your meetings, you will need to close deals with them to turn them into clients. Call #1: Consultation Tying back to when I mentioned you want to take a consultant-first appraoch, you will want to listen closely to their goals and needs and understand their pain points. This would be the first call, and typically I would provide a high level overview of different solutions we could build to tacke these. It really helps to have a presentation available, so you can graphically demonstrate key points and key technologies. I like to use Plus AI for this, it's basically a Google Slides add-on that can generate slide decks for you. I copy and paste my default company messaging, add some key points for the presentation, and it comes out with pretty decent slides. Call #2: Demo The second call would involve a demo of one of these solutions, and typically I'll quickly prototype it with boilerplate code I already have, otherwise I'll cook something up in a no-code tool. If you have a niche where one type of solution is commonly demanded, it helps to have a general demo set up to be able to handle a larger volume of calls, so you aren't burning yourself out. I'll also elaborate on how the final product would look like in comparison to the demo. Call #3 and Beyond: Once the initial consultation and demo is complete, you will want to alleviate any remaining concerns from your prospects and work with them to reach a final work proposal. It's crucial you lay out exactly what you will be building (in writing) and ensure the prospect understands this. Furthermore, be clear and transparent with timelines and communication methods for the project. In terms of pricing, you want to take this from a value-based approach. The same solution may be worth a lot more to client A than client B. Furthermore, you can create "add-ons" such as monthly maintenance/upgrade packages, training sessions for employeees, and so forth, separate from the initial setup fee you would charge. How you can incorporate AI into marketing your businesses Beyond cold sales, I highly recommend creating a funnel to capture warm leads. For instance, I do this currently with my AI tools directory, which links directly to my AI agency and has consistent branding throughout. Warm leads are much more likely to close (and honestly, much nicer to deal with). However, even without an AI-related website, at the very least you will want to create a presence on social media and the web in general. As with any agency, you will want basic a professional presence. A professional virtual address helps, in addition to a Google Business Profile (GBP) and TrustPilot. a GBP (especially for local SEO) and Trustpilot page also helps improve the looks of your search results immensely. For GBP, I recommend using ProfilePro, which is a chrome extension you can use to automate SEO work for your GBP. Aside from SEO optimzied business descriptions based on your business, it can handle Q/A answers, responses, updates, and service descriptions based on local keywords. Privacy and Legal Concerns of the AAA Model Aside from typical concerns for agencies relating to service contracts, there are a few issues (especially when using no-code tools) that will need to be addressed to run a successful AAA. Most of these surround privacy concerns when working with proprietary data. In your terms with your client, you will want to clearly define hosting providers and any third party tools you will be using to build their solution, and a DPA with these third parties listed as subprocessors if necessary. In addition, you will want to implement best practices like redacting private information from data being used for building solutions. In terms of addressing concerns directly from clients, it helps if you host your solutions on their own servers (not possible with AI tools), and address the fact only ChatGPT queries in the web app, not OpenAI API calls, will be used to train OpenAI's models (as reported by mainstream media). The key here is to be open and transparent with your clients about ALL the tools you are using, where there data will be going, and make sure to get this all in writing. have fun, and keep an open mind Before I finish this post, I just want to reiterate the fact that this is NOT an easy way to make money. Running an AI agency will require hours and hours of dedication and work, and constantly rearranging your schedule to meet prospect and client needs. However, if you are looking for a new business to run, and have a knack for understanding business operations and are genuinely interested in the pracitcal applications of generative AI, then I say go for it. The time is ticking before AAA becomes the new dropshipping or SMMA, and I've a firm believer that those who set foot first and establish themselves in this field will come out top. And remember, while 100 thousand people may read this post, only 2 may actually take initiative and start.

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security

Uzair Javaid, a Ph.D. with a passion for data privacy, co-founded Betterdata to tackle one of AI's most pressing challenges: protecting privacy while enabling innovation. Recently, Betterdata secured a lucrative contract with the US Department of Homeland Security, 1 of only 4 companies worldwide to do so and the only one in Asia. Here's how he did it: The Story So what's your story? I grew up in Peshawar, Pakistan, excelling in coding despite studying electrical engineering. Inspired by my professors, I set my sights on studying abroad and eventually earned a Ph.D. scholarship at NUS Singapore, specializing in data security and privacy. During my research, I ethically hacked Ethereum and published 15 papers—three times the requirement. While wrapping up my Ph.D., I explored startup ideas and joined Entrepreneur First, where I met Kevin Yee. With his expertise in generative models and mine in privacy, we founded Betterdata. Now, nearly three years in, we’ve secured a major contract with the U.S. Department of Homeland Security—one of only four companies globally and the only one from Asia. The Startup In a nutshell, what does your startup do? Betterdata is a startup that uses AI and synthetic data generation to address two major challenges: data privacy and the scarcity of high-quality data for training AI models. By leveraging generative models and privacy-enhancing technologies, Betterdata enables businesses, such as banks, to use customer data without breaching privacy regulations. The platform trains AI on real data, learns its patterns, and generates synthetic data that mimics the real thing without containing any personal or sensitive information. This allows companies to innovate and develop AI solutions safely and ethically, all while tackling the growing need for diverse, high-quality data in AI development. How did you conduct ideation and validation for your startup? The initial idea for Betterdata came from personal experience. During my Ph.D., I ethically hacked Ethereum’s blockchain, exposing flaws in encryption-based data sharing. This led me to explore AI-driven deep synthesis technology—similar to deepfakes but for structured data privacy. With GDPR impacting 28M+ businesses, I saw a massive opportunity to help enterprises securely share data while staying compliant. To validate the idea, I spoke to 50 potential customers—a number that strikes the right balance. Some say 100, but that’s impractical for early-stage founders. At 50, patterns emerge: if 3 out of 10 mention the same problem, and this repeats across 50, you have 10–15 strong signals, making it a solid foundation for an MVP. Instead of outbound sales, which I dislike, we used three key methods: Account-Based Marketing (ABM)—targeting technically savvy users with solutions for niche problems, like scaling synthetic data for banks. Targeted Content Marketing—regular customer conversations shaped our thought leadership and outreach. Raising Awareness Through Partnerships—collaborating with NUS, Singapore’s PDPC, and Plug and Play to build credibility and educate the market. These strategies attracted serious customers willing to pay, guiding Betterdata’s product development and market fit. How did you approach the initial building and ongoing product development? In the early stages, we built synthetic data generation algorithms and a basic UI for proof-of-concept, using open-source datasets to engage with banks. We quickly learned that banks wouldn't share actual customer data due to privacy concerns, so we had to conduct on-site installations and gather feedback to refine our MVP. Through continuous consultation with customers, we discovered real enterprise data posed challenges, such as missing values, which led us to adapt our prototype accordingly. This iterative approach of listening to customer feedback and observing their usage allowed us to improve our product, enhance UX, and address unmet needs while building trust and loyalty. Working closely with our customers also gives us a data advantage. Our solution’s effectiveness depends on customer data, which we can't fully access, but bridging this knowledge gap gives us a competitive edge. The more customers we test on, the more our algorithms adapt to diverse use cases, making it harder for competitors to replicate our insights. My approach to iteration is simple: focus solely on customer feedback and ignore external noise like trends or advice. The key question for the team is: which customer is asking for this feature or solution? As long as there's a clear answer, we move forward. External influences, such as AI hype, often bring more confusion than clarity. True long-term success comes from solving real customer problems, not chasing trends. Customers may not always know exactly what they want, but they understand their problems. Our job is to identify these problems and solve them in innovative ways. While customers may suggest specific features, we stay focused on solving the core issue rather than just fulfilling their exact requests. The idea aligns with the quote often attributed to Henry Ford: "If I asked people what they wanted, they would have said faster horses." The key is understanding their problems, not just taking requests at face value. How do you assess product-market fit? To assess product-market fit, we track two key metrics: Customers' Willingness to Pay: We measure both the quantity and quality of meetings with potential customers. A high number of meetings with key decision-makers signals genuine interest. At Betterdata, we focused on getting meetings with people in banks and large enterprises to gauge our product's resonance with the target market. How Much Customers Are Willing to Pay: We monitor the price customers are willing to pay, especially in the early stages. For us, large enterprises, like banks, were willing to pay a premium for our synthetic data platform due to the growing need for privacy tech. This feedback guided our product refinement and scaling strategy. By focusing on these metrics, we refined our product and positioned it for scaling. What is your business model? We employ a structured, phase-driven approach for out business model, as a B2B startup. I initially struggled with focusing on the core value proposition in sales, often becoming overly educational. Eventually, we developed a product roadmap with models that allowed us to match customer needs to specific offerings and justify our pricing. Our pricing structure includes project-based pilots and annual contracts for successful deployments. At Betterdata, our customer engagement unfolds across three phases: Phase 1: Trial and Benchmarking \- We start with outreach and use open-source datasets to showcase results, offering customers a trial period to evaluate the solution. Phase 2: Pilot or PoC \- After positive trial results, we conduct a PoC or pilot using the customer’s private data, with the understanding that successful pilots lead to an annual contract. Phase 3: Multi-Year Contracts \- Following a successful pilot, we transition to long-term commercial contracts, focusing on multi-year agreements to ensure stability and ongoing partnerships. How do you do marketing for your brand? We take a non-conventional approach to marketing, focusing on answering one key question: Which customers are willing to pay, and how much? This drives our messaging to show how our solution meets their needs. Our strategy centers around two main components: Building a network of lead magnets \- These are influential figures like senior advisors, thought leaders, and strategic partners. Engaging with institutions like IMDA, SUTD, and investors like Plug and Play helps us gain access to the right people and foster warm introductions, which shorten our sales cycle and ensure we’re reaching the right audience. Thought leadership \- We build our brand through customer traction, technology evidence, and regulatory guidelines. This helps us establish credibility in the market and position ourselves as trusted leaders in our field. This holistic approach has enabled us to navigate diverse market conditions in Asia and grow our B2B relationships. By focusing on these areas, we drive business growth and establish strong trust with stakeholders. What's your advice for fundraising? Here are my key takeaways for other founders when it comes to fundraising: Fundraise When You Don’t Need To We closed our seed round in April 2023, a time when we weren't actively raising. Founders should always be in fundraising mode, even when they're not immediately in need of capital. Don’t wait until you have only a few months of runway left. Keep the pipeline open and build relationships. When the timing is right, execution becomes much easier. For us, our investment came through a combination of referrals and inbound interest. Even our lead investor initially rejected us, but after re-engaging, things eventually fell into place. It’s crucial to stay humble, treat everyone with respect, and maintain those relationships for when the time is right. Be Mindful of How You Present Information When fundraising, how you present information matters a lot. We created a comprehensive, easily digestible investment memo, hosted on Notion, which included everything an investor might need—problem, solution, market, team, risks, opportunities, and data. The goal was for investors to be able to get the full picture within 30 minutes without chasing down extra details. We also focused on making our financial model clear and meaningful, even though a 5-year forecast might be overkill at the seed stage. The key was clarity and conciseness, and making it as easy as possible for investors to understand the opportunity. I learned that brevity and simplicity are often the best ways to make a memorable impact. For the pitch itself, keep it simple and focus on 4 things: problem, solution, team, and market. If you can summarize each of these clearly and concisely, you’ll have a compelling pitch. Later on, you can expand into market segments, traction, and other metrics, but for seed-stage, focus on those four areas, and make sure you’re strong in at least three of them. If you do, you'll have a compelling case. How do you run things day-to-day? i.e what's your operational workflow and team structure? Here's an overview of our team structure and process: Internally: Our team is divided into two main areas: backend (internal team) and frontend (market-facing team). There's no formal hierarchy within the backend team. We all operate as equals, defining our goals based on what needs to be developed, assigning tasks, and meeting weekly to share updates and review progress. The focus is on full ownership of tasks and accountability for getting things done. I also contribute to product development, identifying challenges and clearing obstacles to help the team move forward. Backend Team: We approach tasks based on the scope defined by customers, with no blame or hierarchy. It's like a sports team—sometimes someone excels, and other times they struggle, but we support each other and move forward together. Everyone has the creative freedom to work in the way that suits them best, but we establish regular meetings and check-ins to ensure alignment and progress. Frontend Team: For the market-facing side, we implement a hierarchy because the market expects this structure. If I present myself as "CEO," it signals authority and credibility. This distinction affects how we communicate with the market and how we build our brand. The frontend team is split into four main areas: Business Product (Software Engineering) Machine Learning Engineering R&D The C-suite sits at the top, followed by team leads, and then the executors. We distill market expectations into actionable tasks, ensuring that everyone is clear on their role and responsibilities. Process: We start by receiving market expectations and defining tasks based on them. Tasks are assigned to relevant teams, and execution happens with no communication barriers between team members. This ensures seamless collaboration and focused execution. The main goal is always effectiveness—getting things done efficiently while maintaining flexibility in how individuals approach their work. In both teams, there's an emphasis on accountability, collaboration, and clear communication, but the structure varies according to the nature of the work and external expectations.

how I built a $6k/mo business with cold email
reddit
LLM Vibe Score0
Human Vibe Score1
Afraid-Astronomer130This week

how I built a $6k/mo business with cold email

I scaled my SaaS to a $6k/mo business in under 6 months completely using cold email. However, the biggest takeaway for me is not a business that’s potentially worth 6-figure. It’s having a glance at the power of cold emails in the age of AI. It’s a rapidly evolving yet highly-effective channel, but no one talks about how to do it properly. Below is the what I needed 3 years ago, when I was stuck with 40 free users on my first app. An app I spent 2 years building into the void. Entrepreneurship is lonely. Especially when you are just starting out. Launching a startup feel like shouting into the dark. You pour your heart out. You think you have the next big idea, but no one cares. You write tweets, write blogs, build features, add tests. You talk to some lukewarm leads on Twitter. You do your big launch on Product Hunt. You might even get your first few sales. But after that, crickets... Then, you try every distribution channel out there. SEO Influencers Facebook ads Affiliates Newsletters Social media PPC Tiktok Press releases The reality is, none of them are that effective for early-stage startups. Because, let's face it, when you're just getting started, you have no clue what your customers truly desire. Without understanding their needs, you cannot create a product that resonates with them. It's as simple as that. So what’s the best distribution channel when you are doing a cold start? Cold emails. I know what you're thinking, but give me 10 seconds to change your mind: When I first heard about cold emailing I was like: “Hell no! I’m a developer, ain’t no way I’m talking to strangers.” That all changed on Jan 1st 2024, when I actually started sending cold emails to grow. Over the period of 6 months, I got over 1,700 users to sign up for my SaaS and grew it to a $6k/mo rapidly growing business. All from cold emails. Mastering Cold Emails = Your Superpower I might not recommend cold emails 3 years ago, but in 2024, I'd go all in with it. It used to be an expensive marketing channel bootstrapped startups can’t afford. You need to hire many assistants, build a list, research the leads, find emails, manage the mailboxes, email the leads, reply to emails, do meetings. follow up, get rejected... You had to hire at least 5 people just to get the ball rolling. The problem? Managing people sucks, and it doesn’t scale. That all changed with AI. Today, GPT-4 outperforms most human assistants. You can build an army of intelligent agents to help you complete tasks that’d previously be impossible without human input. Things that’d take a team of 10 assistants a week can now be done in 30 minutes with AI, at far superior quality with less headaches. You can throw 5000 names with website url at this pipeline and you’ll automatically have 5000 personalized emails ready to fire in 30 minutes. How amazing is that? Beyond being extremely accessible to developers who are already proficient in AI, cold email's got 3 superpowers that no other distribution channels can offer. Superpower 1/3 : You start a conversation with every single user. Every. Single. User. Let that sink in. This is incredibly powerful in the early stages, as it helps you establish rapport, bounce ideas off one another, offer 1:1 support, understand their needs, build personal relationships, and ultimately convert users into long-term fans of your product. From talking to 1000 users at the early stage, I had 20 users asking me to get on a call every week. If they are ready to buy, I do a sales call. If they are not sure, I do a user research call. At one point I even had to limit the number of calls I took to avoid burnout. The depth of the understanding of my customers’ needs is unparalleled. Using this insight, I refined the product to precisely cater to their requirements. Superpower 2/3 : You choose exactly who you talk to Unlike other distribution channels where you at best pick what someone's searching for, with cold emails, you have 100% control over who you talk to. Their company Job title Seniority level Number of employees Technology stack Growth rate Funding stage Product offerings Competitive landscape Social activity (Marital status - well, technically you can, but maybe not this one…) You can dial in this targeting to match your ICP exactly. The result is super low CAC and ultra high conversion rate. For example, My competitors are paying $10 per click for the keyword "HARO agency". I pay $0.19 per email sent, and $1.92 per signup At around $500 LTV, you can see how the first means a non-viable business. And the second means a cash-generating engine. Superpower 3/3 : Complete stealth mode Unlike other channels where competitors can easily reverse engineer or even abuse your marketing strategies, cold email operates in complete stealth mode. Every aspect is concealed from end to end: Your target audience Lead generation methods Number of leads targeted Email content Sales funnel This secrecy explains why there isn't much discussion about it online. Everyone is too focused on keeping their strategies close and reaping the rewards. That's precisely why I've chosen to share my insights on leveraging cold email to grow a successful SaaS business. More founders need to harness this channel to its fullest potential. In addition, I've more or less reached every user within my Total Addressable Market (TAM). So, if any competitor is reading this, don't bother trying to replicate it. The majority of potential users for this AI product are already onboard. To recap, the three superpowers of cold emails: You start a conversation with every single user → Accelerate to PMF You choose exactly who you talk to → Super-low CAC Complete stealth mode → Doesn’t attract competition By combining the three superpowers I helped my SaaS reach product-marketing-fit quickly and scale it to $6k per month while staying fully bootstrapped. I don't believe this was a coincidence. It's a replicable strategy for any startup. The blueprint is actually straightforward: Engage with a handful of customers Validate the idea Engage with numerous customers Scale to $5k/mo and beyond More early-stage founders should leverage cold emails for validation, and as their first distribution channel. And what would it do for you? Update: lots of DM asking about more specifics so I wrote about it here. https://coldstartblueprint.com/p/ai-agent-email-list-building

What's some good AI software for entrepreneurs?
reddit
LLM Vibe Score0
Human Vibe Score1
Moist_Possibility128This week

What's some good AI software for entrepreneurs?

I just started running a smaller business as a side gig and am in need of getting some manual work off my shoulders. This business is basically a hobby turned business as something I've been wanting to get into for a long time but just got the courage to do so this year. I'm making hand-made jewelry that's kind of a niche but has a tiny little tight market with relatively active and supportive buyers. Of course, a huge part of my job is answering all kinds of questions, covering spreadsheets, and doing market research to try and find new customer groups. The majority of this work is relatively simple what I’d call “manual”, which is why I feel like it could be done by AI, at the very least with the precision that I need. I did find some help using Chat GPT 4 so far, especially with handling my spreadsheets and market research. I usually let it do some manual labor on the spreadsheets, and I’ve even managed to train it to do some more complex tasks like researching the market and putting the results in the spreadsheet that I can use. ChatGPT isn’t that good at answering messages however because the answers are pretty generic and I have to manually generate responses and send them which takes arguably even more time than just responding myself. For this task, Personal AI has been proven to be way more useful because it’s literally a personalized AI model that can be trained to accurately respond to anything + once you create your own personal AI, other people can ask questions there instead of messaging me directly and get instant responses from the AI that are based on the knowledge I fed it. Still testing the tool, but so far it has been quite useful and saved me a ton of time. I also used Poll the People a few times to get feedback from my customers, and it worked magnificently. I'd like to hear some recommendations on AI tools that can be useful to someone who's just entering this world so please shoot them!

I have reviewed over 900+ AI Tools for my directory. Here are some of the best ones I have seen for entrepreneurs and startups.
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

I have reviewed over 900+ AI Tools for my directory. Here are some of the best ones I have seen for entrepreneurs and startups.

As one of the co-founders at AI Scout, a platform for AI discovery, I've had the privilege (and challenge) of reviewing over 900 AI tools submitted to our directory. I've filtered these down to some of the top AI tools that I believe could bring value to startups and entrepreneurs. It's worth noting that while these tools are great right out the box, the power of AI is truly realized when these tools are used in tandem and strategically aligned with your business needs. The challenge most people face is not about the lack of AI tools available, but the difficulty in finding the right one that fits their specific needs and workflows. Without further ado, here's my top pick of AI tools you should consider looking into if you are an entrepreneur or run a startup. Chatbase - Custom ChatGPT (Trained on Your Own Data) Taking a step up from traditional support bots, Chatbase combines the power of GPT and your own knowledge base. The result is a ChatGPT-like chatbot that is trained on your own websites and documents. You can embed the chatbot into your own website via an iframe or script in the header of your website code. They also have an API you can take advantage of. We use this personally at AI Scout for ScoutBud (AI assistant to find AI tools), which we trained based on our directory site. It would also work great if you have extensive documentation, papers, etc. that you want to quickly reference by simply asking a chatbot for the info you need instead of having to go through dozens of PDFs. Reply - AI-Powered Sales Engagement Platform Great AI tool to manage your entire sales engagement cycle. They have a large database with about a dozen filters to discover optimal B2B leads. From here, you can use their GPT integration to generate cold emails as well as handle responses and meeting scheduling. What I like personally about Reply are the endless integrations available, including Gmail, Outlook, Zoho, and major social platforms such as Twitter and LinkedIn. Instapage - AI Landing Page Generation, Testing, and Personalization This AI tool allows users to generate content variations for landing pages including headlines, paragraphs, and CTAs based on the target audience. You can also conduct A/B testing for more effective and efficient campaigns. Paired with hundreds of professional and cutomizable layouts, Instapage is definitely something I would recommend for entrepreneurs who want to get a high-converting landing page set up quickly and effectively. SaneBox - AI Emails Management If you feel overwhelmed by the sheer volume of emails you receive like myself and many entrepreneurs, this could be something for you. SaneBox’s AI identifies important emails and declutters your inbox, helping you to stay focused on what truly matters. SocialBee - AI Social Media Manager Think of SocialBee as your all-in-one social media command center, powered by AI. You can manage multiple social media accounts from one platform and generate captions with AI as well. SocialBee not only allows you to schedule posts but also helps you analyze growth and engagement with detailed reports. Works well with all social media platforms, including Facebook, Twitter, Instagram, and Linkedin. I believe they also have integrations for TikTok and YouTube, although I haven't tried these personally. MeetGeek - AI Meeting Assistant Lifesaver if you attend a lot of meetings or calls. Great for transcribing, summarizing, and sharing key insights from meetings. The AI also creates meeting highlights, which I've personally fouund quite useful if you ever need to get a very quick and dirty overview of what happened in a call. It also provides analysis (including sentiment evaluation) for meetings. Taskade - AI Productivity Tool for Task Management An all-in-one AI productivity tool. Multiple AI features available, including a chatbot, writing assistant, and workflow creator. It's a great all-around tool for real-time collaboration and efficient task management. Scribe AI (ScribeHow) - AI Documentation Generator Great for any SaaS applications where you need to create resources/documentations/guides for your app. You simply record your process and Scribe generates a written guide for you. Remember, while AI is an excellent assistant, it's also just a tool. The ultimate success of your venture depends on how effectively you leverage these tools. Happy experimenting!

Detailed Guide - How I've Been Self Employed for 2 Years Selling Posters
reddit
LLM Vibe Score0
Human Vibe Score1
tommo278This week

Detailed Guide - How I've Been Self Employed for 2 Years Selling Posters

Hey everyone, bit of context before you read through this. I have been selling POD posters full time for over 2 years now. My next venture is that I have started my own Print on Demand company for posters, PrintShrimp. As one way of creating customers for our service, we are teaching people for free how to also sell posters. Here is a guide I have written on how to sell posters on Etsy. Feel free to have a read through and then check out PrintShrimp, hopefully can help some of you guys out (and get us some more customers!) All of this is also available in video format on our website too, if you prefer to learn that way. Thanks guys! And as some people asked in other subs, no this isn't written with AI 😅 This took a couple of weeks to put together! Through this guide, we will teach you everything you need to know about starting to sell posters and generate some income. We will also show you why PrintShrimp is the best POD supplier for all of your poster needs. Trust me, you won’t need much convincing.  So, why are posters the best product to sell? Also, just thought I’d quickly answer the question - why posters? If you’ve been researching Print on Demand you’ve probably come across the infinite options of t-shirts, mugs, hats, phone cases, and more. All of these are viable options, however we think posters are the perfect place to start. You can always expand into other areas further down the line! So a brief summary of why posters are the perfect product for Print on Demand: \-They are very easy to design! Posters are a very easy shape to deal with - can’t go wrong with a rectangle. This makes designing products very easy. \-Similarly to this, what you see is what you get with a poster. You can literally see your finished product as you design it in either canva or photoshop. With T-Shirts for example, you have to make your design, and then place it on a t-shirt. Then you have to coordinate with your printers the size you would like the design on the tshirt and many other variables like that. There is no messing about with posters - what you see is what you get. \-The same high quality, everywhere. With other products, if you want to reap the benefits of a printing in various countries, you need to ensure each of your global suppliers stocks the same t-shirts, is able to print in the same way, carries the same sizes etc. Again with posters you avoid all of this hassle- your products will come out the same, no matter which of our global locations are used. \-They have a very favorable profit margin. As you will see later, the cost price of posters is very low. And people are prepared to pay quite a lot for a decent bit of wall art! I have tried out other products, and the profit margin combined with the order quantity of posters makes them my most profitable product, every single time. Using PrintShrimp, you can be sure to enjoy profits of anywhere between £6 - £40 pure profit per sale.  \-They are one of the easiest to print white label. This makes them perfect for Print on Demand. Your posters are simply put in a tube, and off they go. There are no extras you need to faff around with, compared to the extra elements other products come with, such as clothing labels on t-shirts.  Picking your poster niche So, you are ready to start selling posters. Great! Now, the blessing and curse with selling posters is that there are infinite possibilities regarding what you can sell. So, it can easily be quite overwhelming at first.  The first thing I would recommend doing is having a look at what others are selling. Etsy is a wonderful place for this (and will likely be a key part of your poster selling journey). So, log on to Etsy and simply type in ‘poster’ in the search bar. Get ready to write a massive list of the broad categories and type of posters that people are selling.  If you do not have more than 50 categories written down by the end, you are doing something wrong. There are seriously an infinite amount of posters! For example, here are some popular ones to get you started: Star sign posters, Kitchen posters, World map posters, Custom Dog Portrait posters, Music posters, Movie posters, Fine art posters, Skiing posters, Girl Power posters and Football posters.  Now, you have a huge list of potential products to sell. What next? There are a few important things you need to bear in mind when picking your niche: \-Does this interest me?  Don’t make the mistake of going down a niche that didn’t actually interest you just because it would probably be a money maker. Before you know it, what can be a very fun process of making designs can become incredibly \\\monotonous, and feel like a chore\\\. You need to bear in mind that you will be spending a lot of time creating designs - if it is something you are interested in you are much less likely to get burnt out! As well, \\\creativity will flow\\\ far better if it is something you are interested in, which at the end of the day will lead to better designs that are more likely to be purchased by customers.  \-Is this within my design range? Don’t let this put you off too much. We will go through how to get started on design later on in this guide. However, it is important to note that the plain truth of it is that some niches and designs are a hell of a lot more complicated than others. For example, quote posters can essentially be designed by anyone when you learn about how to put nice fonts together in a good color scheme. On the other hand, some posters you see may have been designed with complex illustrations in a program like Illustrator. To start with, it may be better to pick a niche that seems a bit more simple to get into, as you can always expand your range with other stores further down the line. A good way of evaluating the design complexity is by identifying if this poster is \\\a lot of elements put together\\\ or is \\\a lot of elements created by the designer themselves\\\\\.\\ Design can in a lot of cases be like a jigsaw - putting colours, shapes and text together to create an image. This will be a lot easier to start with and can be learnt by anyone, compared to complex drawings and illustrations.  \-Is this niche subject to copyright issues? Time to delve deep into good old copyright. Now, when you go through Etsy, you will without a doubt see hundreds of sellers selling music album posters, car posters, movie posters and more. Obviously, these posters contain the property of musicians, companies and more and are therefore copyrighted. The annoying thing is - these are \\\a complete cash cow.\\\ If you go down the music poster route, I will honestly be surprised if you \\don’t\\ make thousands. However it is only a matter of time before the copyright strikes start rolling in and you eventually get banned from Etsy.  So I would highly recommend \\\not making this mistake\\\. Etsy is an incredible platform for selling posters, and it is a hell of a lot easier to make sales on there compared to advertising your own website. And, you \\\only get one chance on Etsy.\\\ Once you have been banned once, you are not allowed to sign up again (and they do ID checks - so you won’t be able to rejoin again under your own name).  So, don’t be shortsighted when it comes to entering Print on Demand. If you keep your designs legitimate, they will last you a lifetime and you will then later be able to crosspost them to other platforms, again without the worry of ever getting shut down.  So, how do I actually design posters? Now you have an idea of what kind of posters you want to be making, it’s time to get creative and make some designs! Photoshop (and the creative cloud in general) is probably the best for this. However, when starting out it can be a scary investment (it costs about £30 a month unless you can get a student rate!).  So, while Photoshop is preferable in the long term, when starting out you can learn the ropes of design and get going with Canva. This can be great at the start as they have a load of templates that you can use to get used to designing and experimenting (while it might be tempting to slightly modify these and sell them - this will be quite saturated on places like Etsy so we would recommend doing something new).  What size format should I use? The best design format to start with is arguably the A sizes - as all the A sizes (A5, A4, A3, A2, A1, A0) are scalable. This means that you can make all of your designs in one size, for example A3, and these designs will be ready to fit to all other A sizes. For example, if you design an A3 poster and someone orders A1, you can just upload this A3 file to PrintShrimp and it will be ready to print. There is a wide range of other sizes you should consider offering on your shop, especially as these sizes are very popular with the American market. They have a wide range of popular options, which unfortunately aren’t all scalable with each other. This does mean that you will therefore have to make some slight modifications to your design in order to be able to offer them in American sizing, in a few different aspect ratios. What you can do however is design all of your products in UK sizing, and simply redesign to fit American sizing once you have had an order. Essentially: design in UK sizing, but list in both UK and US sizing. Then when you get a non-A size order, you can quickly redesign it on demand. This means that you don’t have to make a few different versions of each poster when first designing, and can simply do a quick redesign for US sizing when you need to. Below is PrintShrimps standard size offering. We can also offer any custom sizing too, so please get in touch if you are looking for anything else. With these sizes, your poster orders will be dispatched domestically in whatever country your customer orders from. Our recommendations for starting design One thing that will not be featured in this guide is a written out explanation or guide on how to design. Honestly, I can’t think of a more boring, or frankly worse, way to learn design. When it comes to getting started, experimenting is your best friend! Just have a play around and see what you can do. It is a really fun thing to get started with, and the satisfaction of when a poster design comes together is like no other. A good way to start is honestly by straight up copying a poster you see for sale online. And we don’t mean copying to sell! But just trying to replicate other designs is a great way to get a feel for it and what you can do. We really think you will be surprised at how easy it is to pull together a lot of designs that at first can appear quite complicated! Your best friend throughout this whole process will be google. At the start you will not really know how to do anything - but learning how to look into things you want to know about design is all part of the process. At first, it can be quite hard to even know how to search for what you are trying to do, but this will come with time (we promise). Learning how to google is a skill that you will learn throughout this process.  Above all, what we think is most important is this golden rule: take inspiration but do not steal. You want to be selling similar products in your niche, but not copies. You need to see what is selling in your niche and get ideas from that, but if you make designs too similar to ones already available, you won’t have much luck. At the end of the day, if two very similar posters are for sale and one shop has 1000 reviews and your newer one has 2, which one is the customer going to buy? You need to make yours offer something different and stand out enough to attract customers. Etsy SEO and maximizing your sales You may have noticed in this guide we have mentioned Etsy quite a few times! That is because we think it is hands down the best place to start selling posters. Why? Etsy is a go to place for many looking to decorate their homes and also to buy gifts. It might be tempting to start selling with your own website straight away, however we recommend Etsy as it brings the customers to you. For example, say you start selling Bathroom Posters. It is going to be a hell of a lot easier to convert sales when you already have customers being shown your page after searching ‘bathroom decor’, compared to advertising your own website. This is especially true as it can be hard to identify your ideal target audience to then advertise to via Meta (Facebook/Instagram) for example. Websites are a great avenue to explore eventually like I now have, but we recommend starting with Etsy and going from there. What costs do I need to be aware of? So, setting up an Etsy sellers account is currently costs £15. The only other upfront cost you will have is the cost of listing a product - this is 20 cents per listing. From then on, every time you make a sale you will be charged a transaction fee of 6.5%, a small payment processing fee, plus another 20 cents for a renewed listing fee. It normally works out to about 10% of each order, a small price to pay for all the benefits Etsy brings. No matter what platform you sell on, you will be faced with some form of transaction fee. Etsy is actually quite reasonable especially as they do not charge you to use their platform on a monthly basis.  What do I need to get selling? Getting your shop looking pretty \-Think of a shop name and design (now you are a professional designer) a logo \-Design a banner for the top of your shop \-Add in some about me info/shop announcement \-I recommend running a sale wherein orders of 3+ items get a 20% of discount. Another big benefit of PrintShrimp is that you receive large discounts when ordering multiple posters. This is great for attracting buyers and larger orders.  Making your products look attractive That is the bulk of the ‘decor’ you will need to do. Next up is placing your posters in mock ups! As you may notice on Etsy, most shops show their posters framed and hanging on walls. These are 99% of the time not real photos, but digital mock ups. This is where Photoshop comes in really handy, as you can automate this process through a plug in called Bulk Mock Up. If you don’t have photoshop, you can do this on Canva, you will just have to do it manually which can be rather time consuming.  Now, where can you get the actual Mock Ups? One platform we highly recommend for design in general is platforms like Envato Elements. These are design marketplaces where you have access to millions of design resources that you are fully licensed to use!  Titles, tags, and descriptions  Now for the slightly more nitty gritty part. You could have the world's most amazing looking poster, however, if you do not get the Etsy SEO right, no one is going to see it! We will take you through creating a new Etsy listing field by field so you can know how to best list your products.  The key to Etsy listing optimisation is to maximise. Literally cram in as many key words as you possibly can! Before you start this process, create a word map of anything you can think of relating to your listing. And come at this from the point of view of, if I was looking for a poster like mine, what would I search? Titles \-Here you are blessed with 140 characters to title your listing. Essentially, start off with a concise way of properly describing your poster. And then afterwards, add in as many key words as you can! Here is an example of the title of a well selling Skiing poster: Les Arcs Skiing Poster, Les Arcs Print, Les Alpes, France Ski Poster, Skiing Poster, Snowboarding Poster, Ski Resort Poster Holiday, French This is 139 characters out of 140 - you should try and maximise this as much as possible! As you can see, this crams in a lot of key words and search terms both related to Skiing as a whole, the poster category, and then the specifics of the poster itself (Les Arcs resort in France). Bear in mind that if you are listing a lot of listings that are of the same theme, you won’t have to spend time creating an entirely new title. For example if your next poster was of a ski resort in Italy, you can copy this one over and just swap out the specifics. For example change “France ski poster” to “Italy ski poster”, change “Les Arcs” to “The Dolomites”, etc.  Description \-Same logic applies for descriptions - try and cram in as many key words as you can! Here is an example for a Formula One poster: George Russell, Mercedes Formula One Poster  - item specific keywords Bright, modern and vibrant poster to liven up your home.  - Describes the style of the poster All posters are printed on high quality, museum grade 200gsm poster paper. Suitable for framing and frames. - Shows the quality of the print. Mentions frames whilst showing it comes unframed Experience the thrill of the racetrack with this stunning Formula One poster. Printed on high-quality paper, this racing car wall art print features a dynamic image of a Formula One car in action, perfect for adding a touch of speed and excitement to any motorsports room or man cave. Whether you're a die-hard fan or simply appreciate the adrenaline of high-speed racing, this poster is sure to impress. Available in a range of sizes, it makes a great addition to your home or office, or as a gift for a fellow Formula One enthusiast. Each poster is carefully packaged to ensure safe delivery, so you can enjoy your new piece of art as soon as possible. - A nice bit of text really highlighting a lot of key words such as gift, motorsports, racetrack etc.  You could go further with this too, by adding in extra things related to the poster such as ‘Perfect gift for a Mercedes F1 fan’ etc.  Tags Now, these are actually probably the most important part of your listing! You get 13 tags (20 character limit for each) and there are essentially search terms that will match your listing with what customers search for when shopping.  You really need to maximize these - whilst Title and Description play a part, these are the main things that will bring buyers to your listing. Once again, it is important to think about what customers are likely to be searching when looking for a poster similar to yours. Life hack alert! You can actually see what tags other sellers are using. All you need to do is go to a listing similar to yours that is selling well, scroll down and you can actually see them listed out at the bottom of the page! Here is an example of what this may look like: So, go through a few listings of competitors and make notes on common denominators that you can integrate into your listing. As you can see here, this seller uses tags such as ‘Birthday Gift’ and ‘Poster Print’. When you first start out, you may be better off swapping these out for more listing specific tags. This seller has been on Etsy for a few years however and has 15,000+ sales, so are more likely to see success from these tags.  If it’s not clear why, think about it this way. If you searched ‘poster print’ on Etsy today, there will be 10s of thousands of results. However, if you searched ‘Russell Mercedes Poster’, you will (as of writing) get 336 results. Etsy is far more likely to push your product to the top of the latter tag, against 300 other listings, rather than the top of ‘Poster Print’ where it is incredibly competitive. It is only when you are a more successful shop pulling in a high quantity of orders that these larger and more generic tags will work for you, as Etsy has more trust in your shop and will be more likely to push you to the front.  SKUs \-One important thing you need to do is add SKUs to all of your products! This is worth doing at the start as it will make your life so much easier when it comes to making sales and using PrintShrimp further down the line. What is an SKU? It is a ‘stock keeping unit’, and is essentially just a product identifier. Your SKUs need to match your file name that you upload to PrintShrimp. For example, if you made a poster about the eiffel tower, you can literally name the SKU eiffel-tower. There is no need to complicate things! As long as your file name (as in the image name of your poster on your computer) matches your SKU, you will be good to go.  \-It may be more beneficial to set up a system with unique identifiers, to make organising your files a lot easier further down the line. Say you get to 1000 posters eventually, you’ll want to be able to quickly search a code, and also ensure every SKU is always unique, so you won’t run into accidentally using the same SKU twice further down the line. For example, you can set it up so at the start of each file name, you have \[unique id\]\[info\], so your files will look like -  A1eiffeltower A2france And further down the line: A99aperolspritz B1potatoart This not only removes the potential issue of duplicating SKUs accidentally (for example if you made a few posters of the same subject), but also keeps your files well organised. If you need to find a file, you can search your files according to the code, so just by searching ‘a1’ for example, rather than having to trawl through a load of different files until you find the correct one. \-If your poster has variations, for example color variations, you can set a different SKU for each variation. Just click the little box when setting up variations that says ‘SKUs vary for each (variation)’. So if you have a poster available either in a white or black background, you can name each file, and therefore each SKU, a1eiffel-tower-black and a1eiffel-tower-white for example. \-The same goes for different sizes. As different American sizes have different aspect ratios, as mentioned above you may have to reformat some posters if you get a sale for one of these sizes. You can then add in the SKU to your listing once you have reformatted your poster. So for example if you sell a 16x20” version of the eiffel tower poster, you can name this file eiffel-tower-white-1620. Whilst this involves a little bit of set up, the time it saves you overall is massive!  Variations and Prices \-So, when selling posters there is a huge variety of sizes that you can offer, as mentioned previously. Non-negotiable is that you should be offering A5-A1. These will likely be your main sellers! Especially in the UK. It is also a good idea to offer inch sizing to appeal to a global audience (as bear in mind with PrintShrimp you will be able to print in multiple countries around the world!).  Below is a recommended pricing structure of what to charge on Etsy. Feel free to mess around with these! You may notice on Etsy that many shops charge a whole lot more for sizes such as A1, 24x36” etc. In my experience I prefer charging a lower rate to attract more sales, but there is validity in going for a lower amount of sales with higher profits. As mentioned above, you can also offer different variations on items - for example different colour schemes on posters. This is always a decent idea (if it suits the design) as it provides the customer with more options, which might help to convert the sale. You can always add this in later however if you want to keep it simple while you start! Setting up shipping profiles Etsy makes it very easy to set up different shipping rates for different countries. However, luckily with PrintShrimp you can offer free shipping to the majority of the major countries that are active on Etsy!  Using PrintShrimp means that your production costs are low enough in each domestic market to justify this. If you look on Etsy you can see there are many shops that post internationally to countries such as the US or Australia. Therefore, they often charge £8-10 in postage, and have a delivery time of 1-2 weeks. This really limits their customer base to their domestic market.  Using PrintShrimp avoids this and means you can offer free shipping (as we absorb the shipping cost in our prices) to the major markets of the UK, Australia, and USA (Europe coming soon!).  We also offer a 1 day processing time, unlike many POD poster suppliers. This means you can set your Etsy processing time to just one day, which combined with our quick shipping, means you will be one of the quickest on Etsy at sending out orders. This is obviously very attractive for customers, who are often very impatient with wanting their orders!  Getting the sales and extra tips \-Don’t list an insane amount of listings when you first get started. Etsy will be like ‘hang on a second’ if a brand new shop suddenly has 200 items in the first week. Warm up your account, and take things slow as you get going. We recommend 5 a day for the first week or so, and then you can start uploading more. You don’t want Etsy to flag your account for suspicious bot-like activity when you first get going.  \-It is very easy to copy listings when creating a new one. Simply select an old listing and press copy, and then you can just change the listing specific details to create a new one, rather than having to start from scratch. It can feel like a bit of a ball-ache setting up your first ever listing, but from then on you can just copy it over and just change the specifics.  \-Try and organize your listings into sections! This really helps the customer journey. Sometimes a customer will click onto your shop after seeing one of your listings, so it really helps if they can easily navigate your shop for what they are looking for. So, you now have a fully fledged Etsy shop. Well done! Time to start making £3,000 a month straight away right? Not quite. Please bear in mind, patience is key when starting out. If you started doing this because you are £10,000 in debt to the Albanian mafia and need to pay it off next week, you have come into this in the wrong frame of mind. If you have however started this to slowly build up a side hustle which hopefully one day become your full time gig, then winner winner chicken dinner.  Starting out on Etsy isn’t always easy. It takes time for your shop to build up trust! As I’ve said before, a buyer is far more likely to purchase from a shop with 1000s of reviews, than a brand new one with 0. But before you know it, you can become one of these shops! One thing you can do at the very start is to encourage your friends and family to buy your posters! This is a slightly naughty way of getting a few sales at the start, of course followed by a few glowing 5\* reviews. It really helps to give your shop this little boost at the start, so if this is something you can do then I recommend it.  Okay, so once you have a fully fledged shop with a decent amount of listings, you might be expecting the sales to start rolling in. And, if you are lucky, they indeed might. However, in my experience, you need to give your listings a little boost. So let us introduce you to: The wonderful world of Etsy ads Ads!! Oh no, that means money!! We imagine some of you more risk averse people are saying to yourself right now. And yes, it indeed does. But more often than not unfortunately you do have to spend money to make money.  Fortunately, in my experience anyway, Etsy ads do tend to work. This does however only apply if your products are actually good however, so if you’re back here after paying for ads for 2 months and are losing money at the same rate as your motivation, maybe go back to the start of this guide and pick another niche.  When you first start out, there are two main strategies.  Number 1: The Safer Option So, with PrintShrimp, you will essentially be making a minimum of £6 profit per order. With this in mind, I normally start a new shop with a safer strategy of advertising my products with a budget of $3-5 dollars a day. This then means that at the start, you only need to make 1 sale to break even, and anything above that is pure profit! This might not seem like the most dazzling proposition right now, but again please bear in mind that growth will be slow at the start. This means that you can gradually grow your shop, and therefore the trust that customers have in your shop, over time with a very small risk of ever actually losing money. Number 2: The Billy Big Balls Option If you were yawning while reading the first option, then this strategy may be for you. This will be better suited to those of you that are a bit more risk prone, and it also helps if you have a bit more cash to invest at the start. Through this strategy, you can essentially pay your way to the top of Etsy's rankings. For this, you’ll probably be looking at spending $20 a day on ads. So, this can really add up quickly and is definitely the riskier option. In my experience, the level of sales with this may not always match up to your spend every day. You may find that some days you rake in about 10 sales, and other days only one. But what this does mean is that as your listings get seen and purchased more, they will begin to rank higher in Etsy’s organic search rankings, at a much quicker rate than option one. This is the beauty of Etsy’s ads. You can pay to boost your products, but then results from this paid promotion feed into the organic ranking of your products. So you may find that you can splash the cash for a while at the start in order to race to the top, and then drop your ad spending later on when your products are already ranking well.  Sending your poster orders So, you’ve now done the hard bit. You have a running Etsy store, and essentially all you need to now on a daily basis is send out your orders and reply to customer messages! This is where it really becomes passive income.  \-Check out the PrintShrimp order portal. Simply sign up, and you can place individual orders through there. \-Bulk upload: We have an option to bulk upload your Esty orders via csv.  Seriously, when you are up and running with your first store, it is really as easy as that.  Once you have your first Etsy store up and running, you can think about expanding. There are many ways to expand your income. You can set up other Etsy stores, as long as the type of posters you are selling varies. You can look into setting up your own Shopify stores, and advertise them through Facebook, Instagram etc. Through this guide, we will teach you everything you need to know about starting to sell posters and generate some income. We will also show you why PrintShrimp is the best POD supplier for all of your poster needs. Trust me, you won’t need much convincing.

Changing Careers, changing products? Age 38, Direction needed, investment advice too.
reddit
LLM Vibe Score0
Human Vibe Score0.667
Salad-BanditThis week

Changing Careers, changing products? Age 38, Direction needed, investment advice too.

Hello, At one point in my life I had a set plan that I had been following in which to design a life that fit my values, but during 2020 the viability was called into question and I have been on bad footing, unable to find stability, since. Though I currently have stable housing without roommate, and enough in savings for a year without any income and three more years in a mutual fund. The question I need help with is about utilizing approximately $40,000 that I would like to invest into a new or existing business venture, or possibly start investing my own hand in selecting stocks. To give context about the parameters of concepts that pertain to me, back in 2005 I graduated highschool and immediately was an entrepreneur, started a sports clothing company, was selling WoW bot accounts, ghillie suits on ebay, and graphic design commissions, and I was proficient in MX Flash. Although the first part of my life plan to start farming three years before 2012 for what I thought would be a peak oil economic collapse, and while watching 2008 unfold, along with my career in MX Flash falling flat, I started farming 2009. From that point I spent a total of 15 years farming, the majority of that was for my own LLC, where I was situated with leases on million dollar properties as Ag tax write off, on an elite island outside a major city, serving local high price wholesale, mainly salad mix and mushrooms, because they are fast turn around. That was truly the best 20s I could have asked for, working mainly for myself, very healthy and was putting away $10-20k in savings/investments per year, plus was earning about $3-5k more per year, while living in a cargo trailer on dirt cheap leases. But it all came to a slow end starting in 2020 when I lost all of my wholesale overnight, and my retail exploded, which burnt me out to the point I couldnt walk, as the sole worker in my LLC. So I do not fully trust the volatility of the wholesale food industry, from a small grower’s perspective, since i don't own land. SO now I am trying to figure out a way forward, because I can always farm in the future, and have taught myself hydroponics, and flat packed farm equipment, so my business is very agile and now I can grow in parking lots closer to the city for more sales opportunities, but I am not sure that is what I want to do in this current moment, because tech is exploding, and we have never had so much information available to us, it's a shame not to spend a moment in life to discover what new opportunities might be out there. I was laid off twice last year, so I've been out of work the past four months, doing thriftstore routes twice a week while making about $500+/wk, really just trying to understand what people still buy and break even, while I continue to study 3d design blender, as well as 2d digital art in the hopes that I can reconnect with my tech art past, because that is what I told myself when I was 18, that I would put off art and computers until I was past 30 and needed to do less with my body. But over the past three years, the better I get at digital art, the better Ai has been getting too. I have some mentors who might give me work and a foot in the door, but most of them are laid off, and scrounging for work if they are not on their own funded indie project. I've thought about continuing to learn 3d modeling despite Ai, and despite seeing Flash, computer program I was proficient in get removed from existence before I could really earn my money back. I assume there will always be a need for Ai models to get cleaned up, mapped and rigged, especially with AR technology coming to consumers soon, but more over it would help if I decided to go to a community college to do CNC certificates, so I can have that as a backup job on CAD at a machining warehouse and do my farm and digital art on the side, but CNC mechanics don't make a crazy amount of money and have a boss. BUT I am an inventor, and have two inventions so far, plus my ultimate goal is to one day have automated hydroponic greenhouses, using all CNC+3d printed parts to create a low time investment agriculture income, with Ai monitored greenhouse, seed to salad product that i can sell to other people, which would tie into my desire to teach people about farming too, as well as do something I enjoy, but it is not a proven concept yet. Anyways if you've read this far I appreciate it, I ultimately would like 3rd party feedback about how I should spend my $40k surplus cash. I originally had it saved and accessible in case I was going to lease land and start my full farm business again from scratch, but I think using the equipment and space I have, and exploring non-perishable products is a smart move for me right now. Should I invest in inventory of products to arbitrage online? Should I invest in the top index funds? Should I buy Silver? Should I invest in inventory of a new product line? Should I spend some money insuring and equipment for a landscaping company? I want to future proof myself the best I can as Ai unfolds, I am pretty set with an income for the rest of my life as long as I can grow food and sell it, but there are currently so many changing opportunities, I want to cast out my net and see what works with my temperment. I’ve thought about getting into cyber security, or maybe be an electrician, or less staple jobs like Landscape Architech (can use art/modeling) and CNC engineer/modeler, but honestly I prefer to make a product and sell it without client service related interaction, and particularly no boss. Thank you for reading

[Ultimate List] A list of Marketing Tools That I’ve tested over the years and found helpful to do better marketing with less work. More than 50 Tools To Help you with Marketing, Copywriting & Sales!
reddit
LLM Vibe Score0
Human Vibe Score0.973
lazymentorsThis week

[Ultimate List] A list of Marketing Tools That I’ve tested over the years and found helpful to do better marketing with less work. More than 50 Tools To Help you with Marketing, Copywriting & Sales!

Starting to focus on marketing for your business, You will come across the same tools mentioned over and over by marketers. I would like to mention here tools that you might haven’t seen going viral in the community but actually will help you grow faster and efficiently. Starting off with My favourite Marketing Channel! #Email Marketing For SMBs Convertkit / Mailerlite / Mailchimp - These 3 Platforms are the best options for SMBs and entrepreneurs just starting out with email marketing. All 3 have free plans up to 1,000 subscribers. Scribe - Email Signature Tool, Create Great Email signatures for your emails. Liramail - Most Email marketing platforms don’t offer great email templates. This tool will help you build great email templates with drag and drop. Quick mail Auto-Warmer - Most Businesses at the beginning don’t know what to do when open rate drops. You need to use an email warmer like this to keep it up. #Email Marketing For Big Businesses SendGrid - Overall Email Marketing Tools, this tool is best for brands that have huge email lists and email marketing is the key marketing channel. Braze - This tool is leading in email marketing for large Email senders. When I was working for agencies, this was one of the best email marketing tools I had used. NeoCertified - Protect your emails for spammers and threats. To keep your email list healthy, this is a must have! Sparkloop - Referral Marketing For Email Campaigns. Email can generate great huge amount of referrals for you and Sparkloop makes it easier. #Cold Emails & Lead Generation Hunter - A Great Tool to scrape emails from domain names. The tool comes with a green free plan but Pro plan is worth the amount of features it provides. Icyleads - It’s better than Hunter as it’s heavily focused on the sales and prospecting to help you derive great results from your campaigns. Mailshake - Beginner Friend Cold Email Tool with Great features like email list warming. #Communication Tools Twilio - One do the best customer engagement platform used by Companies like Stripe and mine too. Chatlio - Use Live chat feature on your website with slack integration. My favourite easier to catch up on conversations through slack integration. Intercom - Used by Most Marketers, Industry Leading customer communication platform. Great for beginners! Chatwoot - Another Amazing Communication Tool but the best part is they have a great free plan useful for new businesses. Loom - Communicate with your audience through Videos. Loom is great for SaaS and to show human interaction to close new visitors effectively. #CRM Outseta - This tool provides great CRM and their billing system is better than other tools out their which makes it stands out! Hubspot - I don’t think this tool needs an introduction because Hubspot’s CRM is the best in industry. Salesflare - This CRM is a great alternative to hubspot as it’s beginner friendly and helpful for SMBs. #SEO Tools Ahrefs - One of the best SEO tool in the industry. They also just launched a bunch of free tools to help SEO beginners. Screaming frog - The only website crawler I have used since I bought my first domain. It’s the best! Ubersuggest- The Tool by Neil Patel is the best SEO tool for you. (I’m Joking, it’s the worst) Contentking - This tool is good at Real-time SEO Auditing, they do a lot of Marketing work through Newsletters. If you are subscribed to any SEO newsletter. You may have seen this tool. SEOquake & Semrush - SEOquake is a great tool to conduct on-page analysis, SERP, and much more. Great tool but it’s owned by Semrush. You should go for Semrush because that tool will cover all SEO aspects for you. #Content Marketing Buzzsumo - This tool is great for content research and but you may find the regular emails pretty annoying sometimes. Contentrow - Analyse Your Content and find it’s strength. Highly recommended who are weak at content structuring like me. Grammarly - If you are not a native English speaker like me, you might think you need it or not. You need it for sure for grammar corrections. #Graphic Design Tools Visme - At agencies, Infographics can be more effective than usual postscript. Visme is a graphic design tool focused on infographics and designs related to B2B and B2C. It’s great for agencies! Glorify - A Graphic Design Tool focused on E-commerce, filled with Designs useful for E-commerce store owners. Canva - All-in-one Industry leading Graphic Design Tool that everyone knows and every template is overused now. Adobe Creative Cloud ( previously Sparkpost) - It’s a great alternative to Canva filled with Amazing Stock images to use in your visuals but the only backlash is the exports in this tool are not high quality. Snaps - A Canva Alternative that might not have overused templates for your Social Accounts. #Advertising Tools Plai - It’s a great PPC tool to create Ads for Instagram and Tiktok. Wordstream - It’s an industry leading PPC Tool, great for Ad Grading and auditing. AdEspresso - This Is a tool by Hootsuite. They have a lot of Data sourced at the backend, which helps in Ad optimisation through this tool. That’s the reason I recommend this tool. #Video Editing Tools Veed Studio - I have been using Veed from last year. It’s one of the best Video Marketing Tool Optimized for Instagram & Tiktok. Synthesia - It’s a new AI video generation platform. From last few months, if you have seen marketing agencies including Videos in Emails. The chances are that’s not a Agency member taking but AI generated Human. Motionbox - It’s also a great video editing tool focused on video editing for Digital Marketers. Jitter Video - It’s a great motion design tool. Comes with great templates, the only place where other tools I mentioned lacks. It’s great and beginner friendly. #Copywriting Jasper AI - Google’s John Mueller says AI generated content is banned on Search but I think with Jasper AI you can generate SEO optimised Content but you have to put in some efforts like at least give 30 minutes for editing the Copy by yourself. Copy AI - Another AI tool to help you write better copy. This one is more focused on helping you write copy suitable for Ads and Social media campaigns. Hemingway App - To help you write more clearly and Bold. This tool is better than Grammarly if you look for writing perspective and it’s free. #Social Media Management App I’ve used a Lot of SMM Tools and that’s why going to mention all of them with a short review. Sprout social - The Best with deep insights coverage. Hootsuite - Great Scheduling tool just under sprout social. Later - Heavily Focused on Instagram from beginning and Now Tiktok too. SkedSocial - It’s like a Later alternative with great addition features like link-in-bio. Facebook’s Business Manager- Great but sometimes bugs can make a huge issue for you and customer support is like dead. Tweet Hunter & Hypefury- Both are Twitter Scheduling tools growing very fast on platform and are great for growth. Buffer - It’s a great tool but I haven’t seen any new updates to help with management. Zoho Social - It’s a great SMM tool and if you use other marketing solutions from Zoho. It’s a must have! #Market Research Tool • SparkToro - That’s the only one I have ever used. It’s great for audience research and comes with great customer service. Founded by Rand Fishkin, it’s one of the best research tool. #Influencer Marketing & UGC InfluenceGrid - A free search engine To find Tiktok & Instagram Influencers for your campaigns. Tiktok Creative Center- TikTok’s in-built tool called “Creative Center” is the best to find content trends, audience demographics and much more. Archive - Find Instagram Stories and Posts mentioning Your brands and use them as Ads for your business Marketing. #Landing Page Builders Leadpages - Its a great landing page builder because the integration and drag-and-drop features makes it easier to work with! Cardd co - A Great Landing page builder with easy step up but it lacks the copywriting and tracking features. Instapage - It’s one of the best out and I think the overall product is effective enough to help you stand out with your landing page. Unbounce - It’s a great alternative to Instapage due its well polished landing page templates that might be helpful for you. #Community Building Mighty Networks - A Great Community building platform, and you can also sell courses within the platform. Circle so - A great alternative to Mighty networks focused on Communities specifically. We are currently using for small community Of ours. #Sales Tools Drift - You can get much more out of Drift than just sales tools but The Sales solutions provided in Drift are one of the best. Salesforce - It’s the industry Sales solution provider. A go-to and have various pricing plans making it suitable for majority of SMBs. #Social Proof Tools People don’t have enough time to search across internet to decide to trust you after seeing your Ad first time. That’s what you might be facing too. Here are two tools I absolutely love for social proof! Use Proof - Show Recent Activities occurring on your website and build the trust of your visitors. Testimonial to - Gather Testimonials across Social Media platforms related to your business with this tool. Capture tweets and comments mentioning your brands and mention them. #Analytics Tools Plausible Analytics- A privacy friendly Analytics alternative to Google Analytics if you hate Analytics 4 like me. Mixpanel - Product Analytics and funnel reports better than Google Analytics. #Reddit Marketing Gummysearch- This tool will help To find your target audience on Reddit and interact with them with its help and close your new customers. Howitzer- It’s another pretty similar tool to Gummysearch focused on Reddit cold outreach to get clients and new customers. Both are great but Gummysearch provides better customer support while Howtizer is helpful on a large scale Reddit Marketing. #Text Marketing Klaviyo - It’s an email + SMS marketing tool, it’s taking up space in marketing industry very quickly as an industry leader due to its great integrations but you need to learn the platform usage to maximise the outcome. Cartloop - This tool provides great text marketing solutions with integration with Spotify and other e-commerce marketing tools. Attentive Mobile - This is my favourite Text marketing tool due to the interactive dashboard + they have a library of Text marketing examples to help you out with your campaigns. #Other Tools I have used throughout my journey! Triple Whale - It’s a great E-commerce marketing tools with Triple pixel to help you track your campaigns more efficiently. Fastory - To create well optimized Instagram & Tiktok Stories for your business. Jotform - Online Form Builder with integrations with leading marketing tools. Gated - As an entrepreneur and marketer, you may receive a bunch of unwanted emails. Use Gated to get rid of them and receive useful mails only! ClickUp- The main Tool for Project Management, one of the best and highly recommended. Riverside - Forget Zoom or Google Meet, For your Podcast Interviews and Marketing conferences. You need riverside with great video quality and recording features. Manychat- Automate your Instagram DMs and interact with your followers more efficiently + sell out your products/ services when you are offline. Calendy - To schedule meetings with your ideal clients. ServiceProviderPro - It’s a client portal for SEO & Growing Agencies, very helpful in scaling agencies. SendCheckit - Compare your Email Subject Lines with 100,000+ others in the database for free. Otter AI - Using AI track your meetings more effectively, you can easily edit, annotate and share notes from the meetings. Ryte - Optimise your website User experience with this tool focused on UX aspects + SEO too. PhantomBuster - Scrape LinkedIn Profile and Data from Facebook/LinkedIn groups. I clearly love this tool! #Honourable Mentions Zapier - The Only tool you need to integrate your favourite tool with a new effective tool. Elementor - That’s what I use for web design and it’s great! Marketer Hire - To hire world class marketers to work with you. InShot & Capcut - I create Instagram Reels and TikTok’s and life without these tools isn’t possible. Nira - It’s a great tool to Manage your workspace and this tool has launched many marketing templates in-built helpful for marketers and also entrepreneurs. X - The tool you love that wasn’t mentioned here is valuable and I honour that tool and share that if you would like to! I mean thanks for reading what I have curated all over my life as a marketer. I share 5 Marketing Tools, 5 Marketing Resources and 1 Free Resourceevery week in my newsletter, you can subscribe here to receive that for free. Also, You can read an expanded list of email marketing tools in this Reddit post!

9 minimalistic habits that will save you 1,000+ hours of your life
reddit
LLM Vibe Score0
Human Vibe Score1
Omeet2This week

9 minimalistic habits that will save you 1,000+ hours of your life

📱 Digital Habits Unsubscribe from Unnecessary Emails Consider using an app like unroll.me to help you declutter your inbox. It's free. Streamline Your Finances Automating your finances can save you a lot of time and stress in the long run. Here's how: \- Automate monthly payments \- Cancel unused subscriptions \- Automate contributions to investments \- Set up a system to automatically deposit 20% of your paycheck into savings Deweaponize Your Device Your phone can be a major time sink. Make it less so with these steps: \- Set screen time limits \- Switch your device to grayscale \- Turn off non-emergency notifications \- Use an app like onesec.app to add friction \- Organize apps into folders based on their purpose Embrace Journaling or a Productivity Coach For journaling, consider the DayOneApp. If you want to delve deeper, try a productivity coach app like Wave.ai. ​ 🏡 Lifestyle Habits Declutter Daily Aim to get rid of just one thing every day. It's a simple way to discover what possessions truly matter to you. Build a Capsule Wardrobe A capsule wardrobe is a minimal collection of versatile items. Here's a quick guide: \- Decide on the number of pieces (up to 50) \- Start with what you already have \- Choose a color palette \- Retain essential/versatile items \- Donate what you don’t need Invest in High-Quality Items Remember, cheap can often end up being more expensive in the long run. ​ 🧠 Mental Habits Prune Your To-Do List As Jim Collins said, "If you have more than 3 priorities, then you don’t have any." Apply the 2-Minute Rule If a task takes less than 2 minutes to complete, do it immediately. Tasks tend to become more daunting the longer we procrastinate. ​ Should I add some more? Edited to add more from comments (From KidBeene): Don't reply to emails for 3 days. Either the issue will resolve itself or it will reduce options down to 1 or 2. No one is shooting at you, no one will die from you not stepping in to an email chain. If it was truly important your phone would be ringing. When people want to meet with you only accept meetings that have an agenda. No meeting should be FYI\- Those are emails or dashboards. Only agenda items that require a decision from you or that require you to step in to descalate or escalate a situation. If an FYI or update to a situation is needed, make it into a paragraph update\- 1st sentence of What happened that made you need to tell me something? Second sentence is historical/has this happened before. 3rd sentence is "so what/why do I care" and 4th sentence is recommendation. ​

I tested hundreds of marketing tools in the last three years and these 50 made it to the list. I'll sum up my top 50 marketing tools with one or two sentences + give you pricings.
reddit
LLM Vibe Score0
Human Vibe Score1
SpicyCopyThis week

I tested hundreds of marketing tools in the last three years and these 50 made it to the list. I'll sum up my top 50 marketing tools with one or two sentences + give you pricings.

Hey guys, I'm working in a growth marketing agency. Marketing tools are 30% of what we do, so we use them a lot and experiment with the new ones as much as possible. There are thousands of tools and it's easy to get lost, so I wanted to share the tools we use most on a daily basis. And divide the list into 14 categories. I thought this could be handy for Entrepreneurs subreddit. Why adopt tools? I see marketing tools as tireless colleagues. If you can't hire an employee, choosing the right tool can solve your problems, because they Are super cheap. Work 7/24 for you. Don’t make mistakes. Don’t need management. (or needless management) Help you to automate the majority of your lead gen process. Onwards to the list. (With the pricings post ended up quite long, you can find a link in the end if you want to check the prices) Email marketing tools #1 ActiveCampaign is armed with the most complicated email automation features and has the most intuitive user experience. It feels like you already know how to use it. \#2 Autopilot is visual marketing automation and customer journey tool that helps you acquire, nurture based on behaviors, interest etc. #3 Mailjet: This is the tool we use to send out bulky email campaigns such as newsletters. It doesn't have sexy features like others but does its job for a cheap price. Email address finders #4 Skrapp finds email of your contacts by name and company. It also works with LinkedIn Sales Navigator and can extract thousands of emails in bulk + have a browser add-on. #5 Hunter: Similar to Skrapp but doesn't work with LinkedIn Sales Navigator directly. In addition, there are email templates and you can set up email campaigns. Prospecting and outreach tools #6 Prospect combines the personal emails, follow-up calls, other social touches and helps you create multichannel campaigns.  #7 Reply is a more intuitive version of Prospect. It is easy to learn and use; their UX makes you feel good and sufficient.  CRM tools #8 Salesflare helps you to stop managing your data and start managing your customers. Not yet popular as Hubspot and etc but the best solution for smaller B2B businesses. (we're fans) \#9 Hubspot: The most popular CRM for good reason and has a broader product range you can adopt in your next steps. Try this if you have a bulky list of customers because it is free. #10 Pardot: Pardot is by Salesforce, it's armed with features that can close the gap between marketing and sales. Sales Tools #11 Salesforce is the best sales automation and lead management software. It helps you to create complicated segmentations and run, track, analyze campaigns from the same dashboard. #12 LinkedIn Sales Navigator gives you full access to LinkedIn's user database. You can even find a kidnapped CEO if you know how to use it with other marketing automation tools like Skrapp. #13 Pipedrive is a simple tool and excels in one thing. It tracks your leads and tells you when to take the next action. It makes sales easier. #14 Qwilr creates great-looking docs, at speed. You can design perfect proposals, quotes, client updates, and more in a flash. We use it a lot to close deals, it's effective. #15 Crystalknows is an add-on that tells you anyone’s personality on LinkedIn and gives you a detailed approach specific to that person. It's eerily accurate. #16 Leadfeeder shows you the companies that visited your website. Tells how they found you and what they’re interested in. It has a free version. Communication Tools #17 Intercom is a sweet and smart host that welcomes your visitors when you’re not home. It’s one of the best chatbot tools in the market. #18 Drift is famous for its conversational marketing features and more sales-focused than Intercom. #19 Manychat is a chatbot that helps you create high converting Facebook campaigns. #20 Plann3r helps you create your personalized meeting page. You can schedule meetings witch clients, candidates, and prospects. #21 Loom is a video messaging tool, it helps you to be more expressive and create closer relationships. #22 Callpage collects your visitors’ phone number and connects you with them in seconds. No matter where you are. Landing page tools #23 Instapage is the best overall landing page builder. It has a broad range of features and even squirrel can build a compelling landing page with templates. No coding needed. #24 Unbounce can do everything that Instapage does and lets you build a great landing page without a developer. But it's less intuitive. Lead generation / marketing automation tools #25 Phantombuster is by far the most used lead generation software in our tool kit. It extracts data, emails, sends requests, customized messages, and does many things on autopilot in any platform. You can check this, this and this if you want to see it in action. #26 Duxsoup is a Google Chrome add-on and can also automate some of LinkedIn lead generation efforts like Phantombuster. But not works in the cloud. #27 Zapier is a glue that holds all the lead generation tools together. With Zapier, You can connect different marketing tools and no coding required. Conversion rate optimization tools #28 Hotjar tracks what people are doing on your website by recording sessions and capturing mouse movements. Then it gives you a heatmap. #29 UsabilityHub shows your page to a digital crowd and measures the first impressions and helps you to validate your ideas. #30 Optinmonster is a top tier conversion optimization tool. It helps you to capture leads and enables you to increase conversions rates with many features. #31 Notifia is one mega tool of widgets that arms your website with the wildest social proof and lead capturing tactics. #32 Sumo is a much simpler version of Notifia. But Sumo has everything to help you capture leads and build your email lists. Web scrapers #33 Data Miner is a Google Chrome browser extension that helps you scrape data from web pages and into a CSV file or Excel spreadsheet. #34 Webscraper does the same thing as Data Miner; however, it is capable of handling more complex tasks. SEO and Content #35 Grammarly: Your English could be your first language and your grammar could be better than Shakespeare. Grammarly still can make your writing better. #36 Hemingwayapp is a copywriting optimization tool that gives you feedback about your copy and improves your readability score, makes your writing bolder and punchier. Free. #37 Ahrefs is an all-rounder search engine optimization tool that helps you with off-page, on-page or technical SEO. #38 SurferSEO makes things easier for your on-page SEO efforts. It’s a tool that analyzes top Google results for specific keywords and gives you a content brief based on that data. Video editing and design tools #39 Canva is a graphic design platform that makes everything easy. It has thousands of templates for anything from Facebook ads, stylish presentations to business cards.  #40 Kapwing is our go-to platform for quick video edits. It works on the browser and can help you to create stylish videos, add subtitles, resize videos, create memes, or remove backgrounds. #41 Animoto can turn your photos and video clip into beautiful video slideshows. It comes handy when you want to create an advertising material but don’t have a budget. Advertising tools #42 AdEspresso lets you create and test multiple ads with few clicks. You can optimize your FB, IG, and Google ads from this tool and measure your ads with in-depth analytics. #43 AdRoll is an AI-driven platform that connects and coordinates marketing efforts across ads, email, and online stores. Other tools #44 Replug helps you to shorten, track, optimize your links with call-to-actions, branded links, and retargeting pixels #45 Draw.io = Mindmaps, schemes, and charts. With Draw.io, you can put your brain in a digital paper in an organized way. #46 Built With is a tool that finds out what websites are built with. So you can see what tools they're using and so on. #47 Typeform can turn data collection into an experience with Typeform. This tool helps you to engage your audience with conversational forms or surveys and help you to collect more data. #48 Livestorm helped us a lot, especially in COVID-19 tiles. It’s a webinar software that works on your browser, mobile, and desktop. #49 Teachable \- If you have an online course idea but hesitating because of the production process, Teachable can help you. It's easy to configure and customizable for your needs. #50 Viral Loops provides a revolutionary referral marketing solution for modern marketers. You can create and run referral campaigns in a few clicks with templates. Remember, most of these tools have a free trial or free version. Going over them one by one can teach you a lot and help you grow your business with less work power in the early stages of your business. I hope you enjoyed the read and can find some tools to make things easier! Let me know about your favorite tools in the comments, so I can try them out. \------ If you want to check the prices and see a broader explanation about the tools, you can go here.

I run an AI automation agency (AAA). My honest overview and review of this new business model
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

I run an AI automation agency (AAA). My honest overview and review of this new business model

I started an AI tools directory in February, and then branched off that to start an AI automation agency (AAA) in June. So far I've come across a lot of unsustainable "ideas" to make money with AI, but at the same time a few diamonds in the rough that aren't fully tapped into yet- especially the AAA model. Thought I'd share this post to shine light into this new business model and share some ways you could potentially start your own agency, or at the very least know who you are dealing with and how to pick and choose when you (inevitably) get bombarded with cold emails from them down the line. Foreword Running an AAA does NOT involve using AI tools directly to generate and sell content directly. That ship has sailed, and unless you are happy with $5 from Fiverr every month or so, it is not a real business model. Cry me a river but generating generic art with AI and slapping it onto a T-shirt to sell on Etsy won't make you a dime. At the same time, the AAA model will NOT require you to have a deep theoretical knowledge of AI, or any academic degree, as we are more so dealing with the practical applications of generative AI and how we can implement these into different workflows and tech-stacks, rather than building AI models from the ground up. Regardless of all that, common sense and a willingness to learn will help (a shit ton), as with anything. Keep in mind - this WILL involve work and motivation as well. The mindset that AI somehow means everything can be done for you on autopilot is not the right way to approach things. The common theme of businesses I've seen who have successfully implemented AI into their operations is the willingess to work with AI in a way that augments their existing operations, rather than flat out replace a worker or team. And this is exactly the train of thought you need when working with AI as a business model. However, as the field is relatively unsaturated and hype surrounding AI is still fresh for enterprises, right now is the prime time to start something new if generative AI interests you at all. With that being said, I'll be going over three of the most successful AI-adjacent businesses I've seen over this past year, in addition to some tips and resources to point you in the right direction. so.. WTF is an AI Automation Agency? The AI automation agency (or as some YouTubers have coined it, the AAA model) at its core involves creating custom AI solutions for businesses. I have over 1500 AI tools listed in my directory, however the feedback I've received from some enterprise users is that ready-made SaaS tools are too generic to meet their specific needs. Combine this with the fact virtually no smaller companies have the time or skills required to develop custom solutions right off the bat, and you have yourself real demand. I would say in practice, the AAA model is quite similar to Wordpress and even web dev agencies, with the major difference being all solutions you develop will incorporate key aspects of AI AND automation. Which brings me to my second point- JUST AI IS NOT ENOUGH. Rather than reducing the amount of time required to complete certain tasks, I've seen many AI agencies make the mistake of recommending and (trying to) sell solutions that more likely than not increase the workload of their clients. For example, if you were to make an internal tool that has AI answer questions based on their knowledge base, but this knowledge base has to be updated manually, this is creating unnecessary work. As such I think one of the key components of building successful AI solutions is incorporating the new (Generative AI/LLMs) with the old (programmtic automation- think Zapier, APIs, etc.). Finally, for this business model to be successful, ideally you should target a niche in which you have already worked and understand pain points and needs. Not only does this make it much easier to get calls booked with prospects, the solutions you build will have much greater value to your clients (meaning you get paid more). A mistake I've seen many AAA operators make (and I blame this on the "Get Rich Quick" YouTubers) is focusing too much on a specific productized service, rather than really understanding the needs of businesses. The former is much done via a SaaS model, but when going the agency route the only thing that makes sense is building custom solutions. This is why I always take a consultant-first approach. You can only build once you understand what they actually need and how certain solutions may impact their operations, workflows, and bottom-line. Basics of How to Get Started Pick a niche. As I mentioned previously, preferably one that you've worked in before. Niches I know of that are actively being bombarded with cold emails include real estate, e-commerce, auto-dealerships, lawyers, and medical offices. There is a reason for this, but I will tell you straight up this business model works well if you target any white-collar service business (internal tools approach) or high volume businesses (customer facing tools approach). Setup your toolbox. If you wanted to start a pressure washing business, you would need a pressure-washer. This is no different. For those without programming knowledge, I've seen two common ways AAA get setup to build- one is having a network of on-call web developers, whether its personal contacts or simply going to Upwork or any talent sourcing agency. The second is having an arsenal of no-code tools. I'll get to this more in a second, but this works beecause at its core, when we are dealing with the practical applications of AI, the code is quite simple, simply put. Start cold sales. Unless you have a network already, this is not a step you can skip. You've already picked a niche, so all you have to do is find the right message. Keep cold emails short, sweet, but enticing- and it will help a lot if you did step 1 correctly and intimately understand who your audience is. I'll be touching base later about how you can leverage AI yourself to help you with outreach and closing. The beauty of gen AI and the AAA model You don't need to be a seasoned web developer to make this business model work. The large majority of solutions that SME clients want is best done using an API for an LLM for the actual AI aspect. The value we create with the solutions we build comes with the conceptual framework and design that not only does what they need it to but integrates smoothly with their existing tech-stack and workflow. The actual implementation is quite straightforward once you understand the high level design and know which tools you are going to use. To give you a sense, even if you plan to build out these apps yourself (say in Python) the large majority of the nitty gritty technical work has already been done for you, especially if you leverage Python libraries and packages that offer high level abstraction for LLM-related functions. For instance, calling GPT can be as little as a single line of code. (And there are no-code tools where these functions are simply an icon on a GUI). Aside from understanding the capabilities and limitations of these tools and frameworks, the only thing that matters is being able to put them in a way that makes sense for what you want to build. Which is why outsourcing and no-code tools both work in our case. Okay... but how TF am I suppposed to actually build out these solutions? Now the fun part. I highly recommend getting familiar with Langchain and LlamaIndex. Both are Python libraires that help a lot with the high-level LLM abstraction I mentioned previously. The two most important aspects include being able to integrate internal data sources/knowledge bases with LLMs, and have LLMs perform autonomous actions. The two most common methods respectively are RAG and output parsing. RAG (retrieval augmented Generation) If you've ever seen a tool that seemingly "trains" GPT on your own data, and wonder how it all works- well I have an answer from you. At a high level, the user query is first being fed to what's called a vector database to run vector search. Vector search basically lets you do semantic search where you are searching data based on meaning. The vector databases then retrieves the most relevant sections of text as it relates to the user query, and this text gets APPENDED to your GPT prompt to provide extra context to the AI. Further, with prompt engineering, you can limit GPT to only generate an answer if it can be found within this extra context, greatly limiting the chance of hallucination (this is where AI makes random shit up). Aside from vector databases, we can also implement RAG with other data sources and retrieval methods, for example SQL databses (via parsing the outputs of LLM's- more on this later). Autonomous Agents via Output Parsing A common need of clients has been having AI actually perform tasks, rather than simply spitting out text. For example, with autonomous agents, we can have an e-commerce chatbot do the work of a basic customer service rep (i.e. look into orders, refunds, shipping). At a high level, what's going on is that the response of the LLM is being used programmtically to determine which API to call. Keeping on with the e-commerce example, if I wanted a chatbot to check shipping status, I could have a LLM response within my app (not shown to the user) with a prompt that outputs a random hash or string, and programmatically I can determine which API call to make based on this hash/string. And using the same fundamental concept as with RAG, I can append the the API response to a final prompt that would spit out the answer for the user. How No Code Tools Can Fit In (With some example solutions you can build) With that being said, you don't necessarily need to do all of the above by coding yourself, with Python libraries or otherwise. However, I will say that having that high level overview will help IMMENSELY when it comes to using no-code tools to do the actual work for you. Regardless, here are a few common solutions you might build for clients as well as some no-code tools you can use to build them out. Ex. Solution 1: AI Chatbots for SMEs (Small and Medium Enterprises) This involves creating chatbots that handle user queries, lead gen, and so forth with AI, and will use the principles of RAG at heart. After getting the required data from your client (i.e. product catalogues, previous support tickets, FAQ, internal documentation), you upload this into your knowledge base and write a prompt that makes sense for your use case. One no-code tool that does this well is MyAskAI. The beauty of it especially for building external chatbots is the ability to quickly ingest entire websites into your knowledge base via a sitemap, and bulk uploading files. Essentially, they've covered the entire grunt work required to do this manually. Finally, you can create a inline or chat widget on your client's website with a few lines of HTML, or altneratively integrate it with a Slack/Teams chatbot (if you are going for an internal Q&A chatbot approach). Other tools you could use include Botpress and Voiceflow, however these are less for RAG and more for building out complete chatbot flows that may or may not incorporate LLMs. Both apps are essentially GUIs that eliminate the pain and tears and trying to implement complex flows manually, and both natively incoporate AI intents and a knowledge base feature. Ex. Solution 2: Internal Apps Similar to the first example, except we go beyond making just chatbots but tools such as report generation and really any sort of internal tool or automations that may incorporate LLM's. For instance, you can have a tool that automatically generates replies to inbound emails based on your client's knowledge base. Or an automation that does the same thing but for replies to Instagram comments. Another example could be a tool that generates a description and screeenshot based on a URL (useful for directory sites, made one for my own :P). Getting into more advanced implementations of LLMs, we can have tools that can generate entire drafts of reports (think 80+ pages), based not only on data from a knowledge base but also the writing style, format, and author voice of previous reports. One good tool to create content generation panels for your clients would be MindStudio. You can train LLM's via prompt engineering in a structured way with your own data to essentially fine tune them for whatever text you need it to generate. Furthermore, it has a GUI where you can dictate the entire AI flow. You can also upload data sources via multiple formats, including PDF, CSV, and Docx. For automations that require interactions between multiple apps, I recommend the OG zapier/make.com if you want a no-code solution. For instance, for the automatic email reply generator, I can have a trigger such that when an email is received, a custom AI reply is generated by MyAskAI, and finally a draft is created in my email client. Or, for an automation where I can create a social media posts on multiple platforms based on a RSS feed (news feed), I can implement this directly in Zapier with their native GPT action (see screenshot) As for more complex LLM flows that may require multiple layers of LLMs, data sources, and APIs working together to generate a single response i.e. a long form 100 page report, I would recommend tools such as Stack AI or Flowise (open-source alternative) to build these solutions out. Essentially, you get most of the functions and features of Python packages such as Langchain and LlamaIndex in a GUI. See screenshot for an example of a flow How the hell are you supposed to find clients? With all that being said, none of this matters if you can't find anyone to sell to. You will have to do cold sales, one way or the other, especially if you are brand new to the game. And what better way to sell your AI services than with AI itself? If we want to integrate AI into the cold outreach process, first we must identify what it's good at doing, and that's obviously writing a bunch of text, in a short amount of time. Similar to the solutions that an AAA can build for its clients, we can take advantage of the same principles in our own sales processes. How to do outreach Once you've identified your niche and their pain points/opportunities for automation, you want to craft a compelling message in which you can send via cold email and cold calls to get prospects booked on demos/consultations. I won't get into too much detail in terms of exactly how to write emails or calling scripts, as there are millions of resources to help with this, but I will tell you a few key points you want to keep in mind when doing outreach for your AAA. First, you want to keep in mind that many businesses are still hesitant about AI and may not understand what it really is or how it can benefit their operations. However, we can take advantage of how mass media has been reporting on AI this past year- at the very least people are AWARE that sooner or later they may have to implement AI into their businesses to stay competitive. We want to frame our message in a way that introduces generative AI as a technology that can have a direct, tangible, and positive impact on their business. Although it may be hard to quantify, I like to include estimates of man-hours saved or costs saved at least in my final proposals to prospects. Times are TOUGH right now, and money is expensive, so you need to have a compelling reason for businesses to get on board. Once you've gotten your messaging down, you will want to create a list of prospects to contact. Tools you can use to find prospects include Apollo.io, reply.io, zoominfo (expensive af), and Linkedin Sales Navigator. What specific job titles, etc. to target will depend on your niche but for smaller companies this will tend to be the owner. For white collar niches, i.e. law, the professional that will be directly benefiting from the tool (i.e. partners) may be better to contact. And for larger organizations you may want to target business improvement and digital transformation leads/directors- these are the people directly in charge of projects like what you may be proposing. Okay- so you have your message, and your list, and now all it comes down to is getting the good word out. I won't be going into the details of how to send these out, a quick Google search will give you hundreds of resources for cold outreach methods. However, personalization is key and beyond simple dynamic variables you want to make sure you can either personalize your email campaigns directly with AI (SmartWriter.ai is an example of a tool that can do this), or at the very least have the ability to import email messages programmatically. Alternatively, ask ChatGPT to make you a Python Script that can take in a list of emails, scrape info based on their linkedin URL or website, and all pass this onto a GPT prompt that specifies your messaging to generate an email. From there, send away. How tf do I close? Once you've got some prospects booked in on your meetings, you will need to close deals with them to turn them into clients. Call #1: Consultation Tying back to when I mentioned you want to take a consultant-first appraoch, you will want to listen closely to their goals and needs and understand their pain points. This would be the first call, and typically I would provide a high level overview of different solutions we could build to tacke these. It really helps to have a presentation available, so you can graphically demonstrate key points and key technologies. I like to use Plus AI for this, it's basically a Google Slides add-on that can generate slide decks for you. I copy and paste my default company messaging, add some key points for the presentation, and it comes out with pretty decent slides. Call #2: Demo The second call would involve a demo of one of these solutions, and typically I'll quickly prototype it with boilerplate code I already have, otherwise I'll cook something up in a no-code tool. If you have a niche where one type of solution is commonly demanded, it helps to have a general demo set up to be able to handle a larger volume of calls, so you aren't burning yourself out. I'll also elaborate on how the final product would look like in comparison to the demo. Call #3 and Beyond: Once the initial consultation and demo is complete, you will want to alleviate any remaining concerns from your prospects and work with them to reach a final work proposal. It's crucial you lay out exactly what you will be building (in writing) and ensure the prospect understands this. Furthermore, be clear and transparent with timelines and communication methods for the project. In terms of pricing, you want to take this from a value-based approach. The same solution may be worth a lot more to client A than client B. Furthermore, you can create "add-ons" such as monthly maintenance/upgrade packages, training sessions for employeees, and so forth, separate from the initial setup fee you would charge. How you can incorporate AI into marketing your businesses Beyond cold sales, I highly recommend creating a funnel to capture warm leads. For instance, I do this currently with my AI tools directory, which links directly to my AI agency and has consistent branding throughout. Warm leads are much more likely to close (and honestly, much nicer to deal with). However, even without an AI-related website, at the very least you will want to create a presence on social media and the web in general. As with any agency, you will want basic a professional presence. A professional virtual address helps, in addition to a Google Business Profile (GBP) and TrustPilot. a GBP (especially for local SEO) and Trustpilot page also helps improve the looks of your search results immensely. For GBP, I recommend using ProfilePro, which is a chrome extension you can use to automate SEO work for your GBP. Aside from SEO optimzied business descriptions based on your business, it can handle Q/A answers, responses, updates, and service descriptions based on local keywords. Privacy and Legal Concerns of the AAA Model Aside from typical concerns for agencies relating to service contracts, there are a few issues (especially when using no-code tools) that will need to be addressed to run a successful AAA. Most of these surround privacy concerns when working with proprietary data. In your terms with your client, you will want to clearly define hosting providers and any third party tools you will be using to build their solution, and a DPA with these third parties listed as subprocessors if necessary. In addition, you will want to implement best practices like redacting private information from data being used for building solutions. In terms of addressing concerns directly from clients, it helps if you host your solutions on their own servers (not possible with AI tools), and address the fact only ChatGPT queries in the web app, not OpenAI API calls, will be used to train OpenAI's models (as reported by mainstream media). The key here is to be open and transparent with your clients about ALL the tools you are using, where there data will be going, and make sure to get this all in writing. have fun, and keep an open mind Before I finish this post, I just want to reiterate the fact that this is NOT an easy way to make money. Running an AI agency will require hours and hours of dedication and work, and constantly rearranging your schedule to meet prospect and client needs. However, if you are looking for a new business to run, and have a knack for understanding business operations and are genuinely interested in the pracitcal applications of generative AI, then I say go for it. The time is ticking before AAA becomes the new dropshipping or SMMA, and I've a firm believer that those who set foot first and establish themselves in this field will come out top. And remember, while 100 thousand people may read this post, only 2 may actually take initiative and start.

AI Voice Platform Comparison for Small Business Use Cases
reddit
LLM Vibe Score0
Human Vibe Score1
Glad-Syllabub6777This week

AI Voice Platform Comparison for Small Business Use Cases

We provide AI voice agent consultation and solutions in Upwork. One of clients’ frequent questions is which platform is best/perfect for their use cases, like lead qualification, AI receptionist, customer support, etc. This post provides our thoughts on this question. Our overall feeling is that the AI agent technology is still not there yet. It seems close but there are many corner cases the AI bot doesn't handle well. Four major players in the AI voice platforms: Bland ($65M funding) Retell ($4.6M funding) Synthflow ($7.4M funding) VAPI ($20M funding) We will only talk about Bland, Retell and VAPI. We firstly tried Synthflow and found the UI was buggy (the prompt editor froze for 20 seconds to 30 seconds when we were editing the prompt). Currently we don't use it anymore. Recommended use cases based on Upwork jobs we delivered: Bland. We recommend Bland for lead qualification as the lead qualification has a strict conversation flow (like asking questions, extracting variables, and making webhook calls). Clients/contractors can draw flow diagrams to build AI voice agents. We also find Bland is not a good fit for a small business with a monthly budget less than 5K. The reason is that common tools (like warm transfer, SMS sending) for AI voice agents are only available to enterprise clients. But warm transfers are critical for small businesses. Retell. We recommend Retell for customer support in contact centers. Retell has the best voice among competitors. One use case we build in Retell is the live translator in the ambulance call center. We tried the same prompt with the same LLM setup in VAPI. We found Retell performs way better than VAPI in terms of the translation quality and reliability. Another common scenario in the customer support domain is to have 3-way merge so that the agent can tell the summary to the transfer number while the caller can hear the conversation. VAPI. We recommend VAPI for AI receptionists and phone answering use cases. We can write a prompt and ask LLM to do the magic if callers ask questions not included in the prompt. We can set up custom tools to trigger automation (like update CRM) and warm transfer to connect to the stakeholders. One feeling we have is that VAPI is way more complicated than the other two platforms. If you don’t have developer experience and have a budget to hire a contractor, it is better to try Retell as Retell has many integrations. If you have any other questions or we miss anything, feel free to comment. We like to explore AI voice agent space together.

I run an AI automation agency (AAA). My honest overview and review of this new business model
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

I run an AI automation agency (AAA). My honest overview and review of this new business model

I started an AI tools directory in February, and then branched off that to start an AI automation agency (AAA) in June. So far I've come across a lot of unsustainable "ideas" to make money with AI, but at the same time a few diamonds in the rough that aren't fully tapped into yet- especially the AAA model. Thought I'd share this post to shine light into this new business model and share some ways you could potentially start your own agency, or at the very least know who you are dealing with and how to pick and choose when you (inevitably) get bombarded with cold emails from them down the line. Foreword Running an AAA does NOT involve using AI tools directly to generate and sell content directly. That ship has sailed, and unless you are happy with $5 from Fiverr every month or so, it is not a real business model. Cry me a river but generating generic art with AI and slapping it onto a T-shirt to sell on Etsy won't make you a dime. At the same time, the AAA model will NOT require you to have a deep theoretical knowledge of AI, or any academic degree, as we are more so dealing with the practical applications of generative AI and how we can implement these into different workflows and tech-stacks, rather than building AI models from the ground up. Regardless of all that, common sense and a willingness to learn will help (a shit ton), as with anything. Keep in mind - this WILL involve work and motivation as well. The mindset that AI somehow means everything can be done for you on autopilot is not the right way to approach things. The common theme of businesses I've seen who have successfully implemented AI into their operations is the willingess to work with AI in a way that augments their existing operations, rather than flat out replace a worker or team. And this is exactly the train of thought you need when working with AI as a business model. However, as the field is relatively unsaturated and hype surrounding AI is still fresh for enterprises, right now is the prime time to start something new if generative AI interests you at all. With that being said, I'll be going over three of the most successful AI-adjacent businesses I've seen over this past year, in addition to some tips and resources to point you in the right direction. so.. WTF is an AI Automation Agency? The AI automation agency (or as some YouTubers have coined it, the AAA model) at its core involves creating custom AI solutions for businesses. I have over 1500 AI tools listed in my directory, however the feedback I've received from some enterprise users is that ready-made SaaS tools are too generic to meet their specific needs. Combine this with the fact virtually no smaller companies have the time or skills required to develop custom solutions right off the bat, and you have yourself real demand. I would say in practice, the AAA model is quite similar to Wordpress and even web dev agencies, with the major difference being all solutions you develop will incorporate key aspects of AI AND automation. Which brings me to my second point- JUST AI IS NOT ENOUGH. Rather than reducing the amount of time required to complete certain tasks, I've seen many AI agencies make the mistake of recommending and (trying to) sell solutions that more likely than not increase the workload of their clients. For example, if you were to make an internal tool that has AI answer questions based on their knowledge base, but this knowledge base has to be updated manually, this is creating unnecessary work. As such I think one of the key components of building successful AI solutions is incorporating the new (Generative AI/LLMs) with the old (programmtic automation- think Zapier, APIs, etc.). Finally, for this business model to be successful, ideally you should target a niche in which you have already worked and understand pain points and needs. Not only does this make it much easier to get calls booked with prospects, the solutions you build will have much greater value to your clients (meaning you get paid more). A mistake I've seen many AAA operators make (and I blame this on the "Get Rich Quick" YouTubers) is focusing too much on a specific productized service, rather than really understanding the needs of businesses. The former is much done via a SaaS model, but when going the agency route the only thing that makes sense is building custom solutions. This is why I always take a consultant-first approach. You can only build once you understand what they actually need and how certain solutions may impact their operations, workflows, and bottom-line. Basics of How to Get Started Pick a niche. As I mentioned previously, preferably one that you've worked in before. Niches I know of that are actively being bombarded with cold emails include real estate, e-commerce, auto-dealerships, lawyers, and medical offices. There is a reason for this, but I will tell you straight up this business model works well if you target any white-collar service business (internal tools approach) or high volume businesses (customer facing tools approach). Setup your toolbox. If you wanted to start a pressure washing business, you would need a pressure-washer. This is no different. For those without programming knowledge, I've seen two common ways AAA get setup to build- one is having a network of on-call web developers, whether its personal contacts or simply going to Upwork or any talent sourcing agency. The second is having an arsenal of no-code tools. I'll get to this more in a second, but this works beecause at its core, when we are dealing with the practical applications of AI, the code is quite simple, simply put. Start cold sales. Unless you have a network already, this is not a step you can skip. You've already picked a niche, so all you have to do is find the right message. Keep cold emails short, sweet, but enticing- and it will help a lot if you did step 1 correctly and intimately understand who your audience is. I'll be touching base later about how you can leverage AI yourself to help you with outreach and closing. The beauty of gen AI and the AAA model You don't need to be a seasoned web developer to make this business model work. The large majority of solutions that SME clients want is best done using an API for an LLM for the actual AI aspect. The value we create with the solutions we build comes with the conceptual framework and design that not only does what they need it to but integrates smoothly with their existing tech-stack and workflow. The actual implementation is quite straightforward once you understand the high level design and know which tools you are going to use. To give you a sense, even if you plan to build out these apps yourself (say in Python) the large majority of the nitty gritty technical work has already been done for you, especially if you leverage Python libraries and packages that offer high level abstraction for LLM-related functions. For instance, calling GPT can be as little as a single line of code. (And there are no-code tools where these functions are simply an icon on a GUI). Aside from understanding the capabilities and limitations of these tools and frameworks, the only thing that matters is being able to put them in a way that makes sense for what you want to build. Which is why outsourcing and no-code tools both work in our case. Okay... but how TF am I suppposed to actually build out these solutions? Now the fun part. I highly recommend getting familiar with Langchain and LlamaIndex. Both are Python libraires that help a lot with the high-level LLM abstraction I mentioned previously. The two most important aspects include being able to integrate internal data sources/knowledge bases with LLMs, and have LLMs perform autonomous actions. The two most common methods respectively are RAG and output parsing. RAG (retrieval augmented Generation) If you've ever seen a tool that seemingly "trains" GPT on your own data, and wonder how it all works- well I have an answer from you. At a high level, the user query is first being fed to what's called a vector database to run vector search. Vector search basically lets you do semantic search where you are searching data based on meaning. The vector databases then retrieves the most relevant sections of text as it relates to the user query, and this text gets APPENDED to your GPT prompt to provide extra context to the AI. Further, with prompt engineering, you can limit GPT to only generate an answer if it can be found within this extra context, greatly limiting the chance of hallucination (this is where AI makes random shit up). Aside from vector databases, we can also implement RAG with other data sources and retrieval methods, for example SQL databses (via parsing the outputs of LLM's- more on this later). Autonomous Agents via Output Parsing A common need of clients has been having AI actually perform tasks, rather than simply spitting out text. For example, with autonomous agents, we can have an e-commerce chatbot do the work of a basic customer service rep (i.e. look into orders, refunds, shipping). At a high level, what's going on is that the response of the LLM is being used programmtically to determine which API to call. Keeping on with the e-commerce example, if I wanted a chatbot to check shipping status, I could have a LLM response within my app (not shown to the user) with a prompt that outputs a random hash or string, and programmatically I can determine which API call to make based on this hash/string. And using the same fundamental concept as with RAG, I can append the the API response to a final prompt that would spit out the answer for the user. How No Code Tools Can Fit In (With some example solutions you can build) With that being said, you don't necessarily need to do all of the above by coding yourself, with Python libraries or otherwise. However, I will say that having that high level overview will help IMMENSELY when it comes to using no-code tools to do the actual work for you. Regardless, here are a few common solutions you might build for clients as well as some no-code tools you can use to build them out. Ex. Solution 1: AI Chatbots for SMEs (Small and Medium Enterprises) This involves creating chatbots that handle user queries, lead gen, and so forth with AI, and will use the principles of RAG at heart. After getting the required data from your client (i.e. product catalogues, previous support tickets, FAQ, internal documentation), you upload this into your knowledge base and write a prompt that makes sense for your use case. One no-code tool that does this well is MyAskAI. The beauty of it especially for building external chatbots is the ability to quickly ingest entire websites into your knowledge base via a sitemap, and bulk uploading files. Essentially, they've covered the entire grunt work required to do this manually. Finally, you can create a inline or chat widget on your client's website with a few lines of HTML, or altneratively integrate it with a Slack/Teams chatbot (if you are going for an internal Q&A chatbot approach). Other tools you could use include Botpress and Voiceflow, however these are less for RAG and more for building out complete chatbot flows that may or may not incorporate LLMs. Both apps are essentially GUIs that eliminate the pain and tears and trying to implement complex flows manually, and both natively incoporate AI intents and a knowledge base feature. Ex. Solution 2: Internal Apps Similar to the first example, except we go beyond making just chatbots but tools such as report generation and really any sort of internal tool or automations that may incorporate LLM's. For instance, you can have a tool that automatically generates replies to inbound emails based on your client's knowledge base. Or an automation that does the same thing but for replies to Instagram comments. Another example could be a tool that generates a description and screeenshot based on a URL (useful for directory sites, made one for my own :P). Getting into more advanced implementations of LLMs, we can have tools that can generate entire drafts of reports (think 80+ pages), based not only on data from a knowledge base but also the writing style, format, and author voice of previous reports. One good tool to create content generation panels for your clients would be MindStudio. You can train LLM's via prompt engineering in a structured way with your own data to essentially fine tune them for whatever text you need it to generate. Furthermore, it has a GUI where you can dictate the entire AI flow. You can also upload data sources via multiple formats, including PDF, CSV, and Docx. For automations that require interactions between multiple apps, I recommend the OG zapier/make.com if you want a no-code solution. For instance, for the automatic email reply generator, I can have a trigger such that when an email is received, a custom AI reply is generated by MyAskAI, and finally a draft is created in my email client. Or, for an automation where I can create a social media posts on multiple platforms based on a RSS feed (news feed), I can implement this directly in Zapier with their native GPT action (see screenshot) As for more complex LLM flows that may require multiple layers of LLMs, data sources, and APIs working together to generate a single response i.e. a long form 100 page report, I would recommend tools such as Stack AI or Flowise (open-source alternative) to build these solutions out. Essentially, you get most of the functions and features of Python packages such as Langchain and LlamaIndex in a GUI. See screenshot for an example of a flow How the hell are you supposed to find clients? With all that being said, none of this matters if you can't find anyone to sell to. You will have to do cold sales, one way or the other, especially if you are brand new to the game. And what better way to sell your AI services than with AI itself? If we want to integrate AI into the cold outreach process, first we must identify what it's good at doing, and that's obviously writing a bunch of text, in a short amount of time. Similar to the solutions that an AAA can build for its clients, we can take advantage of the same principles in our own sales processes. How to do outreach Once you've identified your niche and their pain points/opportunities for automation, you want to craft a compelling message in which you can send via cold email and cold calls to get prospects booked on demos/consultations. I won't get into too much detail in terms of exactly how to write emails or calling scripts, as there are millions of resources to help with this, but I will tell you a few key points you want to keep in mind when doing outreach for your AAA. First, you want to keep in mind that many businesses are still hesitant about AI and may not understand what it really is or how it can benefit their operations. However, we can take advantage of how mass media has been reporting on AI this past year- at the very least people are AWARE that sooner or later they may have to implement AI into their businesses to stay competitive. We want to frame our message in a way that introduces generative AI as a technology that can have a direct, tangible, and positive impact on their business. Although it may be hard to quantify, I like to include estimates of man-hours saved or costs saved at least in my final proposals to prospects. Times are TOUGH right now, and money is expensive, so you need to have a compelling reason for businesses to get on board. Once you've gotten your messaging down, you will want to create a list of prospects to contact. Tools you can use to find prospects include Apollo.io, reply.io, zoominfo (expensive af), and Linkedin Sales Navigator. What specific job titles, etc. to target will depend on your niche but for smaller companies this will tend to be the owner. For white collar niches, i.e. law, the professional that will be directly benefiting from the tool (i.e. partners) may be better to contact. And for larger organizations you may want to target business improvement and digital transformation leads/directors- these are the people directly in charge of projects like what you may be proposing. Okay- so you have your message, and your list, and now all it comes down to is getting the good word out. I won't be going into the details of how to send these out, a quick Google search will give you hundreds of resources for cold outreach methods. However, personalization is key and beyond simple dynamic variables you want to make sure you can either personalize your email campaigns directly with AI (SmartWriter.ai is an example of a tool that can do this), or at the very least have the ability to import email messages programmatically. Alternatively, ask ChatGPT to make you a Python Script that can take in a list of emails, scrape info based on their linkedin URL or website, and all pass this onto a GPT prompt that specifies your messaging to generate an email. From there, send away. How tf do I close? Once you've got some prospects booked in on your meetings, you will need to close deals with them to turn them into clients. Call #1: Consultation Tying back to when I mentioned you want to take a consultant-first appraoch, you will want to listen closely to their goals and needs and understand their pain points. This would be the first call, and typically I would provide a high level overview of different solutions we could build to tacke these. It really helps to have a presentation available, so you can graphically demonstrate key points and key technologies. I like to use Plus AI for this, it's basically a Google Slides add-on that can generate slide decks for you. I copy and paste my default company messaging, add some key points for the presentation, and it comes out with pretty decent slides. Call #2: Demo The second call would involve a demo of one of these solutions, and typically I'll quickly prototype it with boilerplate code I already have, otherwise I'll cook something up in a no-code tool. If you have a niche where one type of solution is commonly demanded, it helps to have a general demo set up to be able to handle a larger volume of calls, so you aren't burning yourself out. I'll also elaborate on how the final product would look like in comparison to the demo. Call #3 and Beyond: Once the initial consultation and demo is complete, you will want to alleviate any remaining concerns from your prospects and work with them to reach a final work proposal. It's crucial you lay out exactly what you will be building (in writing) and ensure the prospect understands this. Furthermore, be clear and transparent with timelines and communication methods for the project. In terms of pricing, you want to take this from a value-based approach. The same solution may be worth a lot more to client A than client B. Furthermore, you can create "add-ons" such as monthly maintenance/upgrade packages, training sessions for employeees, and so forth, separate from the initial setup fee you would charge. How you can incorporate AI into marketing your businesses Beyond cold sales, I highly recommend creating a funnel to capture warm leads. For instance, I do this currently with my AI tools directory, which links directly to my AI agency and has consistent branding throughout. Warm leads are much more likely to close (and honestly, much nicer to deal with). However, even without an AI-related website, at the very least you will want to create a presence on social media and the web in general. As with any agency, you will want basic a professional presence. A professional virtual address helps, in addition to a Google Business Profile (GBP) and TrustPilot. a GBP (especially for local SEO) and Trustpilot page also helps improve the looks of your search results immensely. For GBP, I recommend using ProfilePro, which is a chrome extension you can use to automate SEO work for your GBP. Aside from SEO optimzied business descriptions based on your business, it can handle Q/A answers, responses, updates, and service descriptions based on local keywords. Privacy and Legal Concerns of the AAA Model Aside from typical concerns for agencies relating to service contracts, there are a few issues (especially when using no-code tools) that will need to be addressed to run a successful AAA. Most of these surround privacy concerns when working with proprietary data. In your terms with your client, you will want to clearly define hosting providers and any third party tools you will be using to build their solution, and a DPA with these third parties listed as subprocessors if necessary. In addition, you will want to implement best practices like redacting private information from data being used for building solutions. In terms of addressing concerns directly from clients, it helps if you host your solutions on their own servers (not possible with AI tools), and address the fact only ChatGPT queries in the web app, not OpenAI API calls, will be used to train OpenAI's models (as reported by mainstream media). The key here is to be open and transparent with your clients about ALL the tools you are using, where there data will be going, and make sure to get this all in writing. have fun, and keep an open mind Before I finish this post, I just want to reiterate the fact that this is NOT an easy way to make money. Running an AI agency will require hours and hours of dedication and work, and constantly rearranging your schedule to meet prospect and client needs. However, if you are looking for a new business to run, and have a knack for understanding business operations and are genuinely interested in the pracitcal applications of generative AI, then I say go for it. The time is ticking before AAA becomes the new dropshipping or SMMA, and I've a firm believer that those who set foot first and establish themselves in this field will come out top. And remember, while 100 thousand people may read this post, only 2 may actually take initiative and start.

Watched 8 hours of MrBeast's content. Here are 7 psychological strategies he's used to get 34 billion views
reddit
LLM Vibe Score0
Human Vibe Score1
Positive-Bison5023This week

Watched 8 hours of MrBeast's content. Here are 7 psychological strategies he's used to get 34 billion views

MrBeast can fill giant stadiums and launch 8-figure candy companies on demand. He’s unbelievably popular. Recently, I listened to the brilliant marketer Phill Agnew (from The Nudge podcast) being interviewed on the Creator Science podcast. The episode focused on how MrBeast’s near-academic understanding of audience psychology is the key to his success. Better than anyone, MrBeast knows how to get you: \- Click on his content (increase his click-through rate) \- Get you to stick around (increase his retention rate) He gets you to click by using irresistible thumbnails and headlines. I watched 8 hours of his content. To build upon Phil Agnew’s work, I made a list of 7 psychological effects and biases he’s consistently used to write headlines that get clicked into oblivion. Even the most aggressively “anti-clickbait” purists out there would benefit from learning the psychology of why people choose to click on some content over others. Ultimately, if you don’t get the click, it really doesn’t matter how good your content is. Novelty Effect MrBeast Headline: “I Put 100 Million Orbeez In My Friend's Backyard” MrBeast often presents something so out of the ordinary that they have no choice but to click and find out more. That’s the “novelty effect” at play. Our brain’s reward system is engaged when we encounter something new. You’ll notice that the headline examples you see in this list are extreme. MrBeast takes things to the extreme. You don’t have to. Here’s your takeaway: Consider breaking the reader/viewer’s scrolling pattern by adding some novelty to your headlines. How? Here are two ways: Find the unique angle in your content Find an unusual character in your content Examples: “How Moonlight Walks Skyrocketed My Productivity”. “Meet the Artist Who Paints With Wine and Chocolate.” Headlines like these catch the eye without requiring 100 million Orbeez. Costly Signaling MrBeast Headline: "Last To Leave $800,000 Island Keeps It" Here’s the 3-step click-through process at play here: MrBeast lets you know he’s invested a very significant amount of time and money into his content. This signals to whoever reads the headline that it's probably valuable and worth their time. They click to find out more. Costly signaling is all amount showcasing what you’ve invested into the content. The higher the stakes, the more valuable the content will seem. In this example, the $800,000 island he’s giving away just screams “This is worth your time!” Again, they don’t need to be this extreme. Here are two examples with a little more subtlety: “I built a full-scale botanical garden in my backyard”. “I used only vintage cookware from the 1800s for a week”. Not too extreme, but not too subtle either. Numerical Precision MrBeast knows that using precise numbers in headlines just work. Almost all of his most popular videos use headlines that contain a specific number. “Going Through The Same Drive Thru 1,000 Times" “$456,000 Squid Game In Real Life!” Yes, these headlines also use costly signaling. But there’s more to it than that. Precise numbers are tangible. They catch our eye, pique our curiosity, and add a sense of authenticity. “The concreteness effect”: Specific, concrete information is more likely to be remembered than abstract, intangible information. “I went through the same drive thru 1000 times” is more impactful than “I went through the same drive thru countless times”. Contrast MrBeast Headline: "$1 vs $1,000,000 Hotel Room!" Our brains are drawn to stark contrasts and MrBeast knows it. His headlines often pit two extremes against each other. It instantly creates a mental image of both scenarios. You’re not just curious about what a $1,000,000 hotel room looks like. You’re also wondering how it could possibly compare to a $1 room. Was the difference wildly significant? Was it actually not as significant as you’d think? It increases the audience’s \curiosity gap\ enough to get them to click and find out more. Here are a few ways you could use contrast in your headlines effectively: Transformational Content: "From $200 to a $100M Empire - How A Small Town Accountant Took On Silicon Valley" Here you’re contrasting different states or conditions of a single subject. Transformation stories and before-and-after scenarios. You’ve got the added benefit of people being drawn to aspirational/inspirational stories. Direct Comparison “Local Diner Vs Gourmet Bistro - Where Does The Best Comfort Food Lie?” Nostalgia MrBeast Headline: "I Built Willy Wonka's Chocolate Factory!" Nostalgia is a longing for the past. It’s often triggered by sensory stimuli - smells, songs, images, etc. It can feel comforting and positive, but sometimes bittersweet. Nostalgia can provide emotional comfort, identity reinforcement, and even social connection. People are drawn to it and MrBeast has it down to a tee. He created a fantasy world most people on this planet came across at some point in their childhood. While the headline does play on costly signaling here as well, nostalgia does help to clinch the click and get the view. Subtle examples of nostalgia at play: “How this \[old school cartoon\] is shaping new age animation”. “\[Your favorite childhood books\] are getting major movie deals”. Morbid Curiosity MrBeast Headline: "Surviving 24 Hours Straight In The Bermuda Triangle" People are drawn to the macabre and the dangerous. Morbid curiosity explains why you’re drawn to situations that are disturbing, frightening, or gruesome. It’s that tension between wanting to avoid harm and the irresistible desire to know about it. It’s a peculiar aspect of human psychology and viral content marketers take full advantage of it. The Bermuda Triangle is practically synonymous with danger. The headline suggests a pretty extreme encounter with it, so we click to find out more. FOMO And Urgency MrBeast Headline: "Last To Leave $800,000 Island Keeps It" “FOMO”: the worry that others may be having fulfilling experiences that you’re absent from. Marketers leverage FOMO to drive immediate action - clicking, subscribing, purchasing, etc. The action is driven by the notion that delay could result in missing out on an exciting opportunity or event. You could argue that MrBeast uses FOMO and urgency in all of his headlines. They work under the notion that a delay in clicking could result in missing out on an exciting opportunity or event. MrBeast’s time-sensitive challenge, exclusive opportunities, and high-stakes competitions all generate a sense of urgency. People feel compelled to watch immediately for fear of missing out on the outcome or being left behind in conversations about the content. Creators, writers, and marketers can tap into FOMO with their headlines without being so extreme. “The Hidden Parisian Cafe To Visit Before The Crowds Do” “How \[Tech Innovation\] Will Soon Change \[Industry\] For Good” (Yep, FOMO and urgency are primarily responsible for the proliferation of AI-related headlines these days). Why This All Matters If you don’t have content you need people to consume, it probably doesn’t! But if any aspect of your online business would benefit from people clicking on things more, it probably does. “Yes, because we all need more clickbait in this world - \eye-roll emoji\” - Disgruntled Redditor I never really understood this comment but I seem to get it pretty often. My stance is this: If the content delivers what the headline promises, it shouldn’t be labeled clickbait. I wouldn’t call MrBeast’s content clickbait. The fact is that linguistic techniques can be used to drive people to consume some content over others. You don’t need to take things to the extremes that MrBeast does to make use of his headline techniques. If content doesn’t get clicked, it won’t be read, viewed, or listened to - no matter how brilliant the content might be. While “clickbait” content isn’t a good thing, we can all learn a thing or two from how they generate attention in an increasingly noisy digital world.

AI Voice Platform Comparison for Small Business Use Cases
reddit
LLM Vibe Score0
Human Vibe Score1
Glad-Syllabub6777This week

AI Voice Platform Comparison for Small Business Use Cases

We provide AI voice agent consultation and solutions in Upwork. One of clients’ frequent questions is which platform is best/perfect for their use cases, like lead qualification, AI receptionist, customer support, etc. This post provides our thoughts on this question. Our overall feeling is that the AI agent technology is still not there yet. It seems close but there are many corner cases the AI bot doesn't handle well. Four major players in the AI voice platforms: Bland ($65M funding) Retell ($4.6M funding) Synthflow ($7.4M funding) VAPI ($20M funding) We will only talk about Bland, Retell and VAPI. We firstly tried Synthflow and found the UI was buggy (the prompt editor froze for 20 seconds to 30 seconds when we were editing the prompt). Currently we don't use it anymore. Recommended use cases based on Upwork jobs we delivered: Bland. We recommend Bland for lead qualification as the lead qualification has a strict conversation flow (like asking questions, extracting variables, and making webhook calls). Clients/contractors can draw flow diagrams to build AI voice agents. We also find Bland is not a good fit for a small business with a monthly budget less than 5K. The reason is that common tools (like warm transfer, SMS sending) for AI voice agents are only available to enterprise clients. But warm transfers are critical for small businesses. Retell. We recommend Retell for customer support in contact centers. Retell has the best voice among competitors. One use case we build in Retell is the live translator in the ambulance call center. We tried the same prompt with the same LLM setup in VAPI. We found Retell performs way better than VAPI in terms of the translation quality and reliability. Another common scenario in the customer support domain is to have 3-way merge so that the agent can tell the summary to the transfer number while the caller can hear the conversation. VAPI. We recommend VAPI for AI receptionists and phone answering use cases. We can write a prompt and ask LLM to do the magic if callers ask questions not included in the prompt. We can set up custom tools to trigger automation (like update CRM) and warm transfer to connect to the stakeholders. One feeling we have is that VAPI is way more complicated than the other two platforms. If you don’t have developer experience and have a budget to hire a contractor, it is better to try Retell as Retell has many integrations. If you have any other questions or we miss anything, feel free to comment. We like to explore AI voice agent space together.

Partnership revenue share uncertainty as test before any equity discussions, please help, urgent
reddit
LLM Vibe Score0
Human Vibe Score1
jayn35This week

Partnership revenue share uncertainty as test before any equity discussions, please help, urgent

Hi all, It's brand new relationship to collaborate on work and fast moving situation and i want to be fair and informed about revenue share for this work as startup, new agency, unclear still. Sorry for rushed message, its moving fast. Its starting with revenue share to test and see how things go. I contribute some things as a separate entity/consultant/marketing domain expert who designed some AI products and am able to acquire clients reliably with my marketing skills then they do all development and sales assistance. Details below please can you help with advice on contribution and revenue share thats very fair: The "partnership" non ownership (rev share is best correct?) of delivering custom AI software development solutions to smb b2b clients. As a domain expert i designed a product for myself and then others upsells and want to sell it to other biz, there is interest, its been tested as viable with my outreach which I do and now have 5 clients from last night wanting to meet or receive short video explanations before we meet (its my initial offer, a vid demo). I have designed the product or solution completely and have already developed mvp of the first product that i use myself and is immensely valuable to me. I also acquire all the clients as an client outreach/acquisition expert and perform that entire client acquisition function and marketing up until sales call where they provide assistance/ a joint tech and marketing/product domain specialist (me) sales call, still to be discussed. No dedicated sales function but they have experience. Then I partner with a great desirable professional development agency to deploy the solution and everything that entails hoping for a long-term similar arrangement that mutually beneficial and fair. They also assist with the sales process to close deals, we both contribute on the sales calls but client generation and marketing up to the sales call is my contribution. What would the fair revenue share be in a perfectly fair equal situation and what would it be if I wanted to be generous because i really want to work with them moving forward. Also what would the equity split be if a new entity was formed later to formalize partnership and the contribution remained the same. I dont know much about this or what I should be doing in my situation. As I understand searching revenue share online and a summary from perplexity I perform two of the major functions and they one so something like 30-40 them and the rest me? But if i wanted to be generous and show my appreciation for working with me on this as they are high quality and i foresee more opportunity benefits and capabilities in the future due to their expertise and know they would deliver a superb job, would 50/50 be a fair split? Or am I undervaluing/overvaluing myself,, can you not just offer the logic but advice as well based on the info you have, this is brand new and moving super fast, online info seems clear but i want mine to be super fair even generous for them so they are happy, but also not foolish or irresponsible from my side. Its all new to me. Thank you so much!

Looking for a Developer Co-Founder to Build an AI-Powered Film Budgeting Tool
reddit
LLM Vibe Score0
Human Vibe Score1
Boring_Elephant2767This week

Looking for a Developer Co-Founder to Build an AI-Powered Film Budgeting Tool

Hey everyone, I’m a seasoned producer/line producer with over 10 years in the film industry, specializing in budgeting and production strategy for films, commercials, and music videos. I’ve built over 150 budgets for projects ranging from indie features to large-scale commercials and have worked with major artists, brands, and studios. I’m looking for a developer or AI/ML engineer interested in co-founding a startup with me to build an AI-powered budgeting tool for the film industry. The Problem Creating a budget for a film, music video, or commercial is time-consuming and expensive (typically $3K–$5K per budget for films). Filmmakers, studios, agencies, and managers need a faster, more cost-effective way to estimate production costs without hiring a full-time producer for every project. The Solution The goal is to develop an AI-assisted budgeting tool that takes in scripts, creative decks, or project briefs and outputs a preliminary budget & production schedule. The vision is a hybrid service: • AI-powered script/deck breakdown to extract production elements • Smart reasoning based on real industry budgets • Producer oversight for accuracy before sending budgets to users • Flexible pricing model (lower cost than hiring a full-time producer) What I Bring to the Table Deep industry knowledge – I know how to build accurate budgets & schedules for any type of project. Proven demand – I already have early adopters in indie film, production companies, and agencies. Strong network – I work with studios, reps, and filmmakers who would use this tool. A unique approach – I haven’t seen an AI budgeting tool that truly understands production costs based on creative elements. What I’m Looking For I need a developer partner with experience in AI, automation, and/or SaaS development who can help build this. Ideally, someone interested in co-founding (equity-based, not just a freelance gig). If you have experience with GPT, machine learning, NLP, or building interactive SaaS products, that’s a plus. I’m keeping this low-key for now while I figure out the best path forward. If you’re interested, let’s chat! Even if you’re not a developer but have advice or ideas, I’d love to hear your thoughts. Drop a comment or DM me if this sounds interesting!

The best (actually free to use) AI tools for day-to-day work + productivity
reddit
LLM Vibe Score0
Human Vibe Score0.917
Tapedulema919This week

The best (actually free to use) AI tools for day-to-day work + productivity

I've spent an ungodly amount of time ~~procrastinating~~ trying tons of new/free AI tools from Reddit and various lists of the best AI tools for different use cases. Frankly, most free AI tools (and even paid ones) are gimmicky ChatGPT wrappers with questionable utility in everyday tasks or overpriced enterprise software that don't use AI as anything more than a marketing buzzword. My last list of free AI tools got a good response here, and I wanted to make another with the best AI tools that I actually use day-to-day now that I've spent more time with them. All these tools can be used for free, though most of them have some kind of premium offering if you need more advanced stuff or a ton of queries. To make it easy to sort through, I've also added whether each tool requires signup. ChatPDF: Free Tool to Use ChatGPT on Your Own Documents/PDFs (free no signup) Put simply, ChatPDF lets you upload any PDF and interact with it like ChatGPT. I heard about this one from my nephew who used it to automatically generate flashcards and explain concepts based on class notes and readings. There are a few similar services out there, but I found ChatPDF the easiest to use of those that don't require payment/signup. If you're a student or someone who needs to read through long PDFs regularly, the possibilities to use this are endless. It's also completely free and doesn't require signup. Key Features: Free to upload up to 3 PDFs daily, with up to 120 pages in each PDF Can be used without signing up at all Taskade: AI Task Management, Scheduling, and Notetaking Tool with GPT-4 Built-In (free with signup) Taskade is an all-in-one notetaking, task management, and scheduling platform with built-in AI workflows and templates. Like Notion, Taskade lets you easily create workspaces, documents, and templates for your workflows. Unlike Notion’s GPT-3 based AI, Taskade has built-in GPT-4 based AI that’s trained to structure your documents, create content, and otherwise help you improve your productivity. Key Features: GPT-4 is built in to their free plan and trained to help with document formatting, scheduling, content creation and answering questions through a chat interface. Its AI seems specifically trained to work seamlessly with your documents and workspaces, and understands queries specific to their interface like asking it to turn (text) notes into a mind map. One of the highest usage limits of the free tools: Taskade’s free plan comes with 1000 monthly requests, which is one of the highest I’ve seen for a tool with built-in GPT-4. Because it’s built into a document editor with database, scheduling and chat capabilities, you can use it for pretty much anything you’d use ChatGPT for but without* paying for ChatGPT Premium. Free templates to get you started with actually integrating AI into your workflows: there are a huge number of genuinely useful free templates for workflows, task management, mind mapping, etc. For example, you can add a project and have Taskade automatically map out and schedule a breakdown of the tasks that make up that overall deliverable. Plus AI for Google Slides: AI-generated (and improved) slide decks (free with signup, addon for Google Slides) I've tried out a bunch of AI presentation/slide generating tools. To be honest, most of them leave a lot to be desired and aren't genuinely useful unless you're literally paid to generate a presentation vaguely related to some topic. Plus AI is a (free!) Google Slides addon that lets you describe the kind of slide deck you're making, then generate and fine-tune it based on your exact needs. It's still not at the point where you can literally just tell it one prompt and get the entire finished product, but it saves a bunch of time getting an initial structure together that you can then perfect. Similarly, if you have existing slides made you can tell it (in natural language) how you want it changed. For example, asking it to change up the layout of text on a page, improve the writing style, or even use external data sources. Key Features: Integrates seamlessly into Google Slides: if you’re already using Slides, using Plus AI is as simple as installing the plugin. Their tutorials are easy to follow and it doesn’t require learning some new slideshow software or interface like some other options. Create and* tweak slides using natural language: Plus AI lets you create whole slideshows, adjust text, or change layouts using natural language. It’s all fairly intuitive and the best of the AI slide tools I’ve tried. FlowGPT: Database of AI prompts and workflows (free without signup-though it pushes you to signup!) FlowGPT collects prompts and collections of prompts to do various tasks, from marketing, productivity, and coding to random stuff people find interesting. It uses an upvote system similar to Reddit that makes it easy to find interesting ways to use ChatGPT. It also lets you search for prompts if you have something in mind and want to see what others have done. It's free and has a lot of cool features like showing you previews of how ChatGPT responds to the prompts. Unfortunately, it's also a bit pushy with getting you to signup, and the design leaves something to be desired, but it's the best of these tools I've found. Key Features: Lots of users that share genuinely useful and interesting prompts Upvote system similar to Reddit’s that allows you to find interesting prompts within the categories you’re interested in Summarize.Tech: AI summaries of YouTube Videos (free no signup) Summarize generates AI summaries of YouTube videos, condensing them into relatively short written notes with timestamps. All the summaries I've seen have been accurate and save significant time. I find it especially useful when looking at longer tutorials where I want to find if: ​ The tutorial actually tells me what I'm looking for, and See where in the video I can find that specific part. The one downside I've seen is that it doesn't work for videos that don't have subtitles, but hopefully, someone can build something with Whisper or a similar audio transcription API to solve that. Claude: ChatGPT Alternative with ~75k Word Limit (free with signup) If you've used ChatGPT, you've probably run into the issue of its (relatively low) token limit. Put simply, it can't handle text longer than a few thousand words. It's the same reason why ChatGPT "forgets" instructions you gave it earlier on in a conversation. Claude solves that, with a \~75,000 word limit that lets you input literal novels and do pretty much everything you can do with ChatGPT. Unfortunately, Claude is currently only free in the US or UK. Claude pitches itself as the "safer" AI, which can make it a pain to use for many use cases, but it's worth trying out and better than ChatGPT for certain tasks. Currently, I'm mainly using it to summarize long documents that ChatGPT literally cannot process as a single prompt. Key Features: Much longer word limit than even ChatGPT’s highest token models Stronger guardrails than ChatGPT: if you're into this, Claude focuses a lot more on "trust and safety" than even ChatGPT does. While an AI telling me what information I can and can't have is more of an annoyance for my use cases, it can be useful if you're building apps like customer support or other use cases where it's a top priority to keep the AI from writing something "surprising." Phind: AI Search Engine That Combines Google with ChatGPT (free no signup) Like a combination of Google and ChatGPT. Like ChatGPT, it can understand complex prompts and give you detailed answers condensing multiple sources. Like Google, it shows you the most up-to-date sources answering your question and has access to everything on the internet in real time (vs. ChatGPT's September 2021 cutoff). Unlike Google, it avoids spammy links that seem to dominate Google nowadays and actually answers your question. Key Features: Accesses the internet to get you real-time information vs. ChatGPT’s 2021 cutoff. While ChatGPT is great for content generation and other tasks that you don’t really need live information for, it can’t get you any information from past its cutoff point. Provides actual sources for its claims, helping you dive deeper into any specific points and avoid hallucinations. Phind was the first to combine the best of both worlds between Google and ChatGPT, giving you easy access to actual sources the way Google does while summarizing relevant results the way ChatGPT does. It’s still one of the best places for that, especially if you have technical questions. Bing AI: ChatGPT Alternative Based on GPT-4 (with internet access!) (free no signup) For all the hate Bing gets, they've done the best job of all the major search engines of integrating AI chat to answer questions. Bing's Chat AI is very similar to ChatGPT (it's based on GPT-4). Unlike ChatGPT's base model without plugins, it has access to the internet. It also doesn't require signing in, which is nice. At the risk of sounding like a broken record, Google has really dropped the ball lately in delivering non-spammy search results that actually answer the query, and it's nice to see other search engines like Bing and Phind providing alternatives. Key Features: Similar to Phind, though arguably a bit better for non-technical questions: Bing similarly provides sourced summaries, generates content and otherwise integrates AI and search nicely. Built on top of GPT-4: like Taskade, Bing has confirmed they use GPT-4. That makes it another nice option to get around paying for GPT-4 while still getting much of the same capabilities as ChatGPT. Seamless integration with a standard search engine that’s much better than I remember it being (when it was more of a joke than anything) Honorable Mentions: These are the “rest of the best” free AI tools I've found that are simpler/don't need a whole entry to explain: PdfGPT: Alternative to ChatPDF that also uses AI to summarize and let you interact with PDF documents. Nice to have options if you run into one site’s PDF or page limit and don’t want to pay to do so. Remove.bg: One of the few image AI tools I use regularly. Remove.bg uses simple AI to remove backgrounds from your images. It's very simple, but something I end up doing surprisingly often editing product images, etc. CopyAI and Jasper: both are AI writing tools primarily built for website marketing/blog content. I've tried both but don't use them enough regularly to be able to recommend one over the other. Worth trying if you do a lot of content writing and want to automate parts of it. Let me know if you guys recommend any other free AI tools that you use day-to-day and I can add them to the list. I’m also interested in any requests you guys have for AI tools that don’t exist yet, as I’m looking for new projects to work on at the moment! TL;DR: ChatPDF: Interact with any PDF using ChatGPT without signing up, great for students and anyone who needs to filter through long PDFs. Taskade: All-in-one task management, scheduling, and notetaking with built-in GPT-4 Chat + AI assistant for improving productivity. Plus AI for Google Slides: Addon for Google Slides that generates and fine-tunes slide decks based on your description(s) in natural language. FlowGPT: Database of AI prompts and workflows. Nice resource to find interesting ChatGPT prompts. Summarize.Tech: AI summaries of YouTube videos with timestamps that makes it easier to find relevant information in longer videos. Claude: ChatGPT alternative with a \~75k word limit, ideal for handling long documents and tasks that go above ChatGPT's token limit. Phind: AI search engine similar to a combination of Google and ChatGPT. Built in internet access and links/citations for its claims. Bing AI: Bing's ChatGPT alternative based on GPT-4. Has real-time internet access + integrates nicely with their normal search engine.

I run an AI automation agency (AAA). My honest overview and review of this new business model
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

I run an AI automation agency (AAA). My honest overview and review of this new business model

I started an AI tools directory in February, and then branched off that to start an AI automation agency (AAA) in June. So far I've come across a lot of unsustainable "ideas" to make money with AI, but at the same time a few diamonds in the rough that aren't fully tapped into yet- especially the AAA model. Thought I'd share this post to shine light into this new business model and share some ways you could potentially start your own agency, or at the very least know who you are dealing with and how to pick and choose when you (inevitably) get bombarded with cold emails from them down the line. Foreword Running an AAA does NOT involve using AI tools directly to generate and sell content directly. That ship has sailed, and unless you are happy with $5 from Fiverr every month or so, it is not a real business model. Cry me a river but generating generic art with AI and slapping it onto a T-shirt to sell on Etsy won't make you a dime. At the same time, the AAA model will NOT require you to have a deep theoretical knowledge of AI, or any academic degree, as we are more so dealing with the practical applications of generative AI and how we can implement these into different workflows and tech-stacks, rather than building AI models from the ground up. Regardless of all that, common sense and a willingness to learn will help (a shit ton), as with anything. Keep in mind - this WILL involve work and motivation as well. The mindset that AI somehow means everything can be done for you on autopilot is not the right way to approach things. The common theme of businesses I've seen who have successfully implemented AI into their operations is the willingess to work with AI in a way that augments their existing operations, rather than flat out replace a worker or team. And this is exactly the train of thought you need when working with AI as a business model. However, as the field is relatively unsaturated and hype surrounding AI is still fresh for enterprises, right now is the prime time to start something new if generative AI interests you at all. With that being said, I'll be going over three of the most successful AI-adjacent businesses I've seen over this past year, in addition to some tips and resources to point you in the right direction. so.. WTF is an AI Automation Agency? The AI automation agency (or as some YouTubers have coined it, the AAA model) at its core involves creating custom AI solutions for businesses. I have over 1500 AI tools listed in my directory, however the feedback I've received from some enterprise users is that ready-made SaaS tools are too generic to meet their specific needs. Combine this with the fact virtually no smaller companies have the time or skills required to develop custom solutions right off the bat, and you have yourself real demand. I would say in practice, the AAA model is quite similar to Wordpress and even web dev agencies, with the major difference being all solutions you develop will incorporate key aspects of AI AND automation. Which brings me to my second point- JUST AI IS NOT ENOUGH. Rather than reducing the amount of time required to complete certain tasks, I've seen many AI agencies make the mistake of recommending and (trying to) sell solutions that more likely than not increase the workload of their clients. For example, if you were to make an internal tool that has AI answer questions based on their knowledge base, but this knowledge base has to be updated manually, this is creating unnecessary work. As such I think one of the key components of building successful AI solutions is incorporating the new (Generative AI/LLMs) with the old (programmtic automation- think Zapier, APIs, etc.). Finally, for this business model to be successful, ideally you should target a niche in which you have already worked and understand pain points and needs. Not only does this make it much easier to get calls booked with prospects, the solutions you build will have much greater value to your clients (meaning you get paid more). A mistake I've seen many AAA operators make (and I blame this on the "Get Rich Quick" YouTubers) is focusing too much on a specific productized service, rather than really understanding the needs of businesses. The former is much done via a SaaS model, but when going the agency route the only thing that makes sense is building custom solutions. This is why I always take a consultant-first approach. You can only build once you understand what they actually need and how certain solutions may impact their operations, workflows, and bottom-line. Basics of How to Get Started Pick a niche. As I mentioned previously, preferably one that you've worked in before. Niches I know of that are actively being bombarded with cold emails include real estate, e-commerce, auto-dealerships, lawyers, and medical offices. There is a reason for this, but I will tell you straight up this business model works well if you target any white-collar service business (internal tools approach) or high volume businesses (customer facing tools approach). Setup your toolbox. If you wanted to start a pressure washing business, you would need a pressure-washer. This is no different. For those without programming knowledge, I've seen two common ways AAA get setup to build- one is having a network of on-call web developers, whether its personal contacts or simply going to Upwork or any talent sourcing agency. The second is having an arsenal of no-code tools. I'll get to this more in a second, but this works beecause at its core, when we are dealing with the practical applications of AI, the code is quite simple, simply put. Start cold sales. Unless you have a network already, this is not a step you can skip. You've already picked a niche, so all you have to do is find the right message. Keep cold emails short, sweet, but enticing- and it will help a lot if you did step 1 correctly and intimately understand who your audience is. I'll be touching base later about how you can leverage AI yourself to help you with outreach and closing. The beauty of gen AI and the AAA model You don't need to be a seasoned web developer to make this business model work. The large majority of solutions that SME clients want is best done using an API for an LLM for the actual AI aspect. The value we create with the solutions we build comes with the conceptual framework and design that not only does what they need it to but integrates smoothly with their existing tech-stack and workflow. The actual implementation is quite straightforward once you understand the high level design and know which tools you are going to use. To give you a sense, even if you plan to build out these apps yourself (say in Python) the large majority of the nitty gritty technical work has already been done for you, especially if you leverage Python libraries and packages that offer high level abstraction for LLM-related functions. For instance, calling GPT can be as little as a single line of code. (And there are no-code tools where these functions are simply an icon on a GUI). Aside from understanding the capabilities and limitations of these tools and frameworks, the only thing that matters is being able to put them in a way that makes sense for what you want to build. Which is why outsourcing and no-code tools both work in our case. Okay... but how TF am I suppposed to actually build out these solutions? Now the fun part. I highly recommend getting familiar with Langchain and LlamaIndex. Both are Python libraires that help a lot with the high-level LLM abstraction I mentioned previously. The two most important aspects include being able to integrate internal data sources/knowledge bases with LLMs, and have LLMs perform autonomous actions. The two most common methods respectively are RAG and output parsing. RAG (retrieval augmented Generation) If you've ever seen a tool that seemingly "trains" GPT on your own data, and wonder how it all works- well I have an answer from you. At a high level, the user query is first being fed to what's called a vector database to run vector search. Vector search basically lets you do semantic search where you are searching data based on meaning. The vector databases then retrieves the most relevant sections of text as it relates to the user query, and this text gets APPENDED to your GPT prompt to provide extra context to the AI. Further, with prompt engineering, you can limit GPT to only generate an answer if it can be found within this extra context, greatly limiting the chance of hallucination (this is where AI makes random shit up). Aside from vector databases, we can also implement RAG with other data sources and retrieval methods, for example SQL databses (via parsing the outputs of LLM's- more on this later). Autonomous Agents via Output Parsing A common need of clients has been having AI actually perform tasks, rather than simply spitting out text. For example, with autonomous agents, we can have an e-commerce chatbot do the work of a basic customer service rep (i.e. look into orders, refunds, shipping). At a high level, what's going on is that the response of the LLM is being used programmtically to determine which API to call. Keeping on with the e-commerce example, if I wanted a chatbot to check shipping status, I could have a LLM response within my app (not shown to the user) with a prompt that outputs a random hash or string, and programmatically I can determine which API call to make based on this hash/string. And using the same fundamental concept as with RAG, I can append the the API response to a final prompt that would spit out the answer for the user. How No Code Tools Can Fit In (With some example solutions you can build) With that being said, you don't necessarily need to do all of the above by coding yourself, with Python libraries or otherwise. However, I will say that having that high level overview will help IMMENSELY when it comes to using no-code tools to do the actual work for you. Regardless, here are a few common solutions you might build for clients as well as some no-code tools you can use to build them out. Ex. Solution 1: AI Chatbots for SMEs (Small and Medium Enterprises) This involves creating chatbots that handle user queries, lead gen, and so forth with AI, and will use the principles of RAG at heart. After getting the required data from your client (i.e. product catalogues, previous support tickets, FAQ, internal documentation), you upload this into your knowledge base and write a prompt that makes sense for your use case. One no-code tool that does this well is MyAskAI. The beauty of it especially for building external chatbots is the ability to quickly ingest entire websites into your knowledge base via a sitemap, and bulk uploading files. Essentially, they've covered the entire grunt work required to do this manually. Finally, you can create a inline or chat widget on your client's website with a few lines of HTML, or altneratively integrate it with a Slack/Teams chatbot (if you are going for an internal Q&A chatbot approach). Other tools you could use include Botpress and Voiceflow, however these are less for RAG and more for building out complete chatbot flows that may or may not incorporate LLMs. Both apps are essentially GUIs that eliminate the pain and tears and trying to implement complex flows manually, and both natively incoporate AI intents and a knowledge base feature. Ex. Solution 2: Internal Apps Similar to the first example, except we go beyond making just chatbots but tools such as report generation and really any sort of internal tool or automations that may incorporate LLM's. For instance, you can have a tool that automatically generates replies to inbound emails based on your client's knowledge base. Or an automation that does the same thing but for replies to Instagram comments. Another example could be a tool that generates a description and screeenshot based on a URL (useful for directory sites, made one for my own :P). Getting into more advanced implementations of LLMs, we can have tools that can generate entire drafts of reports (think 80+ pages), based not only on data from a knowledge base but also the writing style, format, and author voice of previous reports. One good tool to create content generation panels for your clients would be MindStudio. You can train LLM's via prompt engineering in a structured way with your own data to essentially fine tune them for whatever text you need it to generate. Furthermore, it has a GUI where you can dictate the entire AI flow. You can also upload data sources via multiple formats, including PDF, CSV, and Docx. For automations that require interactions between multiple apps, I recommend the OG zapier/make.com if you want a no-code solution. For instance, for the automatic email reply generator, I can have a trigger such that when an email is received, a custom AI reply is generated by MyAskAI, and finally a draft is created in my email client. Or, for an automation where I can create a social media posts on multiple platforms based on a RSS feed (news feed), I can implement this directly in Zapier with their native GPT action (see screenshot) As for more complex LLM flows that may require multiple layers of LLMs, data sources, and APIs working together to generate a single response i.e. a long form 100 page report, I would recommend tools such as Stack AI or Flowise (open-source alternative) to build these solutions out. Essentially, you get most of the functions and features of Python packages such as Langchain and LlamaIndex in a GUI. See screenshot for an example of a flow How the hell are you supposed to find clients? With all that being said, none of this matters if you can't find anyone to sell to. You will have to do cold sales, one way or the other, especially if you are brand new to the game. And what better way to sell your AI services than with AI itself? If we want to integrate AI into the cold outreach process, first we must identify what it's good at doing, and that's obviously writing a bunch of text, in a short amount of time. Similar to the solutions that an AAA can build for its clients, we can take advantage of the same principles in our own sales processes. How to do outreach Once you've identified your niche and their pain points/opportunities for automation, you want to craft a compelling message in which you can send via cold email and cold calls to get prospects booked on demos/consultations. I won't get into too much detail in terms of exactly how to write emails or calling scripts, as there are millions of resources to help with this, but I will tell you a few key points you want to keep in mind when doing outreach for your AAA. First, you want to keep in mind that many businesses are still hesitant about AI and may not understand what it really is or how it can benefit their operations. However, we can take advantage of how mass media has been reporting on AI this past year- at the very least people are AWARE that sooner or later they may have to implement AI into their businesses to stay competitive. We want to frame our message in a way that introduces generative AI as a technology that can have a direct, tangible, and positive impact on their business. Although it may be hard to quantify, I like to include estimates of man-hours saved or costs saved at least in my final proposals to prospects. Times are TOUGH right now, and money is expensive, so you need to have a compelling reason for businesses to get on board. Once you've gotten your messaging down, you will want to create a list of prospects to contact. Tools you can use to find prospects include Apollo.io, reply.io, zoominfo (expensive af), and Linkedin Sales Navigator. What specific job titles, etc. to target will depend on your niche but for smaller companies this will tend to be the owner. For white collar niches, i.e. law, the professional that will be directly benefiting from the tool (i.e. partners) may be better to contact. And for larger organizations you may want to target business improvement and digital transformation leads/directors- these are the people directly in charge of projects like what you may be proposing. Okay- so you have your message, and your list, and now all it comes down to is getting the good word out. I won't be going into the details of how to send these out, a quick Google search will give you hundreds of resources for cold outreach methods. However, personalization is key and beyond simple dynamic variables you want to make sure you can either personalize your email campaigns directly with AI (SmartWriter.ai is an example of a tool that can do this), or at the very least have the ability to import email messages programmatically. Alternatively, ask ChatGPT to make you a Python Script that can take in a list of emails, scrape info based on their linkedin URL or website, and all pass this onto a GPT prompt that specifies your messaging to generate an email. From there, send away. How tf do I close? Once you've got some prospects booked in on your meetings, you will need to close deals with them to turn them into clients. Call #1: Consultation Tying back to when I mentioned you want to take a consultant-first appraoch, you will want to listen closely to their goals and needs and understand their pain points. This would be the first call, and typically I would provide a high level overview of different solutions we could build to tacke these. It really helps to have a presentation available, so you can graphically demonstrate key points and key technologies. I like to use Plus AI for this, it's basically a Google Slides add-on that can generate slide decks for you. I copy and paste my default company messaging, add some key points for the presentation, and it comes out with pretty decent slides. Call #2: Demo The second call would involve a demo of one of these solutions, and typically I'll quickly prototype it with boilerplate code I already have, otherwise I'll cook something up in a no-code tool. If you have a niche where one type of solution is commonly demanded, it helps to have a general demo set up to be able to handle a larger volume of calls, so you aren't burning yourself out. I'll also elaborate on how the final product would look like in comparison to the demo. Call #3 and Beyond: Once the initial consultation and demo is complete, you will want to alleviate any remaining concerns from your prospects and work with them to reach a final work proposal. It's crucial you lay out exactly what you will be building (in writing) and ensure the prospect understands this. Furthermore, be clear and transparent with timelines and communication methods for the project. In terms of pricing, you want to take this from a value-based approach. The same solution may be worth a lot more to client A than client B. Furthermore, you can create "add-ons" such as monthly maintenance/upgrade packages, training sessions for employeees, and so forth, separate from the initial setup fee you would charge. How you can incorporate AI into marketing your businesses Beyond cold sales, I highly recommend creating a funnel to capture warm leads. For instance, I do this currently with my AI tools directory, which links directly to my AI agency and has consistent branding throughout. Warm leads are much more likely to close (and honestly, much nicer to deal with). However, even without an AI-related website, at the very least you will want to create a presence on social media and the web in general. As with any agency, you will want basic a professional presence. A professional virtual address helps, in addition to a Google Business Profile (GBP) and TrustPilot. a GBP (especially for local SEO) and Trustpilot page also helps improve the looks of your search results immensely. For GBP, I recommend using ProfilePro, which is a chrome extension you can use to automate SEO work for your GBP. Aside from SEO optimzied business descriptions based on your business, it can handle Q/A answers, responses, updates, and service descriptions based on local keywords. Privacy and Legal Concerns of the AAA Model Aside from typical concerns for agencies relating to service contracts, there are a few issues (especially when using no-code tools) that will need to be addressed to run a successful AAA. Most of these surround privacy concerns when working with proprietary data. In your terms with your client, you will want to clearly define hosting providers and any third party tools you will be using to build their solution, and a DPA with these third parties listed as subprocessors if necessary. In addition, you will want to implement best practices like redacting private information from data being used for building solutions. In terms of addressing concerns directly from clients, it helps if you host your solutions on their own servers (not possible with AI tools), and address the fact only ChatGPT queries in the web app, not OpenAI API calls, will be used to train OpenAI's models (as reported by mainstream media). The key here is to be open and transparent with your clients about ALL the tools you are using, where there data will be going, and make sure to get this all in writing. have fun, and keep an open mind Before I finish this post, I just want to reiterate the fact that this is NOT an easy way to make money. Running an AI agency will require hours and hours of dedication and work, and constantly rearranging your schedule to meet prospect and client needs. However, if you are looking for a new business to run, and have a knack for understanding business operations and are genuinely interested in the pracitcal applications of generative AI, then I say go for it. The time is ticking before AAA becomes the new dropshipping or SMMA, and I've a firm believer that those who set foot first and establish themselves in this field will come out top. And remember, while 100 thousand people may read this post, only 2 may actually take initiative and start.

As a soloproneur, here is how I'm scaling with AI and GPT-based tools
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Scout_OfficialThis week

As a soloproneur, here is how I'm scaling with AI and GPT-based tools

Being a solopreneur has its fair share of challenges. Currently I've got businesses in ecommerce, agency work, and affiliate marketing, and one undeniable truth remains: to truly scale by yourself, you need more than just sheer will. That's where I feel technology, especially AI, steps in. As such, I wanted some AI tools that have genuinely made a difference in my own work as a solo business operator. No fluff, just tried-and-true tools and platforms that have worked for me. The ability for me to scale alone with AI tools that take advantage of GPT in one way, or another has been significant and really changed my game over the past year. They bring in an element of adaptability and intelligence and work right alongside “traditional automation”. Whether you're new to this or looking to optimize your current setup, I hope this post helps. FYI I used multiple prompts with GPT-4 to draft this using my personal notes. Plus AI (add-on for google slides/docs) I handle a lot of sales calls and demos for my AI automation agency. As I’m providing a custom service rather than a product, every client has different pain points and as such I need to make a new slide deck each time. And making slides used to be a huge PITA and pretty much the bane of my existence until slide deck generators using GPT came out. My favorite so far has been PlusAI, which works as a plugin for Google Slides. You pretty much give it a rough idea, or some key points and it creates some slides right within Google Slides. For me, I’ve been pasting the website copy or any information on my client, then telling PlusAI the service I want to propose. After the slides are made, you have a lot of leeway to edit the slides again with AI, compared to other slide generators out there. With 'Remix', I can switch up layouts if something feels off, and 'Rewrite' is there to gently nudge the AI in a different direction if I ever need it to. It's definitely given me a bit of breathing space in a schedule that often feels suffocating. echo.win (web-based app) As a solopreneur, I'm constantly juggling roles. Managing incoming calls can be particularly challenging. Echo.win, a modern call management platform, has become a game-changer for my business. It's like having a 24/7 personal assistant. Its advanced AI understands and responds to queries in a remarkably human way, freeing up my time. A standout feature is the Scenario Builder, allowing me to create personalized conversation flows. Live transcripts and in-depth analytics help me make data-driven decisions. The platform is scalable, handling multiple simultaneous calls and improving customer satisfaction. Automatic contact updates ensure I never miss an important call. Echo.win's pricing is reasonable, offering a personalized business number, AI agents, unlimited scenarios, live transcripts, and 100 answered call minutes per month. Extra minutes are available at a nominal cost. Echo.win has revolutionized my call management. It's a comprehensive, no-code platform that ensures my customers are always heard and never missed MindStudio by YouAi (web app/GUI) I work with numerous clients in my AI agency, and a recurring task is creating chatbots and demo apps tailored to their specific needs and connected to their knowledge base/data sources. Typically, I would make production builds from scratch with libraries such as LangChain/LlamaIndex, however it’s quite cumbersome to do this for free demos. As each client has unique requirements, it means I'm often creating something from scratch. For this, I’ve been using MindStudio (by YouAi) to quickly come up with the first iteration of my app. It supports multiple AI models (GPT, Claude, Llama), let’s you upload custom data sources via multiple formats (PDF, CSV, Excel, TXT, Docx, and HTML), allows for custom flows and rules, and lets you to quickly publish your apps. If you are in their developer program, YouAi has built-in payment infrastructure to charge your users for using your app. Unlike many of the other AI builders I’ve tried, MindStudio basically lets me dictate every step of the AI interaction at a high level, while at the same time simplifying the behind-the-scenes work. Just like how you'd sketch an outline or jot down main points, you start with a scaffold or decide to "remix" an existing AI, and it will open up the IDE. I often find myself importing client data or specific project details, and then laying out the kind of app or chatbot I'm looking to prototype. And once you've got your prototype you can customize the app as much as you want. LLamaIndex (Python framework) As mentioned before, in my AI agency, I frequently create chatbots and apps for clients, tailored to their specific needs and connected to their data sources. LlamaIndex, a data framework for LLM applications, has been a game-changer in this process. It allows me to ingest, structure, and access private or domain-specific data. The major difference over LangChain is I feel like LlamaIndex does high level abstraction much better.. Where LangChain unnecessarily abstracts the simplest logic, LlamaIndex actually has clear benefits when it comes to integrating your data with LLMs- it comes with data connectors that ingest data from various sources and formats, data indexes that structure data for easy consumption by LLMs, and engines that provide natural language access to data. It also includes data agents, LLM-powered knowledge workers augmented by tools, and application integrations that tie LlamaIndex back into the rest of the ecosystem. LlamaIndex is user-friendly, allowing beginners to use it with just five lines of code, while advanced users can customize and extend any module to fit their needs. To be completely honest, to me it’s more than a tool- at its heart it’s a framework that ensures seamless integration of LLMs with data sources while allowing for complete flexibility compared to no-code tools. GoCharlie (web app) GoCharlie, the first AI Agent product for content creation, has been a game-changer for my business. Powered by a proprietary LLM called Charlie, it's capable of handling multi-input/multi-output tasks. GoCharlie's capabilities are vast, including content repurposing, image generation in 4K and 8K for various aspect ratios, SEO-optimized blog creation, fact-checking, web research, and stock photo and GIF pull-ins. It also offers audio transcriptions for uploaded audio/video files and YouTube URLs, web scraping capabilities, and translation. One standout feature is its multiple input capability, where I can attach a file (like a brand brief from a client) and instruct it to create a social media campaign using brand guidelines. It considers the file, prompt, and website, and produces multiple outputs for each channel, each of which can be edited separately. Its multi-output feature allows me to write a prompt and receive a response, which can then be edited further using AI. Overall, very satisfied with GoCharlie and in my opinion it really presents itself as an effective alternative to GPT based tools. ProfilePro (chrome extension) As someone overseeing multiple Google Business Profiles (GBPs) for my various businesses, I’ve been using ProfilePro by Merchynt. This tool stood out with its ability to auto-generate SEO-optimized content like review responses and business updates based on minimal business input. It works as a Chrome extension, and offers suggestions for responses automatically on your GBP, with multiple options for the tone it will write in. As a plus, it can generate AI images for Google posts, and offer suggestions for services and service/product descriptions. While it streamlines many GBP tasks, it still allows room for personal adjustments and refinements, offering a balance between automation and individual touch. And if you are like me and don't have dedicated SEO experience, it can handle ongoing optimization tasks to help boost visibility and drive more customers to profiles through Google Maps and Search

Founder Pitch: AI Agent for Simplifying Public Cloud Management
reddit
LLM Vibe Score0
Human Vibe Score1
rasvi786This week

Founder Pitch: AI Agent for Simplifying Public Cloud Management

Video to understand : https://youtu.be/9ocUjlUrU\w?si=S0ETDbKSdJqlVDyg Are You Ready to Redefine Cloud Management with AI? Imagine an intelligent AI agent that transforms the complexity of managing public cloud infrastructure into simple, natural language commands. No more navigating through endless configurations or deciphering technical documentation—our AI agent is here to revolutionize the way organizations interact with cloud platforms. About the Project We’re building an AI-powered agent designed to handle public cloud management tasks seamlessly. Whether you’re setting up your organization’s cloud foundation or deploying complex workloads, this AI agent makes it as easy as having a conversation. What Can the AI Agent Do? Cloud Foundation Setup: Example: “Please set up a cloud foundation blueprint for my organization on Google Cloud.”* The AI agent will ask key questions (e.g., organization ID) and guide you through authentication. Once authorized, it sets up the foundation using GCP APIs. Workload Deployment: Example: “Spin up a GKE cluster for me.”* The agent will ask for necessary details (e.g., number of nodes, VPC info), authenticate, and deploy the cluster in minutes. Security and Compliance Validation: Example: “Validate my organization’s cloud setup and check for security vulnerabilities.”* The agent audits your setup, identifies potential risks, and provides actionable insights. Current Progress We’ve developed a working prototype that integrates with major cloud providers like Google Cloud. The AI agent can already: Authenticate with cloud APIs Execute foundational tasks such as setting up organizations and spinning up clusters Perform initial security validations Who I’m Looking For I’m searching for a co-founder with enterprise sales experience and a strategic vision to grow our user base. You will be instrumental in helping us: Build relationships with companies willing to pilot our product Develop go-to-market strategies for enterprise adoption Identify opportunities for partnerships with cloud service providers Your Role As a co-founder, you’ll lead efforts to: Secure Pilot Programs: Identify and onboard enterprises for product trials to gather feedback and refine the solution. Drive Growth: Develop scalable strategies to grow our user base across industries. Market Positioning: Work with me to define our unique value proposition and establish thought leadership in the cloud management space. My Background I bring over a decade of experience in tech, with a strong focus on software engineering and infrastructure. My contributions so far include: Developing the core AI engine and cloud integrations Designing workflows that simplify complex cloud tasks Why Join This Project? Revolutionize Cloud Management: Be part of a project that will redefine how organizations interact with public clouds. Tackle Challenging Problems: Work at the cutting edge of AI and cloud computing. High Growth Potential: Join an industry projected to grow exponentially as enterprises embrace AI-driven automation. Build a Company from Scratch: Shape the product, team, and culture as we grow together. What’s Next? Our immediate priorities include: Expanding the AI agent’s capabilities to support multi-cloud setups. Conducting pilot programs with enterprise clients. Iterating on the product based on real-world feedback. What We Need to Succeed Expertise in enterprise sales and partnerships A deep understanding of enterprise challenges and cloud adoption trends A shared passion for leveraging AI to solve complex problems Let’s work together to build the future of cloud management. If you’re excited about this vision and bring the expertise we need, I’d love to connect and discuss how we can take this project to the next level.

Founder Pitch: AI Agent for Simplifying Public Cloud Management
reddit
LLM Vibe Score0
Human Vibe Score1
rasvi786This week

Founder Pitch: AI Agent for Simplifying Public Cloud Management

Video to understand : https://youtu.be/9ocUjlUrU\w?si=S0ETDbKSdJqlVDyg Are You Ready to Redefine Cloud Management with AI? Imagine an intelligent AI agent that transforms the complexity of managing public cloud infrastructure into simple, natural language commands. No more navigating through endless configurations or deciphering technical documentation—our AI agent is here to revolutionize the way organizations interact with cloud platforms. About the Project We’re building an AI-powered agent designed to handle public cloud management tasks seamlessly. Whether you’re setting up your organization’s cloud foundation or deploying complex workloads, this AI agent makes it as easy as having a conversation. What Can the AI Agent Do? Cloud Foundation Setup: Example: “Please set up a cloud foundation blueprint for my organization on Google Cloud.”* The AI agent will ask key questions (e.g., organization ID) and guide you through authentication. Once authorized, it sets up the foundation using GCP APIs. Workload Deployment: Example: “Spin up a GKE cluster for me.”* The agent will ask for necessary details (e.g., number of nodes, VPC info), authenticate, and deploy the cluster in minutes. Security and Compliance Validation: Example: “Validate my organization’s cloud setup and check for security vulnerabilities.”* The agent audits your setup, identifies potential risks, and provides actionable insights. Current Progress We’ve developed a working prototype that integrates with major cloud providers like Google Cloud. The AI agent can already: Authenticate with cloud APIs Execute foundational tasks such as setting up organizations and spinning up clusters Perform initial security validations Who I’m Looking For I’m searching for a co-founder with enterprise sales experience and a strategic vision to grow our user base. You will be instrumental in helping us: Build relationships with companies willing to pilot our product Develop go-to-market strategies for enterprise adoption Identify opportunities for partnerships with cloud service providers Your Role As a co-founder, you’ll lead efforts to: Secure Pilot Programs: Identify and onboard enterprises for product trials to gather feedback and refine the solution. Drive Growth: Develop scalable strategies to grow our user base across industries. Market Positioning: Work with me to define our unique value proposition and establish thought leadership in the cloud management space. My Background I bring over a decade of experience in tech, with a strong focus on software engineering and infrastructure. My contributions so far include: Developing the core AI engine and cloud integrations Designing workflows that simplify complex cloud tasks Why Join This Project? Revolutionize Cloud Management: Be part of a project that will redefine how organizations interact with public clouds. Tackle Challenging Problems: Work at the cutting edge of AI and cloud computing. High Growth Potential: Join an industry projected to grow exponentially as enterprises embrace AI-driven automation. Build a Company from Scratch: Shape the product, team, and culture as we grow together. What’s Next? Our immediate priorities include: Expanding the AI agent’s capabilities to support multi-cloud setups. Conducting pilot programs with enterprise clients. Iterating on the product based on real-world feedback. What We Need to Succeed Expertise in enterprise sales and partnerships A deep understanding of enterprise challenges and cloud adoption trends A shared passion for leveraging AI to solve complex problems Let’s work together to build the future of cloud management. If you’re excited about this vision and bring the expertise we need, I’d love to connect and discuss how we can take this project to the next level.

nine
github
LLM Vibe Score0.406
Human Vibe Score0.000678327714013925
NethermindEthMar 28, 2025

nine

NINE - Neural Interconnected Nodes Engine A flexible framework for building a distributed network of AI agents that work everywhere (STD, WASM, TEE) with a dynamic interface and hot-swappable components. One of the key concepts of the framework is a meta-layer that enables building software systems in a No-code style, where the entire integration is handled by the LLM. Documentation | Telegram | X | Discord Overview Project Structure The project is built using Rust (full-stack) and organized as a workspace consisting of two major groups: substance/ - The core components of the system, responsible for interaction. particles/ - Plugins for the system that enable additional functionalities. examples/ - Usage examples of the framework. Use cases The following cases will have a minimal implementation, and they will be used to track the progress of the framework and its flexibility in building such systems. ☑️ Chatbots - AI-driven natural language chatbots for customer support, virtual assistants, and automation. ☑️ AI-governed blockchains (ChaosChain) - Self-regulating and intelligent blockchain ecosystems with automated decision-making. ⬜ Personal AI Assistant with dynamic UI - AI that generates adaptive and context-aware user interfaces on demand. ☑️ AI-powered trading bots - Autonomous financial agents for high-frequency trading and portfolio management. ⬜ Intelligent email assistant - AI for reading, summarizing, filtering, and responding to emails autonomously. ⬜ Interactivity in home appliances - AI-powered automation for home appliances, making them responsive and adaptive. ⬜ On-demand observability and awareness in DevOps - AI-driven insights, predictive monitoring, and automated issue detection in IT systems. ⬜ AI-powered developer tools - AI agents assisting with code generation, debugging, and software optimization. ⬜ Autonomous research agent - Self-learning AI for data analysis, knowledge discovery, and hypothesis testing. Status: ⬜ Not started | ☑️ In Progress | ✅ Completed Interfaces The platform provides No-code interfaces that automatically adapt to your needs and use LLM for system management. ☑️ Stdio - A console interface that also allows interaction with models through the terminal or via scripts. ☑️ TUI - An advanced console interface with an informative dashboard and the ability to interact more comprehensively with the system. ☑️ GUI - A graphical immediate-state interface suitable for embedded systems with real-time information rendering. ⬜ WEB - The ability to interact with the system through a web browser, such as from a mobile phone. ⬜ Voice - An interface for people with disabilities or those who prefer interaction without a graphical representation (e.g., voice control). ⬜ API - On-the-fly API creation for your system, providing a formal interaction method. This includes encapsulating an entire mesh system into a simple tool for LLM. Features (goals) Built on Rust and implemented as hybrid actor-state machines. Supports various LLMs, tools, and extensibility. Hot model swapping without restarting. Real-time configuration adjustment. Distributed agents, the ability to run components on different machines. Provides a dynamic user interface (UI9) that is automatically generated for interacting with a network of agents. Usage An agent is a substance that assembles from components (particles). Connections automatically form between them, bringing the agent to life: License This project is licensed under the [MIT license]. [MIT license]: https://github.com/NethermindEth/nine/blob/trunk/LICENSE Contribution Unless you explicitly state otherwise, any contribution intentionally submitted for inclusion in this project by you, shall be licensed as MIT, without any additional terms or conditions.

xpert
github
LLM Vibe Score0.457
Human Vibe Score0.0831216059433162
xpert-aiMar 28, 2025

xpert

English | 中文 [uri_license]: https://www.gnu.org/licenses/agpl-3.0.html [urilicenseimage]: https://img.shields.io/badge/License-AGPL%20v3-blue.svg Xpert Cloud · Self-hosting · Documentation · Enterprise inquiry Open-Source AI Platform for Enterprise Data Analysis, Indicator Management and Agents Orchestration Xpert AI is an open-source enterprise-level AI system that perfectly integrates two major platforms: agent orchestration and data analysis. 💡 What's New Agent and Workflow Hybrid Architecture In today's rapidly evolving AI landscape, enterprises face a critical dilemma: how to balance the creativity of LLMs with the stability of processes? While purely agent-based architectures offer flexibility, they are difficult to control; traditional workflows, though reliable, lack adaptability. The Agent and Workflow Hybrid Architecture of the Xpert AI platform is designed to resolve this conflict — it allows AI to possess "free will" while adhering to "rules and order." !agent-workflow-hybrid-architecture Blog - Agent and Workflow Hybrid Architecture Agent Orchestration Platform By coordinating the collaboration of multiple agents, Xpert completes complex tasks. Xpert integrates different types of AI agents through an efficient management mechanism, utilizing their capabilities to solve multidimensional problems. Xpert Agents Data Analysis Platform An agile data analysis platform based on cloud computing for multidimensional modeling, indicator management, and BI display. It supports connecting to various data sources, achieving efficient and flexible data analysis and visualization, and provides multiple intelligent analysis functions and tools to help enterprises quickly and accurately discover business value and make operational decisions. ChatBI ChatBI is an innovative feature we are introducing, combining chat functionality with business intelligence (BI) analysis capabilities. It offers users a more intuitive and convenient data analysis experience through natural language interaction. ChatBI_Demo.mp4 🚀 Quick Start Before installing Xpert, make sure your machine meets the following minimum system requirements: CPU >= 2 Core RAM >= 4 GiB Node.js (ESM and CommonJS) - 18.x, 19.x, 20.x, 22.x The easiest way to start the Xpert server is through docker compose. Before running Xpert with the following commands, make sure that Docker and Docker Compose are installed on your machine: After running, you can access the Xpert dashboard in your browser at http://localhost/onboarding and start the initialization process. Please check our Wiki - Development to get started quickly. 🎯 Mission Empowering enterprises with intelligent collaboration and data-driven insights through innovative AI orchestration and agile analytics. 🌼 Screenshots Show / Hide Screenshots Pareto analysis open in new tab !Pareto analysis Screenshot Product profit analysis open in new tab !Product profit analysis Screenshot Reseller analysis open in new tab !Reseller analysis Screenshot Bigview dashboard open in new tab !Bigview dashboard Screenshot Indicator application open in new tab !Indicator application Screenshot Indicator mobile app open in new tab !Indicator mobile app Screenshot 💻 Demo, Downloads, Testing and Production Demo Xpert AI Platform Demo at . Notes: You can generate samples data in the home dashbaord page. Production (SaaS) Xpert AI Platform SaaS is available at . Note: it's currently in Alpha version / in testing mode, please use it with caution! 🧱 Technology Stack and Requirements TypeScript language NodeJs / NestJs Nx Angular RxJS TypeORM Langchain ECharts Java Mondrian For Production, we recommend: PostgreSQL PM2 See also README.md and CREDITS.md files in relevant folders for lists of libraries and software included in the Platform, information about licenses, and other details 📄 Documentation Please refer to our official Platform Documentation and to our Wiki (WIP). 💌 Contact Us For business inquiries: Xpert AI Platform @ Twitter 🛡️ License We support the open-source community. This software is available under the following licenses: Xpert AI Platform Community Edition Xpert AI Platform Small Business Xpert AI Platform Enterprise Please see LICENSE for more information on licenses. 💪 Thanks to our Contributors Contributors Please give us :star: on Github, it helps! You are more than welcome to submit feature requests in the Xpert AI repo Pull requests are always welcome! Please base pull requests against the develop branch and follow the contributing guide.

AI-Scalpel-Trading-Bot
github
LLM Vibe Score0.491
Human Vibe Score0.09890315835809398
hackobiMar 28, 2025

AI-Scalpel-Trading-Bot

AI-Scalpel-Trading-Bot Disclaimer This software is for educational purposes only. Do not risk money which you are afraid to lose. USE THE SOFTWARE AT YOUR OWN RISK. THE AUTHORS AND ALL AFFILIATES ASSUME NO RESPONSIBILITY FOR YOUR TRADING RESULTS. Always start by running a trading bot in Dry-run and do not engage money before you understand how it works and what profit/loss you should expect. This is an implementation of freqtrade where different machine learning implementations will be tested. Freqtrade is a free and open source crypto trading bot written in Python. It is designed to support all major exchanges and be controlled via Telegram. It contains backtesting, plotting and money management tools as well as strategy optimization by machine learning. !freqtrade Exchange marketplaces supported [X] Bittrex [X] Binance (*Note for binance users) [ ] 113 others to tests. (Some of them might not work) Documentation Documentation. Features [x] Based on Python 3.6+: For botting on any operating system - Windows, macOS and Linux. [x] Persistence: Persistence is achieved through sqlite. [x] Dry-run: Run the bot without playing money. [x] Backtesting: Run a simulation of your buy/sell strategy. [x] Strategy Optimization by machine learning: Use machine learning to optimize your buy/sell strategy parameters with real exchange data. [x] Edge position sizing Calculate your win rate, risk reward ratio, the best stoploss and adjust your position size before taking a position for each specific market. Learn more. [x] Whitelist crypto-currencies: Select which crypto-currency you want to trade or use dynamic whitelists. [x] Blacklist crypto-currencies: Select which crypto-currency you want to avoid. [x] Manageable via Telegram: Manage the bot with Telegram. [x] Display profit/loss in fiat: Display your profit/loss in 33 fiat. [x] Daily summary of profit/loss: Provide a daily summary of your profit/loss. [x] Performance status report: Provide a performance status of your current trades. Quick start Freqtrade provides a Linux/macOS script to install all dependencies and help you to configure the bot. Other installations. Basic Usage Bot commands Telegram RPC commands Telegram is not mandatory. However, this is a great way to control your bot. More details on our documentation /start: Starts the trader /stop: Stops the trader /status [table]: Lists all open trades /count: Displays number of open trades /profit: Lists cumulative profit from all finished trades /forcesell |all: Instantly sells the given trade (Ignoring minimum_roi). /performance: Show performance of each finished trade grouped by pair /balance: Show account balance per currency /daily : Shows profit or loss per day, over the last n days /help: Show help message /version: Show version Development branches The project is currently setup in two main branches: develop - This branch has often new features, but might also cause breaking changes. master - This branch contains the latest stable release. The bot 'should' be stable on this branch, and is generally well tested. feat/* - These are feature branches, which are being worked on heavily. Please don't use these unless you want to test a specific feature. A note on Binance For Binance, please add "BNB/" to your blacklist to avoid issues. Accounts having BNB accounts use this to pay for fees - if your first trade happens to be on BNB, further trades will consume this position and make the initial BNB order unsellable as the expected amount is not there anymore. Support Help / Slack For any questions not covered by the documentation or for further information about the bot, I encourage you to join freqtrade's slack channel. Click here to join Slack channel. Bugs / Issues If you discover a bug in the bot, please search their issue tracker first. If it hasn't been reported, please create a new issue and ensure you follow the template guide so that our team can assist you as quickly as possible. Feature Requests Have you a great idea to improve the bot you want to share? Please, first search if this feature was not already discussed. If it hasn't been requested, please create a new request and ensure you follow the template guide so that it does not get lost in the bug reports. Pull Requests Feel like the bot is missing a feature? Keep em pull requests coming! Please read the Contributing document to understand the requirements before sending pull-requests. Coding is not a neccessity to contribute - maybe start with improving our documentation? Issues labeled good first issue can be good first contributions, and will help get you familiar with the codebase. Note before starting any major new feature work, please open an issue describing what you are planning to do or talk to the team on Slack. This will ensure that interested parties can give valuable feedback on the feature, and let others know that you are working on it. Important: Always create your PR against the develop branch, not master. Requirements Uptodate clock The clock must be accurate, syncronized to a NTP server very frequently to avoid problems with communication to the exchanges. Min hardware required To run this bot we recommend you a cloud instance with a minimum of: Minimal (advised) system requirements: 2GB RAM, 1GB disk space, 2vCPU Software requirements Python 3.6.x pip git TA-Lib virtualenv (Recommended) Docker (Recommended)

aima-python
github
LLM Vibe Score0.575
Human Vibe Score0.33114909407186394
aimacodeMar 28, 2025

aima-python

aima-python Python code for the book Artificial Intelligence: A Modern Approach. You can use this in conjunction with a course on AI, or for study on your own. We're looking for solid contributors to help. Updates for 4th Edition The 4th edition of the book as out now in 2020, and thus we are updating the code. All code here will reflect the 4th edition. Changes include: Move from Python 3.5 to 3.7. More emphasis on Jupyter (Ipython) notebooks. More projects using external packages (tensorflow, etc.). Structure of the Project When complete, this project will have Python implementations for all the pseudocode algorithms in the book, as well as tests and examples of use. For each major topic, such as search, we provide the following files: search.ipynb and search.py: Implementations of all the pseudocode algorithms, and necessary support functions/classes/data. The .py file is generated automatically from the .ipynb file; the idea is that it is easier to read the documentation in the .ipynb file. search_XX.ipynb: Notebooks that show how to use the code, broken out into various topics (the XX). tests/test_search.py: A lightweight test suite, using assert statements, designed for use with py.test, but also usable on their own. Python 3.7 and up The code for the 3rd edition was in Python 3.5; the current 4th edition code is in Python 3.7. It should also run in later versions, but does not run in Python 2. You can install Python or use a browser-based Python interpreter such as repl.it. You can run the code in an IDE, or from the command line with python -i filename.py where the -i option puts you in an interactive loop where you can run Python functions. All notebooks are available in a binder environment. Alternatively, visit jupyter.org for instructions on setting up your own Jupyter notebook environment. Features from Python 3.6 and 3.7 that we will be using for this version of the code: f-strings: all string formatting should be done with f'var = {var}', not with 'var = {}'.format(var) nor 'var = %s' % var. typing module: declare functions with type hints: def successors(state) -> List[State]:; that is, give type declarations, but omit them when it is obvious. I don't need to say state: State, but in another context it would make sense to say s: State. Underscores in numerics: write a million as 1000000 not as 1000000. dataclasses module: replace namedtuple with dataclass. [//]: (There is a sibling [aima-docker]https://github.com/rajatjain1997/aima-docker project that shows you how to use docker containers to run more complex problems in more complex software environments.) Installation Guide To download the repository: git clone https://github.com/aimacode/aima-python.git Then you need to install the basic dependencies to run the project on your system: You also need to fetch the datasets from the aima-data repository: Wait for the datasets to download, it may take a while. Once they are downloaded, you need to install pytest, so that you can run the test suite: pip install pytest Then to run the tests: py.test And you are good to go! Index of Algorithms Here is a table of algorithms, the figure, name of the algorithm in the book and in the repository, and the file where they are implemented in the repository. This chart was made for the third edition of the book and is being updated for the upcoming fourth edition. Empty implementations are a good place for contributors to look for an issue. The aima-pseudocode project describes all the algorithms from the book. An asterisk next to the file name denotes the algorithm is not fully implemented. Another great place for contributors to start is by adding tests and writing on the notebooks. You can see which algorithms have tests and notebook sections below. If the algorithm you want to work on is covered, don't worry! You can still add more tests and provide some examples of use in the notebook! | Figure | Name (in 3rd edition) | Name (in repository) | File | Tests | Notebook |:-------|:----------------------------------|:------------------------------|:--------------------------------|:-----|:---------| | 2 | Random-Vacuum-Agent | RandomVacuumAgent | [agents.py][agents] | Done | Included | | 2 | Model-Based-Vacuum-Agent | ModelBasedVacuumAgent | [agents.py][agents] | Done | Included | | 2.1 | Environment | Environment | [agents.py][agents] | Done | Included | | 2.1 | Agent | Agent | [agents.py][agents] | Done | Included | | 2.3 | Table-Driven-Vacuum-Agent | TableDrivenVacuumAgent | [agents.py][agents] | Done | Included | | 2.7 | Table-Driven-Agent | TableDrivenAgent | [agents.py][agents] | Done | Included | | 2.8 | Reflex-Vacuum-Agent | ReflexVacuumAgent | [agents.py][agents] | Done | Included | | 2.10 | Simple-Reflex-Agent | SimpleReflexAgent | [agents.py][agents] | Done | Included | | 2.12 | Model-Based-Reflex-Agent | ReflexAgentWithState | [agents.py][agents] | Done | Included | | 3 | Problem | Problem | [search.py][search] | Done | Included | | 3 | Node | Node | [search.py][search] | Done | Included | | 3 | Queue | Queue | [utils.py][utils] | Done | No Need | | 3.1 | Simple-Problem-Solving-Agent | SimpleProblemSolvingAgent | [search.py][search] | Done | Included | | 3.2 | Romania | romania | [search.py][search] | Done | Included | | 3.7 | Tree-Search | depth/breadthfirsttree_search | [search.py][search] | Done | Included | | 3.7 | Graph-Search | depth/breadthfirstgraph_search | [search.py][search] | Done | Included | | 3.11 | Breadth-First-Search | breadthfirstgraph_search | [search.py][search] | Done | Included | | 3.14 | Uniform-Cost-Search | uniformcostsearch | [search.py][search] | Done | Included | | 3.17 | Depth-Limited-Search | depthlimitedsearch | [search.py][search] | Done | Included | | 3.18 | Iterative-Deepening-Search | iterativedeepeningsearch | [search.py][search] | Done | Included | | 3.22 | Best-First-Search | bestfirstgraph_search | [search.py][search] | Done | Included | | 3.24 | A\*-Search | astar_search | [search.py][search] | Done | Included | | 3.26 | Recursive-Best-First-Search | recursivebestfirst_search | [search.py][search] | Done | Included | | 4.2 | Hill-Climbing | hill_climbing | [search.py][search] | Done | Included | | 4.5 | Simulated-Annealing | simulated_annealing | [search.py][search] | Done | Included | | 4.8 | Genetic-Algorithm | genetic_algorithm | [search.py][search] | Done | Included | | 4.11 | And-Or-Graph-Search | andorgraph_search | [search.py][search] | Done | Included | | 4.21 | Online-DFS-Agent | onlinedfsagent | [search.py][search] | Done | Included | | 4.24 | LRTA\*-Agent | LRTAStarAgent | [search.py][search] | Done | Included | | 5.3 | Minimax-Decision | minimax_decision | [games.py][games] | Done | Included | | 5.7 | Alpha-Beta-Search | alphabeta_search | [games.py][games] | Done | Included | | 6 | CSP | CSP | [csp.py][csp] | Done | Included | | 6.3 | AC-3 | AC3 | [csp.py][csp] | Done | Included | | 6.5 | Backtracking-Search | backtracking_search | [csp.py][csp] | Done | Included | | 6.8 | Min-Conflicts | min_conflicts | [csp.py][csp] | Done | Included | | 6.11 | Tree-CSP-Solver | treecspsolver | [csp.py][csp] | Done | Included | | 7 | KB | KB | [logic.py][logic] | Done | Included | | 7.1 | KB-Agent | KB_AgentProgram | [logic.py][logic] | Done | Included | | 7.7 | Propositional Logic Sentence | Expr | [utils.py][utils] | Done | Included | | 7.10 | TT-Entails | tt_entails | [logic.py][logic] | Done | Included | | 7.12 | PL-Resolution | pl_resolution | [logic.py][logic] | Done | Included | | 7.14 | Convert to CNF | to_cnf | [logic.py][logic] | Done | Included | | 7.15 | PL-FC-Entails? | plfcentails | [logic.py][logic] | Done | Included | | 7.17 | DPLL-Satisfiable? | dpll_satisfiable | [logic.py][logic] | Done | Included | | 7.18 | WalkSAT | WalkSAT | [logic.py][logic] | Done | Included | | 7.20 | Hybrid-Wumpus-Agent | HybridWumpusAgent | | | | | 7.22 | SATPlan | SAT_plan | [logic.py][logic] | Done | Included | | 9 | Subst | subst | [logic.py][logic] | Done | Included | | 9.1 | Unify | unify | [logic.py][logic] | Done | Included | | 9.3 | FOL-FC-Ask | folfcask | [logic.py][logic] | Done | Included | | 9.6 | FOL-BC-Ask | folbcask | [logic.py][logic] | Done | Included | | 10.1 | Air-Cargo-problem | air_cargo | [planning.py][planning] | Done | Included | | 10.2 | Spare-Tire-Problem | spare_tire | [planning.py][planning] | Done | Included | | 10.3 | Three-Block-Tower | threeblocktower | [planning.py][planning] | Done | Included | | 10.7 | Cake-Problem | havecakeandeatcake_too | [planning.py][planning] | Done | Included | | 10.9 | Graphplan | GraphPlan | [planning.py][planning] | Done | Included | | 10.13 | Partial-Order-Planner | PartialOrderPlanner | [planning.py][planning] | Done | Included | | 11.1 | Job-Shop-Problem-With-Resources | jobshopproblem | [planning.py][planning] | Done | Included | | 11.5 | Hierarchical-Search | hierarchical_search | [planning.py][planning] | Done | Included | | 11.8 | Angelic-Search | angelic_search | [planning.py][planning] | Done | Included | | 11.10 | Doubles-tennis | doubletennisproblem | [planning.py][planning] | Done | Included | | 13 | Discrete Probability Distribution | ProbDist | [probability.py][probability] | Done | Included | | 13.1 | DT-Agent | DTAgent | [probability.py][probability] | Done | Included | | 14.9 | Enumeration-Ask | enumeration_ask | [probability.py][probability] | Done | Included | | 14.11 | Elimination-Ask | elimination_ask | [probability.py][probability] | Done | Included | | 14.13 | Prior-Sample | prior_sample | [probability.py][probability] | Done | Included | | 14.14 | Rejection-Sampling | rejection_sampling | [probability.py][probability] | Done | Included | | 14.15 | Likelihood-Weighting | likelihood_weighting | [probability.py][probability] | Done | Included | | 14.16 | Gibbs-Ask | gibbs_ask | [probability.py][probability] | Done | Included | | 15.4 | Forward-Backward | forward_backward | [probability.py][probability] | Done | Included | | 15.6 | Fixed-Lag-Smoothing | fixedlagsmoothing | [probability.py][probability] | Done | Included | | 15.17 | Particle-Filtering | particle_filtering | [probability.py][probability] | Done | Included | | 16.9 | Information-Gathering-Agent | InformationGatheringAgent | [probability.py][probability] | Done | Included | | 17.4 | Value-Iteration | value_iteration | [mdp.py][mdp] | Done | Included | | 17.7 | Policy-Iteration | policy_iteration | [mdp.py][mdp] | Done | Included | | 17.9 | POMDP-Value-Iteration | pomdpvalueiteration | [mdp.py][mdp] | Done | Included | | 18.5 | Decision-Tree-Learning | DecisionTreeLearner | [learning.py][learning] | Done | Included | | 18.8 | Cross-Validation | cross_validation | [learning.py][learning]\* | | | | 18.11 | Decision-List-Learning | DecisionListLearner | [learning.py][learning]\* | | | | 18.24 | Back-Prop-Learning | BackPropagationLearner | [learning.py][learning] | Done | Included | | 18.34 | AdaBoost | AdaBoost | [learning.py][learning] | Done | Included | | 19.2 | Current-Best-Learning | currentbestlearning | knowledge.py | Done | Included | | 19.3 | Version-Space-Learning | versionspacelearning | knowledge.py | Done | Included | | 19.8 | Minimal-Consistent-Det | minimalconsistentdet | knowledge.py | Done | Included | | 19.12 | FOIL | FOIL_container | knowledge.py | Done | Included | | 21.2 | Passive-ADP-Agent | PassiveADPAgent | [rl.py][rl] | Done | Included | | 21.4 | Passive-TD-Agent | PassiveTDAgent | [rl.py][rl] | Done | Included | | 21.8 | Q-Learning-Agent | QLearningAgent | [rl.py][rl] | Done | Included | | 22.1 | HITS | HITS | [nlp.py][nlp] | Done | Included | | 23 | Chart-Parse | Chart | [nlp.py][nlp] | Done | Included | | 23.5 | CYK-Parse | CYK_parse | [nlp.py][nlp] | Done | Included | | 25.9 | Monte-Carlo-Localization | montecarlolocalization | [probability.py][probability] | Done | Included | Index of data structures Here is a table of the implemented data structures, the figure, name of the implementation in the repository, and the file where they are implemented. | Figure | Name (in repository) | File | |:-------|:--------------------------------|:--------------------------| | 3.2 | romania_map | [search.py][search] | | 4.9 | vacumm_world | [search.py][search] | | 4.23 | onedimstate_space | [search.py][search] | | 6.1 | australia_map | [search.py][search] | | 7.13 | wumpusworldinference | [logic.py][logic] | | 7.16 | hornclausesKB | [logic.py][logic] | | 17.1 | sequentialdecisionenvironment | [mdp.py][mdp] | | 18.2 | waitingdecisiontree | [learning.py][learning] | Acknowledgements Many thanks for contributions over the years. I got bug reports, corrected code, and other support from Darius Bacon, Phil Ruggera, Peng Shao, Amit Patil, Ted Nienstedt, Jim Martin, Ben Catanzariti, and others. Now that the project is on GitHub, you can see the contributors who are doing a great job of actively improving the project. Many thanks to all contributors, especially @darius, @SnShine, @reachtarunhere, @antmarakis, @Chipe1, @ad71 and @MariannaSpyrakou. [agents]:../master/agents.py [csp]:../master/csp.py [games]:../master/games.py [grid]:../master/grid.py [knowledge]:../master/knowledge.py [learning]:../master/learning.py [logic]:../master/logic.py [mdp]:../master/mdp.py [nlp]:../master/nlp.py [planning]:../master/planning.py [probability]:../master/probability.py [rl]:../master/rl.py [search]:../master/search.py [utils]:../master/utils.py [text]:../master/text.py

Vibe Coding FULL Course + WIN MacBook Pro, PlayStation 5 🔥
youtube
LLM Vibe Score0.309
Human Vibe Score0.49
Ishan SharmaMar 27, 2025

Vibe Coding FULL Course + WIN MacBook Pro, PlayStation 5 🔥

I’m organising a UI Hackathon with Outlier that you can participate in: https://bit.ly/uihacks25?utmsource=youtube&utmmedium=paid&pod=coders You simply have to pick a prompt from the ideas list and build a great functional UI on it. Like we built the https://peachpup.vercel.app/ app today in this video with vibe coding. And you can submit your app on this Typeform link: https://form.typeform.com/to/Hljx9wab?utmsource=youtube&utmmedium=paid&pod=coders The top 3 prizes include M4 MacBook Pro, PlayStation 5, and Rayban Meta Glasses. And the top 1% coders will get a chance to work part-time and make up to $27 per hour if you’re in India or $50 per hour anywhere else in the world. You will be judged on the UI you create and how functional it is. The deadline for the submission is Sunday, 30th March I'm sure you must have heard about the word “vibe coding”. Vibe coding is the most practical way to learn coding in today's time. You can build apps and products by just describing your idea in text, and the AI will produce code and do everything on its own. This is really important for a software developer to learn to get to the product quickly and test out your knowledge about coding. In this video, I explain how you can build real apps using tools like Cursor, Replit, Lovable, 10x faster. This can help you become a top-tier developer by developing any application you want in seconds. Watch the video till the end, and don't miss out on the hackathon. 📸 Instagram: https://bit.ly/ishansharma7390ig Join MarkitUpX Discord Server: https://discord.gg/fwSpTje4rh CHAPTERS: 00:00 - Introduction 01:19 - What is Vibe Coding? 02:36 - Tools to get started with Vibe Coding 03:30 - CHECK OUT THIS HACKATHON 08:10 - Building from Scratch 10:54 - Deciding on the Idea 13:09 - Getting Started 18:15 - Step-by-Step Tutorial 01:12:48 - Deploying 01:20:46 - Conclusion 😁 About Me: https://bit.ly/aboutishansharma 📱 Twitter: https://bit.ly/ishansharma7390twt 📝 LinkedIn: https://bit.ly/ishansharma7390li 🌟 Please leave a LIKE ❤️ and SUBSCRIBE for more AMAZING content! 🌟 3 Books You Should Read 📈Psychology of Money: https://amzn.to/30wx4bW 👀Subtle Art of Not Giving a F: https://amzn.to/30zwWbP 💼Rework: https://amzn.to/3ALsAuz Tech I use every day 💻MacBook Air M1: https://amzn.to/2YWKPjG 📺LG 29' Ultrawide Monitor: https://amzn.to/3aG0p5p 🎥Sony ZV1: https://amzn.to/3ANqgDb 🎙Blue Yeti Mic: https://amzn.to/2YYbiNN ⽴Tripod Stand: https://amzn.to/3mVUiQc 🔅Ring Light: https://amzn.to/2YQlzLJ 🎧Marshall Major II Headphone: https://amzn.to/3lLhTDQ 🖱Logitech mouse: https://amzn.to/3p8edOC 💺Green Soul Chair: https://amzn.to/3mWIxZP ✨ Tags ✨ coding,coding hackathon,vibe coding,vibe coding explained,how to use replit,how to use cursor,how to build app on replit,lovable,web development,app development,build with ai,no code ai tools,artificial intelligence,cursor ai tutorial,andrej karpathy,coding hackathon 2025,ai coding hackathon,frontend ui hackathon,claude,how to build app without coding,build app with cursor ai,build app with no code,best no code app builder,BUILD Apps in Minutes w/ Cursor ✨ Hashtags ✨ #vibecoding #coding #artificialintelligence

machine-learning-blackjack-solution
github
LLM Vibe Score0.42
Human Vibe Score0.022610872675250356
GregSommervilleMar 27, 2025

machine-learning-blackjack-solution

machine-learning-blackjack-solution Introduction A genetic algorithm is a type of artificial intelligence programming that uses ideas from evolution to solve complex problems. It works by creating a population of (initially random) candidate solutions, then repeatedly selecting pairs of candidates and combining their solutions using a process similar to genetic crossover. Sometimes candidate solutions even go through mutation, just to introduce new possibilities into the population. After a large number of generations, the best solution found up to that point is often the optimal, best solution possible. Genetic algorithms are particularly well-suited for combinatorial problems, where there are huge numbers of potential solutions to a problem. The evolutionary process they go through is, in essence, a search through a huge solution space. A solution space so large that you simply could never use a brute force approach. This project is a demonstration of using a genetic algorithm to find an optimal strategy for playing the casino game Blackjack. Please see this article for a story about how this program was used, and what the results were. The article describes some of the available settings, and shows how different values for those settings affect the final result. The source code is for a Windows application written in Cthat allows you to play with different settings like population size, selection style and mutation rate. Each generation's best solution is displayed, so you can watch the program literally evolve a solution. !blackjack strategy tester screenshot The property grid located at the upper left of the screen is where you adjust settings. There's an informational area below that, and the right side of the screen is the display area for the three tables that represent a strategy for playing Blackjack. The tall table on the left is for hard hands, the table in the upper right is for soft hands, and the table in the lower right is for pairs. We'll talk more about how to interpret this strategy in a bit. The columns along the tops of the three tables are for the dealer upcard. When you play Blackjack the dealer has one of his two cards initially turned face up, and the rank of that card has a big impact on recommended strategy. Notice that the upcard ranks don't include Jack, Queen or King. That's because those cards all count 10, so we group them and the Ten together and simplify the tables. To use the tables, first, determine if you have a pair, soft hand, or hard hand. Then look in the appropriate table, with the correct dealer upcard column. The cell in the table will be "H" when the correct strategy is to hit, "S" when the correct strategy is to stand, "D" for double-down, and (in the pairs table only) "P" for split. A Word About This "Optimal" Strategy Before we go any further, it needs to be stated that this problem of finding an optimal Blackjack strategy has already been solved. Back in the 1960s, a mathematician named Edward O. Thorp authored a book called Beat the Dealer, which included charts showing the optimal "Basic" strategy. That strategy looks like this: !optimal blackjack strategy So we're solving a problem that has already been solved, but that's actually good. That means we can compare our results to the known best solution. For example, if our result strategy tells us to do anything but stand when holding a pair of Tens, Jacks, Queens or Kings, we know there's a problem. There's one other thing to get out of the way before we go any further, and that's the idea of nondeterministic code. That means that if we run the same code twice in a row, we're likely to get two different results. That's something that happens with genetic algorithms due to their inherent randomness. There's no guarantee you'll find the absolute optimal solution, but it is assured that you will find an optimal or near-optimal solution. It's something that isn't typical when writing code, so it takes some adjustment for most programmers. Genetic Algorithms Now let's talk about the details of a genetic algorithm. Fitness Scores First of all, we need a way to evaluate candidates so we can compare them to each other. That means a numeric fitness score, which in this case is quite simple: you simulate playing a certain number of hands using the strategy, and then count the number of chips you have at the end. The big question is, how many hands should we test with? The challenge of trying to test a strategy is that due to the innate randomness of Blackjack, you could use the same strategy ten times and get ten completely different results. Obviously, the more hands you play, the more the randomness gets smoothed out, and the quality of the underlying strategy starts to emerge. If you doubt this, just think about flipping a coin. If you only flip it five times, there's certainly a possibility that it'll come up heads all five times (in fact, that happens just over 3% of the time). However, if you flip it 500 times, there's no way it's going to end up all heads - the odds of it happening are 0.5500, which works out to be roughly once every 3 x 10150 times you try it. After some testing and analysis, it was determined that a minimum of 100,000 hands per test is needed for a reasonable level of accuracy. There's still variance even at that number, but in order to cut the variance in half, you'd need to bump the number of hands to 500,000. One reason this accuracy is important is that in the later generations, the differences between candidates are very small. Evolution has caused the main parts of the strategy to converge on a particular approach, and towards the end all it's doing is refining the minor details. In those cases it's important to accurately determine the difference between two similar candidates. Representation Representation is simply the idea that we need to use a data structure for a candidate solution that can be combined via crossover, and possibly mutated. In this case, that's also quite simple because the way that human beings represent a Blackjack strategy is to use three tables, as we've seen. Representing those in code with three two-dimensional arrays is the obvious approach. Each cell in those three tables will have "Hit", "Stand", "Double-Down", or (only for pairs) "Split". By the way, since there are 160 cells in the hard hands table, and 80 cells in the soft hands table, and 100 cells in the pairs table, we can calculate exactly how many possible distinct strategies there are for Blackjack: 4100 x 380 x 3160 = 5 x 10174 possible Blackjack strategies That's a big number, which is obviously impossible to search using brute force. Genetic algorithms (GAs) are extremely helpful when trying to find an optimal solution from a very large set of possible solutions like this. Blackjack Rules and Strategies The rules of Blackjack are fairly simple. The dealer and the player both are dealt two cards. The player sees both of their cards (they are usually dealt face up), and one of the dealer's cards is dealt face up. Each card has a value - for cards between 2 and 10, the value is the same as the card's rank (so an Eight of Spades counts as 8, for example). All face cards count as 10, and an Ace can either be 1 or 11 (it counts as 11 only when that does not result in a hand that exceeds 21). The suit of a card does not matter. After the cards are dealt, if the player has Blackjack (a total of 21) and the dealer does not, the player is immediately paid 1.5 times their original bet, and a new hand is dealt. If the player has 21 and the dealer does also, then it's a tie and the player gets their original bet back, and a new hand is dealt. If the player wasn't dealt a Blackjack, then play continues with the player deciding whether to Stand (not get any more cards), Hit (receive an additional card), Double-down (place an additional bet, and receive one and only one more card), or, in the case of holding a pair, splitting the hand, which means placing an additional bet and receiving two new cards, so the end result is that the player is now playing two (or, in the case of multiple splits, more than two) hands simultaneously. If the player hits or double-downs and has a resulting hand that exceeds 21, then they lose and play continues with the next hand. If not, then the dealer draws until their hand totals at least 17. If the dealer exceeds 21 at this point, the player receives a payment equal to twice their original bet. If the dealer doesn't exceed 21, then the hands are compared and the player with the highest total that doesn't exceed 21 wins. Because of these rules, certain effective strategies emerge. One common strategy is that if you hold a hard hand with a value of 20, 19 or 18, you should Stand, since you avoid busting by going over 21, and you have a nice hand total that might win in a showdown with the dealer. Another common strategy is to split a pair of Aces, since Aces are so powerful (due to the fact that count as 11 or 1, you can often Hit a hand with a soft Ace with no risk of busting). Likewise, splitting a pair of 8s is a good idea because with a hard total of 16, it's likely you will bust if you take a Hit (since so many cards count as 10). As a human being, all it takes is a little knowledge about the rules in order to construct a strategy. The GA program doesn't have that advantage, and operates completely without any pre-programmed knowledge of Blackjack. It simply uses the relative fitness scores and the mechanism of evolution to find the solution. GA Settings There are many variables or settings for a GA. You can adjust population size, how parent candidates are selected, how the resulting children may be mutated, and several other items. The following sections describe some of these settings: Setting: Selection Style Once we've solved representation and have a fitness function, the next step is to select two candidates for crossover during the process of building a new generation. There are three common styles for selection, and this program supports all of them. First, you can choose Roulette Wheel selection. It's named for a Roulette wheel because you can imagine each candidate's fitness score being a wedge in a pie chart, with a size proportionate to its relative fitness compared to the other candidates. (Of course, this assumes that all fitness scores are positive, which we will talk about shortly). The main benefit of Roulette Wheel selection is that selection is fitness-proportionate. Imagine if you had only three candidates, with fitness scores of 1, 3, and 8. The relative selection probabilities for those candidates will be 1/12, 3/12, and 8/12. The downside of Roulette Wheel selection is that it tends to be somewhat slow in terms of processing. The selection process is done by iterating through the candidates until a particular condition is matched - in other words, O(N) performance. Another potential problem with Roulette Wheel selection is that there may be situations where fitness scores vary widely, to such an extent that only certain candidates have any reasonable chance of being selected. This happens frequently in early generations, since the majority of candidates are mostly random. Although this might sound like a positive (since you ultimately want to select candidates with high fitness scores), it also results in a loss of genetic diversity. In other words, even though a particular candidate may have a low fitness score in an early generation, it may contain elements that are needed to find the ultimate solution in later generations. Ranked Selection is the solution to this problem. Instead of using raw fitness scores during the selection process, the candidates are sorted by fitness, with the worst candidate receiving a score of 0, the second worse receiving 1, and so forth, all the way to the best candidate, which has a score equal to the population size - 1. Ranked Selection is quite slow, since it combines the O(N) performance of Roulette Wheel, with the additional requirement that the candidates be sorted before selection. However, there may be circumstances where it performs better than other selection approaches. Finally, the fastest selection method of all is called Tournament Selection. This method simply selects N random candidates from the current generation, and then uses the one with the best fitness score. A tournament size of 2 means two random candidates are selected, and the best of those two is used. If you have a large tournament size (like 10), then 10 different candidates will be selected, with the best of those being the ultimate selection. That obviously tilts the balance between randomness and quality. Tournament selection works well in most cases, but it does require some experimentation to find the best tourney size. Setting: Elitism Elitism is a technique that helps ensure that the best candidates are always maintained. Since all selection methods are random to some degree, it is possible to completely lose the best candidates from one generation to another. By using Elitism, we automatically advance a certain percentage of the best candidates to the next generation. Elitism does have a negative impact on performance since all of the candidates must be sorted by fitness score. Typically Elitism is done before filling the rest of a new generation with new candidates created by crossover. Crossover Details Once two candidate solutions have been selected, the next step in building a new generation is to combine those two into a single new candidate, hopefully using the best of both parent strategies. There are a number of ways to do crossover, but the method used in this program is quite straightforward - the two fitness scores are compared, and crossover happens in a relatively proportionate way. If one candidate has a fitness of 10, and the other has a fitness of 5, then the one with fitness 10 contributes twice as much to the child as the parent with a fitness of 5. Since the fitness scores in this program are based on how much the strategy would win over thousands of hands, almost all fitness scores will be negative. (This is obviously because the rules are set up so the house always wins.) This makes it difficult to calculate relative fitnesses (how do you compare a positive number with a negative, and find relative proportions?), and also causes problems with selection methods like Roulette Wheel or Ranked. To solve this, we find the lowest fitness score of the generation and add that value to each candidate. This results in an adjusted fitness score of 0 for the very worse candidate, so it never gets selected. Mutation As has been mentioned a few times, maintaining genetic diversity in our population of candidate solutions is a good thing. It helps the GA ultimately find the very best solution, by occasionally altering a candidate in a positive direction. There are two settings for mutation. MutationRate controls what percentage of new candidates have mutation done on them. MutationImpact controls what percentage of their strategy is randomized. Population Size Population size has a significant impact on performance. The smaller the population size, the faster the GA will execute. On the other hand, if the size is too low the population may not have enough genetic diversity to find the ultimate solution. During testing, it looks like 700 to 1000 is a good balance between speed and correctness. Performance Notes This program consumes a lot of processing power. Running tests of hundreds of thousands of hands of Blackjack for hundreds or thousands of candidates consumes a lot of time. It's really imperative to write the code so that it works as efficiently as possible. If your CPU isn't consistently at or above 95% usage, there's still room for improvement. Multi-threading is a natural fit for genetic algorithms because we often want to perform the same action on each candidate. The best example of this is when we calculate fitness scores. This is often an operation that takes quite a bit of time. In our case, we're dealing out 100,000 hands, and each hand has to be played until the end. If we're single-threading that code, it's going to take a long time. Multi-threading is really the way to go. Luckily, there's a ridiculously simple way to efficiently use all of your processors for an operation like this. This code loops over all of the candidates in the currentGeneration list, calls the fitness function and sets the fitness property for each: Regardless of the number of items in the list or the number of processors on your machine, the code will efficiently run the code in a multi-threaded manner, and continue only when all of the threads are complete. One of the side effects of making this code multi-threaded is that all of the code relating to evaluating a candidate must be thread-safe, including any Singleton objects. When making code thread-safe, pay attention that you don't accidentally introduce code that will slow your program down unintentionally, because sometimes it can be quite subtle. Random numbers are central to how genetic algorithms work, so it's critical that they can be used correctly from a multithreaded environment. That means that each random number generator must be separate from the others, and it also means that each must produce a distinct series of random numbers. Random number generators use seed values which are usually time-based, like the number of milliseconds the computer has been turned on. Starting with that seed, subsequent calls will return a series of numbers that look random, but really aren't. If you start with the same seed, you get the same sequence. And that's a problem because if you create multiple random number generator objects in a loop using the default time-based seed, several of them will have the same time-based initial seed value, which will result in the same sequence of "random" numbers. That's a bug, because it can reduce the true randomness of the program a great deal, and that's vital to a genetic algorithm. There are a couple of ways to solve this problem. First, you can make the random object truly a singleton, and restrict access to it by using a Clock statement. The makes all access serialized for any random number need, which reduces performance. Another approach is to make the variable static per thread. By declaring the variable as static and also marking it with the [ThreadStatic] attribute, the .NET runtime allocates one static variable per thread. That eliminates the locking/serialization, but also has performance issues. The approach used in this application is to use a non-default seed value. In this case we call Guid.NewGuid().GetHashCode(), which generates a new, unique GUID, then gets an integer hashcode value that should be unique, depending on how GetHashCode is implemented. While multithreading really helps performance, there are also other things we can do to improve performance. For example, when dealing with large populations, the hundreds or thousands of objects that will be generated each generation can quickly turn into a huge problem related to garbage collection. In the end, the easiest way to solve that is to look through the code and find objects being allocate inside a loop. It's better to declare the variable outside of the loop, and then clear it in the loop, rather than reallocate it. In a program like this one where you could be looping hundreds of thousands of times, this can result in a very significant performance boost. For example, in an early version of this code, a Deck object was created for each hand. Since there are hundreds of candidate solutions running hundreds of thousands of trial hands, this was a huge inefficiency. The code was changed to allocate one deck per test sequence. The deck was shuffled as needed, so it never needs to be reallocated. Beyond the cards in the deck, another object type that was repeatedly created and destroyed were the candidate strategies. To mitigate this problem, a StrategyPool class was created that handles allocation and deallocation. This means that strategy objects are reused, rather than dynamically created when needed. The pool class has to be thread-safe, so it does serialize access to its methods via a Clock statement, but overall using the pool approach produced a good performance increase. Finally, a subtle form of object allocation is conversion. In an early version of the code, a utility card function used Convert.ToInt32(rankEnum). Obviously, the easiest way to convert from an enum to an int is simply to cast it, like (int)rankEnum. But it's hard to know exactly what the difference is between that approach, int.Parse(), int.TryParse(), or Convert.ToInt32(), since they can all be used and are roughly equivalent. Perhaps the compiler was boxing the enum value before passing it to Convert.ToInt32(), because the profiler identified this as a function that had large amounts of thread contention waiting - and the problem got much, much worse as the generations passed. By rewriting the conversion to use a simple cast, the program performance increased threefold (3x). Contributing Please read CONTRIBUTING.md for details on our code of conduct, and the process for submitting pull requests to us. Author Greg Sommerville - Initial work* License This project is licensed under the Apache 2.0 License - see the LICENSE.md file for details

bubbln_network-automation
github
LLM Vibe Score0.421
Human Vibe Score0.004537250556463098
olasupoMar 14, 2025

bubbln_network-automation

Bubbln: An AI-driven Network Automation In the world of network engineering, automation has completely transformed the way things work. But, before automation, setting up and managing networks was a tedious job filled with challenges. Engineers had to manually type out configurations, often doing the same tasks repeatedly on different devices. This led to mistakes and wasted time. Then came automation tools like Ansible, Chef, and Puppet, which changed everything. They made network management much easier and allowed for scalability. But there was still a problem: creating automation scripts required a lot of technical know-how and was prone to errors because it relied on human input. And that's why we built Bubbln. It's a game-changer in network engineering, integrating AI into Ansible to take automation to the next level. With Bubbln, we can automatically generate and execute playbooks with incredible accuracy, thereby improving automation efficiency and increasing network engineer’s productivity. It was developed using Python programming language and acts as a bridge between ChatGPT and network systems, making interactions seamless and deployments effortless. Current Capabilities AI-Driven Playbook Generation for OSPF and EIGRP based networks: Bubbln has been rigorously tested to leverage ChatGPT for generation of playbooks for networks based on OSPF and EIGRP networks, with a very high accuracy rate. Auto-creation of Inventory files: Users do not need to prepare the hosts file. Bubbln will auto-generate this file from input provided by the user. Customizable Configurations: Users can input specific router protocols (OSPF or EIGRP), interface configurations, and other network details to tailor the generated playbooks. Documentation: Bubbln automatically creates a report that contains the network configurations, prompts, and generated playbooks for easy reference in future. No expertise required: By auto-generation of the playbooks and inventory file, Bubbln has been able to eliminate a major hurdle to network automation – need for users to learn the automation tools e.g Ansible, Chef. Improved Efficiency: With AI automation, Bubbln speeds up the deployment of network configurations, reducing the time required for manual playbook creation, thereby increasing the productivity of network engineers. Getting Started There are two main approaches to installing Bubbln on your local machine. Docker Container Bubbln has been packaged using docker containers for easy distribution and usage. The following steps can be followed to deploy the Bubbln container on your local machine. Ensure docker is installed on your local machine by entering the below command. This command works for windows and linux OS: The version of docker would be displayed if it is installed. Otherwise, please follow the link below to install docker on your machine: Windows: Docker Desktop for Windows Ubuntu: Docker Engine for Ubuntu CentOS: Docker Engine for CentOS Debian: Docker Engine for Debian Fedora: Docker Engine for Fedora Download the docker image: Create a directory for the project and download Bubbln image using the below command: Run the docker container using the below command: Install nano Update the sshipaddresses.txt file: Update the ssh_addresses.txt file with the SSH IP addresses of the routers you want to configure. Bubbln will utilize this information along with the login credentials (inputted at runtime) to automatically generate a hosts.yml file required by ansible for network configuration. To do this enter the below command to edit the file: Obtain an OpenAPI API Key: You may follow this guide to sign up and obtain an API key: Utilizing a Virtualization machine of choice, setup a network with the following basic configurations: Enable SSH on each of the routers. Configure IP addresses and enable only interfaces required for connectivity by Bubbln. Configure static routes to enable Bubbln reach the routers on the network. Ensure all the routers can be reached by ping and SSH from your host machine. Initialize Bubbln by entering the below command: Github Repository Clone You can clone Bubbln’s GitHub repository by following the below steps: Prerequisites Bubbln works well with Python 3.10. You need to ensure python3.10 is installed on your local machine. This can be confirmed by entering the below command: If it is not Installed, then the below command can be utilized to install python 3.10: Build and Prepare the Project Clone the Bubbln repository from GitHub: To clone the repository, first verify you have git installed on your machine by issuing the following commands: If git is installed, the version number would be displayed, otherwise, you can issue the following commands to have git installed on your machine: Navigate or create a directory for the project on your machine and issue the following commands to clone the Bubbln git repository: Create a Virtual Environment for the application Firstly, confirm virtualenv is installed on your machine by inputting the following command: If the output shows something similar to the below, then go to the next step to install virtualenv ` WARNING: Package(s) not found: env, virtual ` Issue the below command to install virtualenv: Create a virtual environment for the project: Activate the virtual environment: Install the dependencies You can then run the below command to install the necessary packages for the app. Update the sshipaddresses.txt file: Update the ssh_addresses.txt file with the SSH IP addresses of the routers you want to configure. Bubbln will utilize this information along with the login credentials (inputted at runtime) to automatically generate a hosts.yml file required by ansible for network configuration. Obtain an OpenAPI API Key: You may follow this guide to sign up and obtain an API key OpenAI Key: OpenAI Key Utilizing a Virtualization machine of choice, setup a network with the following basic configurations: Enable SSH on each of the routers. Configure IP addresses and enable only interfaces required for connectivity by Bubbln Configure static routes to enable Bubbln reach the routers on the network. Ensure all the routers can be reached by ping and SSH from your host machine. Initialize Bubbln While ensuring that python virtual environment is activated as stated in step 5, run the below command to initialize Bubbln How Bubbln Works Bubbln serves as an intermediary between ChatGPT and a network infrastructure, providing logic, control functions, and facilitating network automation. Its operation can be summarized as follows: !image Figure 1Bubbln architecture and interaction with a network of four routers. Initialization: When Bubbln is initialized, it checks the “userconfig.pkl” file to see if Bubbln has ever been initiated. This is indicated by the presence of a welcome message status in the file. If it exists, Bubbln jumps straight to request the user to input the OpenAI key. Otherwise, it displays a welcome message, and updates the userconfig.pkl file accordingly. Upon successful input of the API key, the user is prompted for the SSH credentials of the routers. These parameters are then encrypted and saved in the user_config.pkl file. The SSH credential is later decrypted and parsed as input to dynamically generate a hosts.yml file at runtime. Responsible Code Section: bubbln.py: welcomemessagefeature() !image Figure 2 Bubbln's welcome message. Parameter Input & Validation: In the parameter input stage, Bubbln first checks for the existence of a file called “router_configuration.pkl”. If it exists, the user is prompted to decide whether to load an existing configuration or input a new set of configurations. If the file is empty or non-existent, then users are prompted to input the configuration parameters for each router on the network. These parameters serve as variables that are combined with hardcoded instructions written in natural language to form the prompt sent to ChatGPT. Key parameters include: Router Configurations: OSPF Area OSPF Process ID Number of networks to advertise (OSPF/EIGRP) AS Number (EIGRP) Interface names IP Addresses (in CIDR format) This module also ensures that parameters are keyed in using the correct data type and format e.g. IP addresses are expected in CIDR format and OSPF Area should be of type integer. Upon completion of parameter input, all parameters are saved into a file called “router_configuration.pkl” upon validation of accuracy by the user. Responsible Code Section: parameter_input.py !image Figure 3 Bubbln receiving Network Parameters. Before generating the prompt, a summary of the inputted parameters is displayed for user validation. This step ensures accuracy and minimizes errors. Users are given the option to make corrections if any discrepancies are found. Responsible Code Section: parameterinput.py: validateinputs() !image Figure 4 Bubbln Awaiting Validation of Inputted Network Parameters. Auto-Generation of Prompt: After validation of inputted parameters, Bubbln composes the prompt by combining the inputted parameters with a set of well-engineered hardcoded instructions written in natural language. Responsible Code Section: prompt_generator.py ChatGPT Prompting: The auto-composed prompt is then sent to ChatGPT utilizing gpt-4 chatCompletions model with a temperature parameter of 0.2 and maximum tokens of 1500. The following functions were designed into this process stage Responsible Code Section: chatGPT_prompting.py !image Figure 5 ChatGPT prompting in progress Playbook Generation & Extraction: After ChatGPT processes the prompt from Bubbln, it provides a response which usually contains the generated playbook and explanatory notes. Bubbln then extracts the playbook from the explanatory notes by searching for “---” which usually connotes the start of playbooks and saves each generated playbook uniquely using the nomenclature RouteriPlaybook.yml. Responsible Code Section: playbook_extractor.py !image Figure 6 ChatGPT-generated playbook. Playbook Execution: Bubbln loads the saved “RouteriPlaybook.yml” playbook and dynamically generates the hosts.yml file and parses them to the python library ansiblerunner for further execution on the configured network. Bubbln generates the hosts.yml file at run time by using the pre-inputted SSH credentials in userconfig.pkl file - and decrypts them, as well as IP addresses from the sshipaddresses.txt file, as inputs Responsible Code Section: playbook_execution.py !image Figure 7 Playbook execution in progress Sample result of Executed Playbook Upon successful execution of all playbooks, a query of the routing table on router 4 indicates that router 4 could reach all the prefixes on the network. !image Figure 8 Output of 'sh ip route' executed on R1 File Management and Handling Throughout the execution process, Bubbln manages the creation, saving, and loading of various files to streamline the network automation process. user_config.pkl: This dictionary file dynamically created at run time is used to store encrypted API keys, SSH credentials and initial welcome message information. router_configuration.pkl: It is auto created by Bubbln and used to store network configuration parameters for easy loading during subsequent sessions. hosts.yml: This is a runtime autogenerated file that contains inventory of the network devices. It is auto deleted after the program runs. networkconfigurationreport.pdf: This auto-generated report by Bubbln is a documentation of all the routers configured their parameters, generated playbooks, and prompt for each execution of the Bubbln application. It is created after a successful execution of playbooks and network testing and is meant for auditing and documentation purposes. RouteriPlaybook.yml: After extraction of generated playbooks from ChatGPT’s raw response, Bubbln automatically saves a copy of the generated playbook using unique names for each playbook. !image Figure 9 File structure after successful deployment of a four-router network Providing Feedback We are glad to hear your thoughts and suggestions. Kindly do this through the discussion section of our GitHub - https://github.com/olasupo/bubbln_network-automation/discussions/1#discussion-6487475 We can also be reached on: Olasupo Okunaiya – olasupo.o@gmail.com

Coding Is OVER!🤯 Replit AI Agent Builds Apps In Minutes! Vibe Coding Explained
youtube
LLM Vibe Score0.422
Human Vibe Score0.9
Ishan SharmaFeb 22, 2025

Coding Is OVER!🤯 Replit AI Agent Builds Apps In Minutes! Vibe Coding Explained

Check out the apps I built: 📚 Learning App: https://learn-flash-master-ishanclips7390.replit.app/ 💪 Fitness Tracker: https://fitness-companion-ishanclips7390.replit.app/ 💰 Finance Tracker: https://mindful-spendings.lovable.app/ In this video, I'll show you 2 powerful and completely free AI tools that will help you build professional applications without any coding knowledge! Instead of spending hours writing complex code, you can now simply describe what you want to build, while AI takes care of the technical stuff. This new approach, called "Vibe Coding," is a great way to bring your ideas to life. Watch the full tutorial to learn how easily you can start building your own apps today. CHAPTERS: 00:00 - Introduction 01:17 - Replit: AI Tool 1 01:45 - Creating a Learning App 07:56 - Lovable: AI Tool 2 08:14 - Creating a Finance Tracker 10:58 - More Examples 12:47 - Conclusion 📸 Instagram: https://bit.ly/ishansharma7390ig Join MarkitUpX Discord Server: https://discord.gg/fwSpTje4rh 😁 About Me: https://bit.ly/aboutishansharma 📱 Twitter: https://bit.ly/ishansharma7390twt 📝 LinkedIn: https://bit.ly/ishansharma7390li 🌟 Please leave a LIKE ❤️ and SUBSCRIBE for more AMAZING content! 🌟 3 Books You Should Read 📈Psychology of Money: https://amzn.to/30wx4bW 👀Subtle Art of Not Giving a F: https://amzn.to/30zwWbP 💼Rework: https://amzn.to/3ALsAuz Tech I use every day 💻MacBook Air M1: https://amzn.to/2YWKPjG 📺LG 29' Ultrawide Monitor: https://amzn.to/3aG0p5p 🎥Sony ZV1: https://amzn.to/3ANqgDb 🎙Blue Yeti Mic: https://amzn.to/2YYbiNN ⽴Tripod Stand: https://amzn.to/3mVUiQc 🔅Ring Light: https://amzn.to/2YQlzLJ 🎧Marshall Major II Headphone: https://amzn.to/3lLhTDQ 🖱Logitech mouse: https://amzn.to/3p8edOC 💺Green Soul Chair: https://amzn.to/3mWIxZP ✨ Tags ✨ ishan sharma,ai agents,ai agents explained,ai agents 2025,ai assistant,ai agents tutorial,ai agents full guide,ai agent,ai,artificial intelligence,ai agents use cases,replit ai agent,lovable ai tutorial,replit ai tutorial,build app with ai,build app without coding,ai website builder,coding with AI,lovable,lovable tutorial,web development,replit ai agent tutorial,vibe coding,vibe coding tutorial,vibe coding ai,no code app builder,no code, Coding Is OVER! Replit AI Agent Builds Apps In Minutes! Vibe Coding Explained ✨ Hashtags ✨ #ai #aitools #coding

How I'd Use AI in 2025 (If I Could Start Over)
youtube
LLM Vibe Score0.415
Human Vibe Score0.86
Ishan SharmaFeb 12, 2025

How I'd Use AI in 2025 (If I Could Start Over)

Check out the Artificial Intelligence and Machine Learning Courses by Simplilearn: https://bit.ly/Ishan-AIML With tools like Gemini, DeepSeek, Perplexity, NotebookLM, and many others that are exploding in 2025, it's becoming insanely easier to get things done faster and better. It would be a very long and tiring video if I started talking about every single AI tool on the rise. However, a better option is to talk about how you can actually use these AI tools in your work to achieve maximum output in the shortest period. And that's what you'll be learning today through this video. I've shared a complete step-by-step guide that will give you a better understanding of using AI, resources, and tools to help you get started. This is the perfect time to experiment and experience where AI can actually help us. 📸 Instagram: https://bit.ly/ishansharma7390ig Join MarkitUpX Discord Server: https://discord.gg/fwSpTje4rh CHAPTERS: 00:00 - Introduction 02:00 - Step 1 05:36 - Step 2 07:15 - Step 3 09:42 - Conclusion 😁 About Me: https://bit.ly/aboutishansharma 📱 Twitter: https://bit.ly/ishansharma7390twt 📝 LinkedIn: https://bit.ly/ishansharma7390li 🌟 Please leave a LIKE ❤️ and SUBSCRIBE for more AMAZING content! 🌟 3 Books You Should Read 📈Psychology of Money: https://amzn.to/30wx4bW 👀Subtle Art of Not Giving a F: https://amzn.to/30zwWbP 💼Rework: https://amzn.to/3ALsAuz Tech I use every day 💻MacBook Air M1: https://amzn.to/2YWKPjG 📺LG 29' Ultrawide Monitor: https://amzn.to/3aG0p5p 🎥Sony ZV1: https://amzn.to/3ANqgDb 🎙Blue Yeti Mic: https://amzn.to/2YYbiNN ⽴Tripod Stand: https://amzn.to/3mVUiQc 🔅Ring Light: https://amzn.to/2YQlzLJ 🎧Marshall Major II Headphone: https://amzn.to/3lLhTDQ 🖱Logitech mouse: https://amzn.to/3p8edOC 💺Green Soul Chair: https://amzn.to/3mWIxZP ✨ Tags ✨ ishan sharma,ai,ml,artificial intelligence,machine learning,ai engineering,ai career,ai ml jobs,machine learning jobs,machine learning career,how to become ai ml engineer,how to become ai engineer,developer,development,ai developer,ml developer,how to be an ai dev,how to become an ai engineer,ai developer roadmap,ai engineer roadmap,ai developer course,ai developer guide,ai for beginners,how to learn ai,free courses,ai courses,ml courses ✨ Hashtags ✨ #ai #artificialintelligence #aitools

15 EASY Business Ideas to GET RICH with AI Agents in 2025
youtube
LLM Vibe Score0.432
Human Vibe Score0.73
Ishan SharmaFeb 3, 2025

15 EASY Business Ideas to GET RICH with AI Agents in 2025

Check out the AI Income Ideas Guide by HubSpot Today: https://clickhubspot.com/6rig In 2025, so many companies want to automate their work and they are looking for tools that can help them with that. Imagine earning your first thousand dollars just by helping a business automate its repetitive tasks with AI agents. In this video, I’ve shared 15 business ideas that you can start using AI agents and help companies in different aspects and earn money doing that! You don't need years of technical skills to master this - you just need to understand how to solve real problems with AI. And right now, there's so much opportunity because most people are not even aware of this. Watch the video till the end and take notes 📸 Instagram: https://bit.ly/ishansharma7390ig Join MarkitUpX Discord Server: https://discord.gg/fwSpTje4rh CHAPTERS: 00:00 - Introduction 00:49 - Business Idea 1 01:53 - Business Idea 2 03:02 - Business Idea 3 04:40 - Business Idea 4 05:31 - Business Idea 5 06:05 - Business Idea 6 07:38 - Business Idea 7 09:00 - Business Idea 8 10:18 - Business Idea 9 10:52 - Business Idea 10 12:07 - Business Idea 11 13:45 - Business Idea 12 15:55 - Business Idea 13 16:52 - Business Idea 14 17:38 - Business Idea 15 18:42 - Conclusion 😁 About Me: https://bit.ly/aboutishansharma 📱 Twitter: https://bit.ly/ishansharma7390twt 📝 LinkedIn: https://bit.ly/ishansharma7390li 🌟 Please leave a LIKE ❤️ and SUBSCRIBE for more AMAZING content! 🌟 3 Books You Should Read 📈Psychology of Money: https://amzn.to/30wx4bW 👀Subtle Art of Not Giving a F: https://amzn.to/30zwWbP 💼Rework: https://amzn.to/3ALsAuz Tech I use every day 💻MacBook Air M1: https://amzn.to/2YWKPjG 📺LG 29' Ultrawide Monitor: https://amzn.to/3aG0p5p 🎥Sony ZV1: https://amzn.to/3ANqgDb 🎙Blue Yeti Mic: https://amzn.to/2YYbiNN ⽴Tripod Stand: https://amzn.to/3mVUiQc 🔅Ring Light: https://amzn.to/2YQlzLJ 🎧Marshall Major II Headphone: https://amzn.to/3lLhTDQ 🖱Logitech mouse: https://amzn.to/3p8edOC 💺Green Soul Chair: https://amzn.to/3mWIxZP ✨ Tags ✨ ishan sharma,ai agents,ai agents explained,ai agents 2025,ai assistant,get rich with ai agents,make money ai agents,make money online,earn money online,ai agents tutorial,ai agent,ai,ai tools,make money with ai,make money with ai tools,artificial intelligence,deepseek r1,ai agents use cases,ai agents projects,business ideas with low investment,zero investment business ideas,business ideas for students,business ideas for beginners,best business ideas,how to start a business,online business ideas,startup business ideas,ai business ideas,business ideas using ai ✨ Hashtags ✨ #ai #artificialintelligence #businessideas

I ranked every AI Coder: Bolt vs. Cursor vs. Replit vs Lovable
youtube
LLM Vibe Score0.399
Human Vibe Score0.77
Greg IsenbergJan 24, 2025

I ranked every AI Coder: Bolt vs. Cursor vs. Replit vs Lovable

v0 vs windsurf vs replit vs bolt vs lovable vs tempolabs - which one should you use? Ras Mic breaks down the AI coding platforms based on how tech-savvy you are and how much control you want. He splits the tools into three groups: no-code options for non-techies, hybrid platforms for those with a mix of skills, and advanced tools for developers. None of them are quite ready for full-on production yet, but the video highlights what each one does best—whether it’s integrations, teamwork, or deployment features. Timestamps: 00:00 - Intro 01:00 - Overview of Popular Tools 02:29 - Technical vs. non-technical user classification 05:37 - Production readiness discussion 09:50 - Mapping Tools to User Profiles 12:52 - Platform comparisons and strengths 15:15 - Pricing discussion 16:43 - AI agents in coding platforms 19:04 - Final Recommendations and User Alignment Key Points: • Comprehensive comparison of major AI coding platforms (Lovable, Bolt, V0, Replit, Tempo Labs, Onlook, Cursor, Windsurf) • Tools categorized by technical expertise required and level of control offered • None of the platforms are 100% production-ready, but Replit and Tempo Labs are closest • All platforms offer similar base pricing ($20-30/month) with generous free tiers 1) First, understand the 3 MAJOR CATEGORIES of AI coding tools: • No-code (non-technical friendly) • Middle-ground (hybrid) • Technical (developer-focused) Your choice depends on TWO key factors: How much control you want Your technical expertise 2) THE CONTROL SPECTRUM Less Control → More Control • Lovable (basic control) • Bolt/V0 (code tweaking) • Replit (file management) • Tempo/Onlook (design control) • Cursor/Windsurf (full code control) 3) PRODUCTION READINESS STATUS Most honest take: None are 100% there yet, but some are close: Top contenders: • Replit • Tempo Labs Runner-ups: • Bolt • Lovable Pro tip: Start building now to be ready when they mature! 4) BEST TOOLS BY USER TYPE Non-technical: • Lovable • Bolt Product-minded non-technical: • Tempo Labs • Replit Technical folks: • Cursor • Windsurf 5) WINNING FEATURES BY PLATFORM Integrations: Lovable (crushing it!) Replit Tempo Labs Collaboration: Tempo Labs Replit Deployment: All solid, but Tempo needs work 6) PRICING INSIDER TIP All platforms hover around $20-30/month for basic tiers SECRET: They ALL have generous free tiers! Pro tip: Test drive everything before committing to paid plans 7) FINAL ADVICE Build a simple todo app on each platform Use free tiers to test Choose based on: Your technical comfort Desired level of control Specific project needs Remember: There's no "perfect" tool - just the right one for YOU! Notable Quotes: "None of the tools are there yet. I cannot confidently say you can build something to production easily, simply without a ton of roadblocks." - Ras Mic "Control is not for everybody. Did you like the assumptions that AI product was making for you? Or do you want to be able to tell it exactly what to do?" - Ras Mic LCA helps Fortune 500s and fast-growing startups build their future - from Warner Music to Fortnite to Dropbox. We turn 'what if' into reality with AI, apps, and next-gen products https://latecheckout.agency/ BoringAds — ads agency that will build you profitable ad campaigns http://boringads.com/ BoringMarketing — SEO agency and tools to get your organic customers http://boringmarketing.com/ Startup Empire - a membership for builders who want to build cash-flowing businesses https://www.startupempire.co FIND ME ON SOCIAL X/Twitter: https://twitter.com/gregisenberg Instagram: https://instagram.com/gregisenberg/ LinkedIn: https://www.linkedin.com/in/gisenberg/ FIND MIC ON SOCIAL X/Twitter: https://x.com/rasmickyy Youtube: https://www.youtube.com/@rasmic

YT_Emerging_Technologies_Introduction_to_AI
github
LLM Vibe Score0.461
Human Vibe Score0.039054583141409485
zusmaniJan 17, 2025

YT_Emerging_Technologies_Introduction_to_AI

YouTube Channel: Emerging Technologies Playlist: Introduction to AI Instructor: Zeeshan-ul-hassan Usmani Dear Students, I have uploaded all relevant material here for your quick access and learning. I hope you will find it beneficiary Yours Truly, Zeeshan =========================================== Video title: Resources Books to Order: Artificial Intelligence by Zeeshan Usmani - https://gufhtugu.com/artificial-intelligence Artificial Intelligence by Baqir Naqvi - https://gufhtugu.com/masnoi-zahanat/ Recommended Books • Gödel, Escher, Bach : An Eternal Golden Braid by Douglas R. Hofstadter A classic, poetic, philosophical defense of AI. • Machines Who Think by Pamela McCorduck. A good review of early AI history. • Robot: Mere Machine to Transcendent Mind by Hans P. Moravec Somewhat hyped book by a CMU robotics researcher. • Flesh and Machines: How Robots Will Change Us by Rodney Allen Brooks Reasonably decent book by MIT's leading robotics researcher. • Wired for War by Peter Warren Singer Reviews growing use of robots and unmanned vehicles in warfare. • Behind Deep Blue: Building the Computer That Defeated the World Chess Champion by Feng-Hsiung Hsu Autobiographical book on the development of a history making game-playing system. Interesting personal story of the hard engineering work that went into the system, with a few interesting facts on the technical aspects. • The Age of Spiritual Machines : When Computers Exceed Human Intelligence by Ray Kurzweil A recent view by an AI entrepreneur that has content if you ignore all the hype and overly-optimistic trust that Moore's law will magically solve all of the major problems. • Hal's Legacy : 2001's Computer As Dream and Reality An interesting collection of edited articles written to celebrate the fictional birthday of a famous intelligent computer who's true birthday must unfortunately be delayed, pending AI's inevitable progress. • The Sciences of the Artificial by Herbert Simon AI as science by one of its founders. • Models of My Life by Herbert Simon. An autobiography of one of AI's founders who's intellectual contributions also include fundamental contributions to economics (for which he won the Nobel prize), cognitive psychology, and computer science (such as co-inventing the linked list in the 1950's). • Alan Turing: The Enigma by Alan Hodges. A biography of one of the founders of CS and originator of the Turing test. Also a testimony to the tragic implications of homophobia. • The Emperor's New Mind : Concerning Computers, Minds, and the Laws of Physics and Shadows of the Mind : A Search for the Missing Science of Consciousness and The Large, the Small and the Human Mind by Roger Penrose A completely bogus argument against AI by a hopelessly Platonic mathematician. The last book contains an appended article by Stephen Hawking (a colleague of Penrose's) who of course doesn't buy his bogus argument. • The Mind's New Science : A History of the Cognitive Revolution by Howard Gardner A nice history of the development of cognitive science. • How the Mind Works , The Language Instinct , and Words and Rules : The Ingredients of Language by Steven Pinker Fun reading on lots of interesting issues in modern Cognitive Science and Linguistics if you don't take his exaggerated beliefs in nativism and evolutionary psychology too seriously. • Bots : The Origin of New Species by Andrew Leonard A light, somewhat hyped book on on Internet agents, chatterbots, etc. with a few funny stories. • Mathematics: The Loss of Certainty by Morris Kline A very nice book on the failed enterprise of using logic to build a firm foundation for infallible mathematics and the role of Gödel's Incompleteness Theorem in the philosophy of mathematics. • Incompleteness: The Proof and Paradox of Kurt Gödel by Rebecca Goldstein An interesting biography of Kurt Gödel. Too bad he was such a Platonist that, unlike Turing, he did not understand the true implications of his own theorems (interesting author connection: Goldstein is Pinker's wife). Links: • AAAI AI Topics Basic info on AI from the American Association for Artificial Intelligence: http://www.aaai.org/AITopics/html/welcome.html • Loebner Prize for limited Turing test: http://www.loebner.net/Prizef/loebner-prize.html • IBM's Deep Blue Page: http://www.research.ibm.com/deepblue/ • Robocup: Robotic Soccer Competition: http://www.robocup.org/ • NY Times Article on Proof of the Robbins Theorem: http://www.nytimes.com/library/cyber/week/1210math.html • NY Times article on Bayes Nets at Microsoft Research: http://www.nytimes.com/library/tech/00/07/biztech/articles/17lab.html =========================================== Video title: Numbers Infinity Video Link - •https://www.youtube.com/watch?v=hlXHwMgS06c https://www.cbs.com/shows/numb3rs/ http://numb3rs.wolfram.com/ =========================================== Video title: 20 Hours Rule and Assisgnemnt Assignment - https://www.urdufake2020.cicling.org/ =========================================== Video title: Assignments – P1 Mostly Human - https://money.cnn.com/mostly-human =========================================== Video title: Assignments – P2 Assignment – 2 - https://replika.ai/ Assignment – 3 – Teachable Machines https://teachablemachine.withgoogle.com/ Assignment – 4 – Tensor Flow Playground https://playground.tensorflow.org Assignment – 5 – GPT-3 Paper (175B Parameters) https://debuild.co/ Assignment – 6 - Image GPT-3 https://openai.com/blog/image-gpt/ =========================================== Video title: Create your own Deep Fake 1.https://colab.research.google.com/drive/1mGg_fmvhTpvkPkclw2yKkhALVzmawfvT?usp=sharing 2.https://drive.google.com/drive/folders/1wW1bxRV2S7Ce8gc3VDTzMQABE3-WCc_Y?usp=sharing •go into you gdrive > find cloned folder and ensure that this folder must have: vox-adv-cpk.pth.tar & vox-cpk.pth.tar failes •Aliaksandr Siarohin : https://github.com/AliaksandrSiarohin/first-order-model

DO THIS To Get RICH With AI in 2025
youtube
LLM Vibe Score0.358
Human Vibe Score0.31
Ishan SharmaJan 12, 2025

DO THIS To Get RICH With AI in 2025

Ishan Sharma: DO THIS To Get RICH With AI in 2025 How AI is CHANGING the Startup World! 🤯 Sam Altman, CEO of Open AI, predicts how one person could build a billion dollar startup, only using AI tools and software. It is crazy to think that the next billion dollar company might just be yours or mine with our AI toolset. This is a glimpse from the podcast where me and Saheli discussed freelancing, how to master personal branding as a beginner, how to talk with clients and much more. 📸 Instagram: https://bit.ly/ishansharma7390ig Join MarkitUpX Discord Server: https://discord.gg/fwSpTje4rh 😁 About Me: https://bit.ly/aboutishansharma 📱 Twitter: https://bit.ly/ishansharma7390twt 📝 LinkedIn: https://bit.ly/ishansharma7390li 🌟 Please leave a LIKE ❤️ and SUBSCRIBE for more AMAZING content! 🌟 3 Books You Should Read 📈Psychology of Money: https://amzn.to/30wx4bW 👀Subtle Art of Not Giving a F: https://amzn.to/30zwWbP 💼Rework: https://amzn.to/3ALsAuz Tech I use every day 💻MacBook Air M1: https://amzn.to/2YWKPjG 📺LG 29' Ultrawide Monitor: https://amzn.to/3aG0p5p 🎥Sony ZV1: https://amzn.to/3ANqgDb 🎙Blue Yeti Mic: https://amzn.to/2YYbiNN ⽴Tripod Stand: https://amzn.to/3mVUiQc 🔅Ring Light: https://amzn.to/2YQlzLJ 🎧Marshall Major II Headphone: https://amzn.to/3lLhTDQ 🖱Logitech mouse: https://amzn.to/3p8edOC 💺Green Soul Chair: https://amzn.to/3mWIxZP ✨ Tags ✨ ishan sharma,DO THIS To Get RICH With AI in 2025,ai agent,ai agents,low investment business ideas,business ideas with low investment,zero investment business ideas,best business ideas 2024,business ideas for students,business ideas for beginners,best business ideas,how to start a business,online business ideas,new business ideas 2024,startup business ideas,money,ai business ideas,business ideas using ai,ai,artificial intelligence,chatgpt,bard,gemini,google ✨ Hashtags ✨ #business #businessideas #ai

teach-AI-in-business
github
LLM Vibe Score0.443
Human Vibe Score0.018525334165293606
aenyneJan 9, 2025

teach-AI-in-business

Teaching AI in Business ![HitCount] I am collecting material for teaching AI-related issues to non-tech people. The links should provide for a general understanding of AI without going too deep into technical issues. Please contribute! Make this Issue your First Issue I am collecting material for teaching AI-related issues to non-tech people. The links should have provide for a general understanding of AI without going too deep into technical issues. Please contribute! Kindly use only those Resources with NO CODE NEW Check out also the AI Wiki NEW Online Videos & Courses | Link to Issue | Description | |---|---| | Top Trending Technologies | Youtube Channel to master top trending technologyies including artificial intelligence | | AI4All | AI 4 All is a resource for AI facilitators to bring AI to scholars and students | | Elements of AI | Elements of AI is a free open online course to teach AI principles | | Visual Introduction to Machine Learning | Visual introduction to Machine Learning is a beautiful website that gives a comprehensive introduction and easily understood first encounter with machine learning | | CS50's Introduction to Artificial Intelligence with Python | Learn to use machine learning in Python in this introductory course on artificial intelligence.| | Crash course for AI | This is a fun video series that introduces students and educators to Artificial Intelligence and also offers additional more advanced videos. Learn about the basics, neural networks, algorithms, and more. | Youtuber Channel Machine Learning Tutorial | Youtube Channel Turorial Teachable Machine for beginner | | Artificial Intelligence (AI) |Learn the fundamentals of Artificial Intelligence (AI), and apply them. Design intelligent agents to solve real-world problems including, search, games, machine learning, logic, and constraint satisfaction problems | | AI For Everyone by Andrew Ng | AI For Everyone is a course especially for people from a non-technical background to understand AI strategies | | How far is too far? The age of AI| This is a Youtube Orignals series by Robert Downey| | Fundamentals of Artificial Intelligence|This course is for absolute beginners with no technical knowledge.| | Bandit Algorithm (Online Machine Learning)|No requirement of technical knowledge, but a basic understending of Probability Ttheory would help| | An Executive's Guide to AI|This is an interactive guide to teaching business professionals how they might employ artificial intelligence in their business| | AI Business School|Series of videos that teach how AI may be incorporated in various business industries| | Artificial Intelligence Tutorial for Beginners | This video will provide you with a comprehensive and detailed knowledge of Artificial Intelligence concepts with hands-on examples. | | Indonesian Machine Learning Tutorial | Turorial Teachable Machine to train a computer for beginner | | Indonesian Youtube Playlist AI Tutorial | Youtube Playlist AI Tutorial For Beginner | | Artificial Intelligence Search Methods For Problem Solving By Prof. Deepak Khemani|These video lectures are for absolute beginners with no technical knowledge| | AI Basics Tutorial | This video starts from the very basics of AI and ML, and finally has a hands-on demo of the standard MNIST Dataset Number Detection model using Keras and Tensorflow.| | Simple brain.js Tutorial | This video explains a very simple javascript AI library called brain.js so you can easily run AI in the browser.| | Google AI| A complete kit for by google official for non-tech guy to start all over from basics, till advanced | | Microsoft AI for Beginners| A self-driven curriculum by Microsoft, which includes 24 lessons on AI. | Train Your Own AI | Link to Issue | Description | |---|---| | Teachable Machine | Use Teachable Machine to train a computer to recognize your own images, sounds, & poses | | eCraft2Learn | Resource and interactive space (Snap, a visual programming environment like Scratch) to learn how to create AI programs | | Google Quick Draw | Train an AI to guess from drawings| | Deepdream Generator| Merge Pictures to Deep Dreams using the Deepdream Generator| | Create ML|Quickly build and train Core ML models on your Mac with no code.| | What-If Tool|Visually probe the behavior of trained machine learning models, with minimal coding.| | Metaranx|Use and build artificial intelligence tools to analyze and make decisions about your data. Drag-and-drop. No code.| | obviously.ai|The total process of building ML algorithms, explaining results, and predicting outcomes in one single click.| Articles | By & Title | Description | |---|---| | Artificial Intelligence | Wikipedia Page of AI | | The Non-Technical AI Guide | One of the good blog post that could help AI more understandable for people without technical background | | LIAI | A detailed introduction to AI and neural networks | | Layman's Intro | A layman's introduction to AI | | AI and Machine Learning: A Nontechnical Overview | AI and Machine Learning: A Nontechnical Overview from OREILLY themselves is a guide to learn anyone everything they need to know about AI, focussed on non-tech people | | What business leaders need to know about artifical intelligence|Short article that summarizes the essential aspects of AI that business leaders need to understand| | How Will No-Code Impact the Future of Conversational AI | A humble explanation to the current state of converstational AI i.e.Chatbots and how it coul evolve with the current trend of no coding. | | Investopedia | Basic explanation of what AI is in a very basic and comprehensive way | | Packtpub | A non programmer’s guide to learning Machine learning | | Builtin | Artificial Intelligence.What is Artificial Intelligence? How Does AI Work? | | Future Of Life | Benefits & Risks of Artificial Intelligence | | NSDM India -Arpit | 100+ AI Tools For Non-Coders That Will Make Your Marketing Better. | | AI in Marketing for Startups & Non-technical Marketers | A practical guide for non-technical people | | Blog - Machine Learning MAstery | Blogs and Articles by Jason Browniee on ML | | AI Chatbots without programming| Chatbots are increasingly in demand among global businesses. This course will teach you how to build, analyze, deploy and monetize chatbots - with the help of IBM Watson and the power of AI.| Book Resources for Further Reading | Author | Book | Description & Notes | |---|---|---| | Ethem Alpaydin|Machine Learning: The New AI | Graph Theory with Applications to Engineering & Computer Science. A concise overview of machine learning—computer programs that learn from data—which underlies applications that include recommendation systems, face recognition, and driverless cars. | | Charu C. Aggarwal| Neural Networks and Deep Learning | This book covers both classical and modern models in deep learning. The primary focus is on the theory and algorithms of deep learning. The book is also rich in discussing different applications in order to give the practitioner a flavor of how neural architectures are designed for different types of problems. | | Hal Daumé III | A Course in Machine Learning | The purpose of this book is to provide a gentle and pedagogically organized introduction to the field. A second goal of this book is to provide a view of machine learning that focuses on ideas and models, not on math. | | Ian Goodfellow and Yoshua Bengio and Aaron Courville| Deep Learning | The book starts with a discussion on machine learning basics, including the applied mathematics and algorithms needed to effectively study deep learning from an academic perspective. There is no code covered in the book, making it perfect for a non-technical AI enthusiast. | | Peter Harrington|Machine Learning in Action| (Source: https://github.com/kerasking/book-1/blob/master/ML%20Machine%20Learning%20in%20Action.pdf) This book acts as a guide to walk newcomers through the techniques needed for machine learning as well as the concepts behind the practices.| | Jeff Heaton| Artificial Intelligence for Humans |This book helps its readers get an overview and understanding of AI algorithms. It is meant to teach AI for those who don’t have an extensive mathematical background. The readers need to have only a basic knowledge of computer programming and college algebra.| | John D. Kelleher, Brian Mac Namee and Aoife D'Arcy|Fundamentals of Machine Learning for Predictive Data Analytics: Algorithms, Worked Examples, and Case Studies (The MIT Press)|This book covers all the fundamentals of machine learning, diving into the theory of the subject and using practical applications, working examples, and case studies to drive the knowledge home.| | Deepak Khemani| [A First Course in Artificial Intelligence] | It is an introductory course on Artificial Intelligence, a knowledge-based approach using agents all across and detailed, well-structured algorithms with proofs. This book mainly follows a bottom-up approach exploring the basic strategies needed problem-solving on the intelligence part. | | Maxim Lapan | Deep Reinforcement Learning Hands-On - Second Edition | Deep Reinforcement Learning Hands-On, Second Edition is an updated and expanded version of the bestselling guide to the very latest reinforcement learning (RL) tools and techniques. It provides you with an introduction to the fundamentals of RL, along with the hands-on ability to code intelligent learning agents to perform a range of practical tasks. | | Tom M Mitchell | Machine Learning | This book covers the field of machine learning, which is the study of algorithms that allow computer programs to automatically improve through experience. The book is intended to support upper level undergraduate and introductory level graduate courses in machine learning. | | John Paul Mueller and Luca Massaron|Machine Learning For Dummies|This book aims to get readers familiar with the basic concepts and theories of machine learning and how it applies to the real world. And "Dummies" here refers to absolute beginners with no technical background.The book introduces a little coding in Python and R used to teach machines to find patterns and analyze results. From those small tasks and patterns, we can extrapolate how machine learning is useful in daily lives through web searches, internet ads, email filters, fraud detection, and so on. With this book, you can take a small step into the realm of machine learning and we can learn some basic coding in Pyton and R (if interested)| | Michael Nielsen| Neural Networks and Deep Learning |Introduction to the core principles of Neural Networks and Deep Learning in AI| | Simon Rogers and Mark Girolami| A Course in Machine Learning |A First Course in Machine Learning by Simon Rogers and Mark Girolami is the best introductory book for ML currently available. It combines rigor and precision with accessibility, starts from a detailed explanation of the basic foundations of Bayesian analysis in the simplest of settings, and goes all the way to the frontiers of the subject such as infinite mixture models, GPs, and MCMC.| |Peter Norvig| Paradigm of Artificial Intelligence Programming |Paradigms of AI Programming is the first text to teach advanced Common Lisp techniques in the context of building major AI systems. By reconstructing authentic, complex AI programs using state-of-the-art Common Lisp, the book teaches students and professionals how to build and debug robust practical programs, while demonstrating superior programming style and important AI concepts.| | Stuart Russel & Peter Norvig | Artificial Intelligence: A Modern Approach, 3rd Edition | This is the prescribed text book for my Introduction to AI university course. It starts off explaining all the basics and definitions of what AI is, before launching into agents, algorithms, and how to apply them. Russel is from the University of California at Berkeley. Norvig is from Google.| | Richard S. Sutton and Andrew G. Barto| Reinforcement Learning: An Introduction |Reinforcement learning, one of the most active research areas in artificial intelligence, is a computational approach to learning whereby an agent tries to maximize the total amount of reward it receives while interacting with a complex, uncertain environment.| | Alex Smola and S.V.N. Vishwanathan | Introduction to Machine Learning | Provides the reader with an overview of the vast applications of ML, including some basic tools of statistics and probability theory. Also includes discussions on sophisticated ideas and concepts. | | Shai Shalev-Shwartz and Shai Ben-David | Understanding Machine Learning From Theory to Algorithms |The primary goal of this book is to provide a rigorous, yet easy to follow, introduction to the main concepts underlying machine learning. | | Chandra S.S.V | Artificial Intelligence and Machine Learning | This book is primarily intended for undergraduate and postgraduate students of computer science and engineering. This textbook covers the gap between the difficult contexts of Artificial Intelligence and Machine Learning. It provides the most number of case studies and worked-out examples. In addition to Artificial Intelligence and Machine Learning, it also covers various types of learning like reinforced, supervised, unsupervised and statistical learning. It features well-explained algorithms and pseudo-codes for each topic which makes this book very useful for students. | | Oliver Theobald|Machine Learning For Absolute Beginners: A Plain English Introduction|This is an absolute beginners ML guide.No mathematical background is needed, nor coding experience — this is the most basic introduction to the topic for anyone interested in machine learning.“Plain” language is highly valued here to prevent beginners from being overwhelmed by technical jargon. Clear, accessible explanations and visual examples accompany the various algorithms to make sure things are easy to follow.| | Tom Taulli | Artificial Intelligence Basics: A Non-Technical Introduction | This book equips you with a fundamental grasp of Artificial Intelligence and its impact. It provides a non-technical introduction to important concepts such as Machine Learning, Deep Learning, Natural Language Processing, Robotics and more. Further the author expands on the questions surrounding the future impact of AI on aspects that include societal trends, ethics, governments, company structures and daily life. | |Cornelius Weber, Mark Elshaw, N. Michael Mayer| Reinforcement Learning |Learning is a very important aspect. This book is on reinforcement learning which involves performing actions to achieve a goal. The first 11 chapters of this book describe and extend the scope of reinforcement learning.| |John D. Kelleher, Brian Mac Namee, Aoife D'arcy| Algorithms, Worked Examples, and Case Studies | A comprehensive introduction to the most important machine learning approaches used in predictive data analytics, covering both theoretical concepts and practical applications. |

FORGET ChatGPT, This AI TOOL is a GAMECHANGER 🔥
youtube
LLM Vibe Score0.299
Human Vibe Score0.32
Ishan SharmaOct 19, 2024

FORGET ChatGPT, This AI TOOL is a GAMECHANGER 🔥

Ishan Sharma: FORGET ChatGPT, This AI TOOL is a GAMECHANGER 🔥 Google just dropped NotebookLM and it is changing everything. I was using ChatGPT so far for research and learning, but NotebookLM has stolen the charm. NotebookLM lets you convert PDFs, YouTube videos, or Websites into Audio Podcasts This is a 2 person conversational podcast about the topic And Trust me, it sounds too good to be AI generated. It's powered by Google's latest Gemini 1.5 model. But that’s not all! You can add multiple sources in a notebook And also get a summary, table of contents, study guide And MORE in seconds! THIS is a game changer for learners, researchers, and creators! Helping you skyrocket your productivity. It's also great for school students and college students to learn anything faster. Try it out on notebooklm.google.com 📸 Instagram: https://bit.ly/ishansharma7390ig Join MarkitUpX Discord Server: https://discord.gg/fwSpTje4rh Timestamps 😁 About Me: https://bit.ly/aboutishansharma 📱 Twitter: https://bit.ly/ishansharma7390twt 📝 LinkedIn: https://bit.ly/ishansharma7390li 🌟 Please leave a LIKE ❤️ and SUBSCRIBE for more AMAZING content! 🌟 3 Books You Should Read 📈Psychology of Money: https://amzn.to/30wx4bW 👀Subtle Art of Not Giving a F: https://amzn.to/30zwWbP 💼Rework: https://amzn.to/3ALsAuz Tech I use every day 💻MacBook Air M1: https://amzn.to/2YWKPjG 📺LG 29' Ultrawide Monitor: https://amzn.to/3aG0p5p 🎥Sony ZV1: https://amzn.to/3ANqgDb 🎙Blue Yeti Mic: https://amzn.to/2YYbiNN ⽴Tripod Stand: https://amzn.to/3mVUiQc 🔅Ring Light: https://amzn.to/2YQlzLJ 🎧Marshall Major II Headphone: https://amzn.to/3lLhTDQ 🖱Logitech mouse: https://amzn.to/3p8edOC 💺Green Soul Chair: https://amzn.to/3mWIxZP ✨ Tags ✨ ishan sharma,FORGET ChatGPT This AI TOOL is a GAMECHANGER,chatgpt,gpt-4o,chatgpt 4o,gpt4o,openai,gpt 4,openai sora,microsoft openai,artificial intelligence,ai,chatbot,gpt-4,chatgpt-4,new gpt ai model,chatgpt vision,chatgpt chatbot,chatgpt4o,new ai,chat gpt,chatgpt 4,gpt update,chat gpt 4o,google notebooklm,google notebook,google notebook app,google notebooklm tool,google keep,google ai,google app ai,google notes app,google notes ai ✨ Hashtags ✨ #google #chatgpt #ai

LearnAI-KnowledgeMiningBootcamp
github
LLM Vibe Score0.438
Human Vibe Score0.05521136990708693
sithukyaw007Jan 29, 2024

LearnAI-KnowledgeMiningBootcamp

LearnAI: Build an Enterprise Knowledge Mining Solution using the Microsoft AI Platform Build an enterprise scale intelligent search solution for searching business documents using Microsoft Azure and Cognitive Search About this Course In this course, you will learn to build an enterprise search solution by applying knowledge mining approach to search an organization’s business documents like Microsoft Office, PDFs and images using Azure search and Cognitive search skillsets and expose the results via a Bot interface. You will learn to perform entity recognition, image analysis, text translation and indexed search on enterprise business documents using Microsoft Cognitive Services and Azure Search. This approach can be used with almost any Azure service to augment a customer’s scenario involving intelligent search. While this course focusses on Azure and Cognitive search capabilities, a depth course on building Bots and integrating various cognitive services is available here - Building Intelligent Agents and Apps. In this course you will learn Fundamentals of Azure Search and its capabilities. Understand Microsoft Cognitive Search and its key scenarios for using them. Build an enriched data pipeline for search using predefined and custom skillsets: a. Text skills like entity recognition, language detection, text manipulation and key phrase extraction. b. Image skills like OCR. c. Language skills like text translation. d. Content moderation skills to block documents with incompliant content. Use the enriched data pipeline for a knowledge mining solution on business documents within an enterprise. Expose the knowledge mining solution using a bot interface for document search and consumption. Architecture !Architecture Technologies Covered !Technology Industry application Intelligent search is relevant to many major industries. Some are listed below. Retail and health care industries employ chatbots with advanced multi-language support capabilities to service their customers. Retail, Housing and Automotive industries for sales/listing. Entertainment industry uses search for relevant/contextual on-demand streaming. Pre-requisites Fundamental working knowledge of Azure Portal, Functions and Azure Search. Familiarity with Visual Studio. Familiarity with Azure Bots and Microsoft Bot Framework v4. If you do not have any familiarity with the above pre-requisites, please find below links To Read (10 minutes): Visual Studio Tutorial To Read (4 minutes): Azure Functions Overview To Read (10 minutes): Azure Search Overview To Read (7 minutes): Postman Tutorial To Do (30 minutes): CQuickstart Pre-Setup before you attend the class Mandatory To Create: You need a Microsoft Azure account to create the services we use in our solution. You can create a free account, use your MSDN account or use any other subscription where you have permission to create services. To Install: Visual Studio 2017 version version 15.5 or later, including the Azure development workload. To Install: Postman. To call the labs APIs. Course Details Primary Audience: Azure AI Developers, Architects. Secondary Audience: Any professional interested in learning AI. Level This content is designed as an intermediate to advanced level course for AI developers and/or architects. Type This course, in its full form, is designed to be taught in-person but you can also use the materials in a self-paced fashion. There are assignments and multiple reference links throughout the materials that support the concepts and skills you will learn. Length Full Course classroom training: 16 hours Related LearnAI Courses Building Intelligent Agents and Apps Course Modules Introduction – Overview of Azure Search, Cognitive Search, Scenarios and industry specific applications. Fundamentals of Azure Search. Architecture – Solution Architecture for building enterprise search solution. Cognitive Search Skillset – Applying text skills. Cognitive Search Skillset – Applying image skills. Cognitive Search Skillset – Applying Language skills. Cognitive Search Skillset – Applying Moderation skills. Build and Integrate a Bot with Cognitive Search API. Group Hands-on Lab to practice skills acquired.

Learn AI in Just 3 HOURS 🚀| ChatGPT & Generative AI | Ishan Sharma #shorts
youtube
LLM Vibe Score0.318
Human Vibe Score0.31
Ishan SharmaNov 3, 2023

Learn AI in Just 3 HOURS 🚀| ChatGPT & Generative AI | Ishan Sharma #shorts

BEST FREE AI Course For EVERYONE 🚀| Ishan Sharma 📸 Instagram: https://bit.ly/ishansharma7390ig Join MarkitUpX Discord Server: https://discord.gg/fwSpTje4rh 😁 About Me: https://bit.ly/aboutishansharma 📱 Twitter: https://bit.ly/ishansharma7390twt 📝 LinkedIn: https://bit.ly/ishansharma7390li 🌟 Please leave a LIKE ❤️ and SUBSCRIBE for more AMAZING content! 🌟 3 Books You Should Read 📈Psychology of Money: https://amzn.to/30wx4bW 👀Subtle Art of Not Giving a F: https://amzn.to/30zwWbP 💼Rework: https://amzn.to/3ALsAuz Tech I use every day 💻MacBook Air M1: https://amzn.to/2YWKPjG 📺LG 29' Ultrawide Monitor: https://amzn.to/3aG0p5p 🎥Sony ZV1: https://amzn.to/3ANqgDb 🎙Blue Yeti Mic: https://amzn.to/2YYbiNN ⽴Tripod Stand: https://amzn.to/3mVUiQc 🔅Ring Light: https://amzn.to/2YQlzLJ 🎧Marshall Major II Headphone: https://amzn.to/3lLhTDQ 🖱Logitech mouse: https://amzn.to/3p8edOC 💺Green Soul Chair: https://amzn.to/3mWIxZP ✨ Tags ✨ ishan sharma,artificial intelligence,Artificial Intelligence Tutorial for Beginners,artificial intelligence course for beginners,what is artificial intelligence,artificial intelligence for beginners,ai developer,ai course,coding,programming,machine learning,data science,developer,development,coding courses,learn to code,ai for beginners,chatgpt,google bard,free google course,free courses,ai engineer,aiml,best,ai,ai courses,BEST FREE AI Course For EVERYONE ✨ Hashtags ✨ #ai #artificialintelligence #course

AI Will Help you EARN CRORES! CRAZY BUSINESS IDEA! 🤑 | Ishan Sharma #shorts
youtube
LLM Vibe Score0.393
Human Vibe Score0.62
Ishan SharmaOct 14, 2023

AI Will Help you EARN CRORES! CRAZY BUSINESS IDEA! 🤑 | Ishan Sharma #shorts

📸 Instagram: https://bit.ly/ishansharma7390ig Join MarkitUpX Discord Server: https://discord.gg/fwSpTje4rh 😁 About Me: https://bit.ly/aboutishansharma 📱 Twitter: https://bit.ly/ishansharma7390twt 📝 LinkedIn: https://bit.ly/ishansharma7390li 🌟 Please leave a LIKE ❤️ and SUBSCRIBE for more AMAZING content! 🌟 3 Books You Should Read 📈Psychology of Money: https://amzn.to/30wx4bW 👀Subtle Art of Not Giving a F: https://amzn.to/30zwWbP 💼Rework: https://amzn.to/3ALsAuz Tech I use every day 💻MacBook Air M1: https://amzn.to/2YWKPjG 📺LG 29' Ultrawide Monitor: https://amzn.to/3aG0p5p 🎥Sony ZV1: https://amzn.to/3ANqgDb 🎙Blue Yeti Mic: https://amzn.to/2YYbiNN ⽴Tripod Stand: https://amzn.to/3mVUiQc 🔅Ring Light: https://amzn.to/2YQlzLJ 🎧Marshall Major II Headphone: https://amzn.to/3lLhTDQ 🖱Logitech mouse: https://amzn.to/3p8edOC 💺Green Soul Chair: https://amzn.to/3mWIxZP ✨ Tags ✨ ishan sharma,THIS Will Help You EARN CRORES!,make money online,make money online 2023,ai tools for photo editing,ai tools for photos,earn money online,ishan sharma motivation,ishan sharma speech,ai tools,ai tools to earn money,money making apps,ai apps to make money,photo generator ai,ai photo generator,ai photo generator free,using ai tools,how to earn money using ai,best app store apps,app store,ishan sharma freelance,how to sell anything ✨ Hashtags ✨ #makemoneyonline #ai #growth

OpenAI GPT-4: THE SECRET PROMPT You Need To Know 🤐 #shorts
youtube
LLM Vibe Score0.326
Human Vibe Score0.36
Ishan SharmaApr 2, 2023

OpenAI GPT-4: THE SECRET PROMPT You Need To Know 🤐 #shorts

OpenAI GPT-4: THE SECRET PROMPT You Need To Know 🤐 #shorts 📸 Instagram: https://bit.ly/ishansharma7390ig Join MarkitUpX Discord Server: https://discord.gg/fwSpTje4rh 😁 About Me: https://bit.ly/aboutishansharma 📱 Twitter: https://bit.ly/ishansharma7390twt 📝 LinkedIn: https://bit.ly/ishansharma7390li 🌟 Please leave a LIKE ❤️ and SUBSCRIBE for more AMAZING content! 🌟 3 Books You Should Read 📈Psychology of Money: https://amzn.to/30wx4bW 👀Subtle Art of Not Giving a F: https://amzn.to/30zwWbP 💼Rework: https://amzn.to/3ALsAuz Tech I use every day 💻MacBook Air M1: https://amzn.to/2YWKPjG 📺LG 29' Ultrawide Monitor: https://amzn.to/3aG0p5p 🎥Sony ZV1: https://amzn.to/3ANqgDb 🎙Blue Yeti Mic: https://amzn.to/2YYbiNN ⽴Tripod Stand: https://amzn.to/3mVUiQc 🔅Ring Light: https://amzn.to/2YQlzLJ 🎧Marshall Major II Headphone: https://amzn.to/3lLhTDQ 🖱Logitech mouse: https://amzn.to/3p8edOC 💺Green Soul Chair: https://amzn.to/3mWIxZP ✨ Tags ✨ ishan sharma,OpenAI GPT-4: THE SECRET PROMPT You Need To Know 🤐,ai tools,artificial intelligence,microsoft,microsoft openai,microsoft chatgpt,what is chatgpt,chatgpt tutorial,chatgpt 4,make money with chatgpt,chatgpt,open ai,gpt 4,gpt 4 vs gpt 3,openai,openai chatbot gpt,chatgpt explained,chatgpt examples,chat gpt,open ai gpt 4,chat gpt uses,chatgpt4 tutorial,gpt 4 demo,gpt 4 image input,gpt 4 review,gpt 4 how to use,gpt 4 prompt ✨ Hashtags ✨ #gpt4 #chatgpt #ai