VibeBuilders.ai Logo
VibeBuilders.ai

Secret

Explore resources related to secret to help implement AI solutions for your business.

[N] Inside DeepMind's secret plot to break away from Google
reddit
LLM Vibe Score0
Human Vibe Score0
MassivePellfishThis week

[N] Inside DeepMind's secret plot to break away from Google

Article https://www.businessinsider.com/deepmind-secret-plot-break-away-from-google-project-watermelon-mario-2021-9 by Hugh Langley and Martin Coulter For a while, some DeepMind employees referred to it as "Watermelon." Later, executives called it "Mario." Both code names meant the same thing: a secret plan to break away from parent company Google. DeepMind feared Google might one day misuse its technology, and executives worked to distance the artificial-intelligence firm from its owner for years, said nine current and former employees who were directly familiar with the plans. This included plans to pursue an independent legal status that would distance the group's work from Google, said the people, who asked not to be identified discussing private matters. One core tension at DeepMind was that it sold the business to people it didn't trust, said one former employee. "Everything that happened since that point has been about them questioning that decision," the person added. Efforts to separate DeepMind from Google ended in April without a deal, The Wall Street Journal reported. The yearslong negotiations, along with recent shake-ups within Google's AI division, raise questions over whether the search giant can maintain control over a technology so crucial to its future. "DeepMind's close partnership with Google and Alphabet since the acquisition has been extraordinarily successful — with their support, we've delivered research breakthroughs that transformed the AI field and are now unlocking some of the biggest questions in science," a DeepMind spokesperson said in a statement. "Over the years, of course we've discussed and explored different structures within the Alphabet group to find the optimal way to support our long-term research mission. We could not be prouder to be delivering on this incredible mission, while continuing to have both operational autonomy and Alphabet's full support." When Google acquired DeepMind in 2014, the deal was seen as a win-win. Google got a leading AI research organization, and DeepMind, in London, won financial backing for its quest to build AI that can learn different tasks the way humans do, known as artificial general intelligence. But tensions soon emerged. Some employees described a cultural conflict between researchers who saw themselves firstly as academics and the sometimes bloated bureaucracy of Google's colossal business. Others said staff were immediately apprehensive about putting DeepMind's work under the control of a tech giant. For a while, some employees were encouraged to communicate using encrypted messaging apps over the fear of Google spying on their work. At one point, DeepMind's executives discovered that work published by Google's internal AI research group resembled some of DeepMind's codebase without citation, one person familiar with the situation said. "That pissed off Demis," the person added, referring to Demis Hassabis, DeepMind's CEO. "That was one reason DeepMind started to get more protective of their code." After Google restructured as Alphabet in 2015 to give riskier projects more freedom, DeepMind's leadership started to pursue a new status as a separate division under Alphabet, with its own profit and loss statement, The Information reported. DeepMind already enjoyed a high level of operational independence inside Alphabet, but the group wanted legal autonomy too. And it worried about the misuse of its technology, particularly if DeepMind were to ever achieve AGI. Internally, people started referring to the plan to gain more autonomy as "Watermelon," two former employees said. The project was later formally named "Mario" among DeepMind's leadership, these people said. "Their perspective is that their technology would be too powerful to be held by a private company, so it needs to be housed in some other legal entity detached from shareholder interest," one former employee who was close to the Alphabet negotiations said. "They framed it as 'this is better for society.'" In 2017, at a company retreat at the Macdonald Aviemore Resort in Scotland, DeepMind's leadership disclosed to employees its plan to separate from Google, two people who were present said. At the time, leadership said internally that the company planned to become a "global interest company," three people familiar with the matter said. The title, not an official legal status, was meant to reflect the worldwide ramifications DeepMind believed its technology would have. Later, in negotiations with Google, DeepMind pursued a status as a company limited by guarantee, a corporate structure without shareholders that is sometimes used by nonprofits. The agreement was that Alphabet would continue to bankroll the firm and would get an exclusive license to its technology, two people involved in the discussions said. There was a condition: Alphabet could not cross certain ethical redlines, such as using DeepMind technology for military weapons or surveillance. In 2019, DeepMind registered a new company called DeepMind Labs Limited, as well as a new holding company, filings with the UK's Companies House showed. This was done in anticipation of a separation from Google, two former employees involved in those registrations said. Negotiations with Google went through peaks and valleys over the years but gained new momentum in 2020, one person said. A senior team inside DeepMind started to hold meetings with outside lawyers and Google to hash out details of what this theoretical new formation might mean for the two companies' relationship, including specifics such as whether they would share a codebase, internal performance metrics, and software expenses, two people said. From the start, DeepMind was thinking about potential ethical dilemmas from its deal with Google. Before the 2014 acquisition closed, both companies signed an "Ethics and Safety Review Agreement" that would prevent Google from taking control of DeepMind's technology, The Economist reported in 2019. Part of the agreement included the creation of an ethics board that would supervise the research. Despite years of internal discussions about who should sit on this board, and vague promises to the press, this group "never existed, never convened, and never solved any ethics issues," one former employee close to those discussions said. A DeepMind spokesperson declined to comment. DeepMind did pursue a different idea: an independent review board to convene if it were to separate from Google, three people familiar with the plans said. The board would be made up of Google and DeepMind executives, as well as third parties. Former US president Barack Obama was someone DeepMind wanted to approach for this board, said one person who saw a shortlist of candidates. DeepMind also created an ethical charter that included bans on using its technology for military weapons or surveillance, as well as a rule that its technology should be used for ways that benefit society. In 2017, DeepMind started a unit focused on AI ethics research composed of employees and external research fellows. Its stated goal was to "pave the way for truly beneficial and responsible AI." A few months later, a controversial contract between Google and the Pentagon was disclosed, causing an internal uproar in which employees accused Google of getting into "the business of war." Google's Pentagon contract, known as Project Maven, "set alarm bells ringing" inside DeepMind, a former employee said. Afterward, Google published a set of principles to govern its work in AI, guidelines that were similar to the ethical charter that DeepMind had already set out internally, rankling some of DeepMind's senior leadership, two former employees said. In April, Hassabis told employees in an all-hands meeting that negotiations to separate from Google had ended. DeepMind would maintain its existing status inside Alphabet. DeepMind's future work would be overseen by Google's Advanced Technology Review Council, which includes two DeepMind executives, Google's AI chief Jeff Dean, and the legal SVP Kent Walker. But the group's yearslong battle to achieve more independence raises questions about its future within Google. Google's commitment to AI research has also come under question, after the company forced out two of its most senior AI ethics researchers. That led to an industry backlash and sowed doubt over whether it could allow truly independent research. Ali Alkhatib, a fellow at the Center for Applied Data Ethics, told Insider that more public accountability was "desperately needed" to regulate the pursuit of AI by large tech companies. For Google, its investment in DeepMind may be starting to pay off. Late last year, DeepMind announced a breakthrough to help scientists better understand the behavior of microscopic proteins, which has the potential to revolutionize drug discovery. As for DeepMind, Hassabis is holding on to the belief that AI technology should not be controlled by a single corporation. Speaking at Tortoise's Responsible AI Forum in June, he proposed a "world institute" of AI. Such a body might sit under the jurisdiction of the United Nations, Hassabis theorized, and could be filled with top researchers in the field. "It's much stronger if you lead by example," he told the audience, "and I hope DeepMind can be part of that role-modeling for the industry."

Secret behind Airbnb's Billion-Dollar Empire? Spamming Craigslist
reddit
LLM Vibe Score0
Human Vibe Score1
deadcoder0904This week

Secret behind Airbnb's Billion-Dollar Empire? Spamming Craigslist

Silicon Valley wants you to believe that their unicorn startups succeeded doing things legally. But that couldn't be far from truth. For starters, Airbnb used multiple Gmail accounts to spam Craigslist. "They posted unrealistically (fake) cheap rentals of beautiful apartments in places where normal rent should be 10x more. Once people replied, they auto-responded that the unit has been rented, but they should be looking for another unit on AirBnB." The Game of Blackhat is a cat-and-mouse game. You need a lot of guardrails to protect yourself from people using your Social Site by spamming their products. Craigslist is a team of 30 people. There's stuff AI can automate now with such a small team but back then, it wasn't possible. Airbnb used Craigslist as its playground to spam Craigslist visitors to grow their supply-side. In a 2-sided marketplace, growing both supply and demand is very important. And both must grow at the same time for the marketplace to work. A Blackhat Marketer created a new test site to get vacation rental owners to sign-up so that he can test his Airbnb theory. He grabbed their real email-addresses (not Craigslist anonymous addresses) via Craigslist by specifically targeting those who were advertising their vacation rentals on Craigslist. He skipped over the other categories that were directly related to AirBnB's business model because they didn't fit with the test site he built. Once he got 1000+ sign-ups, he then took it upon himself to post it to the advertising section on Craigslist. The email said this: I am emailing you because you have one of the nicest listings on Craigslist in Idaho and I want to recommend you feature it (for free) on one of the largest Idaho housing sites on the web, Airbnb. The site already has 3,000,000 pages views a month. Check it out here to list now: airbnb(dot)com Sarah Surpisingly, all emails were by ladies. He did the same in Week 2 and Week 3 to test if it wasn't a one-time thing. Surely, it wasn't a fluke. After posting 4 ads on Craigslist in 3 weeks, he received 5 identical emails from 2 ladies who were raving fans of AirBnB and spent their days emailing Craigslist advertisers. This is one of the greatest blackhat strategies used in the real world to build a billion-dollar marketplace by growing the supply-side with pure blackhat. These strategies are not mentioned in Press Interviews, Media, or any Founder stories but this is probably the most important piece of the puzzle. Without it, Airbnb probably wouldn't have survived. "Some very famous investors have alluded to the fact that they look for a dangerous streak in the entrepreneurs they invest in…and while those investors will never come out and tell you what they mean, this kind of thing is probably what they mean." It definitely violates CAN-SPAM act. Some comments from Hacker News: "CAN-SPAM, sending from a fake address (illegal headers). CA has a specific law that pre-empts CAN-SPAM that definitely makes this illegal if sent from CA." But I guess it worked in Airbnb's favour lol as they were never caught or fined until after. "It's commercial email 100%. Probably a fake sender name (illegal), against gmail ToS, against CL ToS and no unsubscribe link and no one even subscribed in the first place. 100% against CAN-SPAM." Thanks for reading. If you'd like to learn more blackhat tactics like this, check this site which is a growth hacking newsletter with real-world blackhat examples. PS: Actual emails & screenshots from the Airbnb x Craigslist spam can be found here.

Secret behind Airbnb's Billion-Dollar Empire? Spamming Craigslist
reddit
LLM Vibe Score0
Human Vibe Score1
deadcoder0904This week

Secret behind Airbnb's Billion-Dollar Empire? Spamming Craigslist

Silicon Valley wants you to believe that their unicorn startups succeeded doing things legally. But that couldn't be far from truth. For starters, Airbnb used multiple Gmail accounts to spam Craigslist. "They posted unrealistically (fake) cheap rentals of beautiful apartments in places where normal rent should be 10x more. Once people replied, they auto-responded that the unit has been rented, but they should be looking for another unit on AirBnB." The Game of Blackhat is a cat-and-mouse game. You need a lot of guardrails to protect yourself from people using your Social Site by spamming their products. Craigslist is a team of 30 people. There's stuff AI can automate now with such a small team but back then, it wasn't possible. Airbnb used Craigslist as its playground to spam Craigslist visitors to grow their supply-side. In a 2-sided marketplace, growing both supply and demand is very important. And both must grow at the same time for the marketplace to work. A Blackhat Marketer created a new test site to get vacation rental owners to sign-up so that he can test his Airbnb theory. He grabbed their real email-addresses (not Craigslist anonymous addresses) via Craigslist by specifically targeting those who were advertising their vacation rentals on Craigslist. He skipped over the other categories that were directly related to AirBnB's business model because they didn't fit with the test site he built. Once he got 1000+ sign-ups, he then took it upon himself to post it to the advertising section on Craigslist. The email said this: I am emailing you because you have one of the nicest listings on Craigslist in Idaho and I want to recommend you feature it (for free) on one of the largest Idaho housing sites on the web, Airbnb. The site already has 3,000,000 pages views a month. Check it out here to list now: airbnb(dot)com Sarah Surpisingly, all emails were by ladies. He did the same in Week 2 and Week 3 to test if it wasn't a one-time thing. Surely, it wasn't a fluke. After posting 4 ads on Craigslist in 3 weeks, he received 5 identical emails from 2 ladies who were raving fans of AirBnB and spent their days emailing Craigslist advertisers. This is one of the greatest blackhat strategies used in the real world to build a billion-dollar marketplace by growing the supply-side with pure blackhat. These strategies are not mentioned in Press Interviews, Media, or any Founder stories but this is probably the most important piece of the puzzle. Without it, Airbnb probably wouldn't have survived. "Some very famous investors have alluded to the fact that they look for a dangerous streak in the entrepreneurs they invest in…and while those investors will never come out and tell you what they mean, this kind of thing is probably what they mean." It definitely violates CAN-SPAM act. Some comments from Hacker News: "CAN-SPAM, sending from a fake address (illegal headers). CA has a specific law that pre-empts CAN-SPAM that definitely makes this illegal if sent from CA." But I guess it worked in Airbnb's favour lol as they were never caught or fined until after. "It's commercial email 100%. Probably a fake sender name (illegal), against gmail ToS, against CL ToS and no unsubscribe link and no one even subscribed in the first place. 100% against CAN-SPAM." Thanks for reading. If you'd like to learn more blackhat tactics like this, check this site which is a growth hacking newsletter with real-world blackhat examples. PS: Actual emails & screenshots from the Airbnb x Craigslist spam can be found here.

Secret behind Airbnb's Billion-Dollar Empire? Spamming Craigslist
reddit
LLM Vibe Score0
Human Vibe Score1
deadcoder0904This week

Secret behind Airbnb's Billion-Dollar Empire? Spamming Craigslist

Silicon Valley wants you to believe that their unicorn startups succeeded doing things legally. But that couldn't be far from truth. For starters, Airbnb used multiple Gmail accounts to spam Craigslist. "They posted unrealistically (fake) cheap rentals of beautiful apartments in places where normal rent should be 10x more. Once people replied, they auto-responded that the unit has been rented, but they should be looking for another unit on AirBnB." The Game of Blackhat is a cat-and-mouse game. You need a lot of guardrails to protect yourself from people using your Social Site by spamming their products. Craigslist is a team of 30 people. There's stuff AI can automate now with such a small team but back then, it wasn't possible. Airbnb used Craigslist as its playground to spam Craigslist visitors to grow their supply-side. In a 2-sided marketplace, growing both supply and demand is very important. And both must grow at the same time for the marketplace to work. A Blackhat Marketer created a new test site to get vacation rental owners to sign-up so that he can test his Airbnb theory. He grabbed their real email-addresses (not Craigslist anonymous addresses) via Craigslist by specifically targeting those who were advertising their vacation rentals on Craigslist. He skipped over the other categories that were directly related to AirBnB's business model because they didn't fit with the test site he built. Once he got 1000+ sign-ups, he then took it upon himself to post it to the advertising section on Craigslist. The email said this: I am emailing you because you have one of the nicest listings on Craigslist in Idaho and I want to recommend you feature it (for free) on one of the largest Idaho housing sites on the web, Airbnb. The site already has 3,000,000 pages views a month. Check it out here to list now: airbnb(dot)com Sarah Surpisingly, all emails were by ladies. He did the same in Week 2 and Week 3 to test if it wasn't a one-time thing. Surely, it wasn't a fluke. After posting 4 ads on Craigslist in 3 weeks, he received 5 identical emails from 2 ladies who were raving fans of AirBnB and spent their days emailing Craigslist advertisers. This is one of the greatest blackhat strategies used in the real world to build a billion-dollar marketplace by growing the supply-side with pure blackhat. These strategies are not mentioned in Press Interviews, Media, or any Founder stories but this is probably the most important piece of the puzzle. Without it, Airbnb probably wouldn't have survived. "Some very famous investors have alluded to the fact that they look for a dangerous streak in the entrepreneurs they invest in…and while those investors will never come out and tell you what they mean, this kind of thing is probably what they mean." It definitely violates CAN-SPAM act. Some comments from Hacker News: "CAN-SPAM, sending from a fake address (illegal headers). CA has a specific law that pre-empts CAN-SPAM that definitely makes this illegal if sent from CA." But I guess it worked in Airbnb's favour lol as they were never caught or fined until after. "It's commercial email 100%. Probably a fake sender name (illegal), against gmail ToS, against CL ToS and no unsubscribe link and no one even subscribed in the first place. 100% against CAN-SPAM." Thanks for reading. If you'd like to learn more blackhat tactics like this, check this site which is a growth hacking newsletter with real-world blackhat examples. PS: Actual emails & screenshots from the Airbnb x Craigslist spam can be found here.

OpenAI GPT-4: THE SECRET PROMPT You Need To Know 🤐 #shorts
youtube
LLM Vibe Score0.326
Human Vibe Score0.36
Ishan SharmaApr 2, 2023

OpenAI GPT-4: THE SECRET PROMPT You Need To Know 🤐 #shorts

OpenAI GPT-4: THE SECRET PROMPT You Need To Know 🤐 #shorts 📸 Instagram: https://bit.ly/ishansharma7390ig Join MarkitUpX Discord Server: https://discord.gg/fwSpTje4rh 😁 About Me: https://bit.ly/aboutishansharma 📱 Twitter: https://bit.ly/ishansharma7390twt 📝 LinkedIn: https://bit.ly/ishansharma7390li 🌟 Please leave a LIKE ❤️ and SUBSCRIBE for more AMAZING content! 🌟 3 Books You Should Read 📈Psychology of Money: https://amzn.to/30wx4bW 👀Subtle Art of Not Giving a F: https://amzn.to/30zwWbP 💼Rework: https://amzn.to/3ALsAuz Tech I use every day 💻MacBook Air M1: https://amzn.to/2YWKPjG 📺LG 29' Ultrawide Monitor: https://amzn.to/3aG0p5p 🎥Sony ZV1: https://amzn.to/3ANqgDb 🎙Blue Yeti Mic: https://amzn.to/2YYbiNN ⽴Tripod Stand: https://amzn.to/3mVUiQc 🔅Ring Light: https://amzn.to/2YQlzLJ 🎧Marshall Major II Headphone: https://amzn.to/3lLhTDQ 🖱Logitech mouse: https://amzn.to/3p8edOC 💺Green Soul Chair: https://amzn.to/3mWIxZP ✨ Tags ✨ ishan sharma,OpenAI GPT-4: THE SECRET PROMPT You Need To Know 🤐,ai tools,artificial intelligence,microsoft,microsoft openai,microsoft chatgpt,what is chatgpt,chatgpt tutorial,chatgpt 4,make money with chatgpt,chatgpt,open ai,gpt 4,gpt 4 vs gpt 3,openai,openai chatbot gpt,chatgpt explained,chatgpt examples,chat gpt,open ai gpt 4,chat gpt uses,chatgpt4 tutorial,gpt 4 demo,gpt 4 image input,gpt 4 review,gpt 4 how to use,gpt 4 prompt ✨ Hashtags ✨ #gpt4 #chatgpt #ai

From “Green” to “Smart” – Tom Gorski’s Word of Advice
reddit
LLM Vibe Score0
Human Vibe Score1
DanielleHarrison1This week

From “Green” to “Smart” – Tom Gorski’s Word of Advice

Sharing this interview with entrepreneur Tom Gorski. I think it contains a few nice tips for beginner entrepreneurs. What is the problem with the term “Green?” what are the top 3 mistakes entrepreneurs make that can prevent them from enjoying the sweet taste of success? And what should young entrepreneurs always keep in mind? Continuing our expert interview series, we asked entrepreneur Tom Gorski to share some of his secrets to success with us. Gorski is the CEO and Co-Founder at SaaSGenius.com, and an Inbound Marketer & Growth Hacker at InboundWay.com. His career spans over 12 years of developing and implementing online marketing, SEO and conversion optimization campaigns. He defines his biggest accomplishment to date as “achieving 4500% growth for one of my clients over a three­year period.” logo-saasgenius Q: It’s no secret that the SaaS market is saturated, as new companies are having very hard time acquiring, retaining and monetizing users. In your view – what are the top 3 mistakes SaaS companies make? What are some key differentiators you recognize in a successful product? A: Mistake No. 1: Product-market fit is not good enough There are a number of reasons for this, including the fact that inertia, incumbency and bureaucracy are all working against you. For emerging companies, this means finding a way to be exponentially better with fewer resources. As a result, focus is key. Mistake No. 2: Not Specializing Your Sales Roles When you specialize your sales people, you allow them to focus, which creates greater output form your sales team. Mistake No. 3: You Need a Niche To be able to market and sell well, you need to have a niche. The world is noisy and messy, and you’ll struggle if you don’t have a sharp, direct message. When you try to speak to everyone, no one can hear you. Q: Which innovative trends do you recognize in the high tech world nowadays? A: “Green” was a mega trend of the last decade and while it will continue to be very important, there will be a shift towards “smart” solutions, which are intelligent, connected and have the ability to sense, report, and take the right action. Smart solutions will be everywhere around us from smart clothing, phones, to smart homes and smart cities. Q: What is the most significant advice you can give young entrepreneurs? A: Being very successful means learning from those who have already achieved success. Having a mentor is an amazing blessing to an entrepreneur, but not everyone can find one in person. My advice is to work smarter, not harder. This is the most non-intuitive observation I will probably make. If you want to compete in the arena, hard work isn’t enough. And judging yourself on how hard you work, rather than how smart you work can be fatal. Q: We are flooded with buzzwords lately – VR / AI / Bots… where do you think the software world is heading? A: AI and bots are a very hot topic in 2016 and it’s sometimes hard to distinguish the real potential behind the hype. My point of view is that, like with many things, there’s no revolution but evolution. It’s unrealistic to think that AI can become mainstream in SaaS products without proper AI infrastructure. SaaS delivery will significantly outpace traditional software product delivery, growing nearly five times faster than the traditional software market and will become a significant growth driver for all functional software markets. By 2019, the SaaS software model will account for $1 of every $4 spent on software. Q: Let us in on some of your secrets… where do you look for innovation? For inspiration and revolutionary ideas? A: Ideas for new startups often begin with a real problem that needs to be solved. And they don’t come while you’re sitting around sipping coffee and contemplating life. They tend to reveal themselves while you’re at work on something else. Start with brainstorming with problems that you are personally invested in. Building a business is hard and takes the kind of relentless dedication that comes from personal passion. Perhaps the greatest factor that determines whether or not an entrepreneur will be successful isn’t the business idea itself, but rather the entrepreneur’s willingness to try to turn the idea into reality. Great ideas are abundant, but it’s what we decide to do with them that counts. Original post: http://saasaddict.walkme.com/from-green-to-smart-tom-gorskis-words-of-advice/

A Structured Approach to Ideation and Validation (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

A Structured Approach to Ideation and Validation (I will not promote)

Hi all, I used to work in VC and wanted to share some startup knowledge and insights from startup founders I know. Recently, I interviewed a friend of mine who built an AI Robotics startup ("Hivebotics") that creates automated toilet-cleaning robots. I can't post the full article because of Reddit's word limit, so I'll be posting it in sections here instead. This first section of the transcript goes through his approach to ideation and validation. Enjoy and let me know what you think! (I will not promote) \\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\ (1) Ideation and Validation Problem-Market-Solution Framework I like to think of startup ideation and validation using this framework: Problem– What exactly are you solving? Observation– How you identify a problem to work on User Research– How you further understand that problem Market– Is there a large enough market for solving this problem? Size– How many people experience this same problem? Demand– How many of those people are willing to pay for the solution? Solution– Your answer to the problem Desirability– Whether people actually want your solution Feasibility– Whether building the solution is practical and realistic Viability– Whether your solution can generate revenue Problem You always need to start problem-first, which is something that was really drilled into me during my time at Stanford. Too often, founders rush to build solutions first—apps or products they find exciting—without confirming whether there's any real demand for it. The first step is always to identify a specific problem, then further understand its scale, urgency and further details by talking to potential users. Observation– To find problems, observation is key. People may not even realise the inefficiencies in their processes until you point them out. That’s why interviews and field research are so important. There are problems all around us, so it's simply a matter of going out, paying attention and being attuned to them as they occur. User Research– To further understand the problem, conducting user research by interviewing potential customers is essential. Personally, I like to use the "Mom Test" when I conduct interviews to avoid biased and generic feedback. Don’t just ask theoretical questions and avoid being too specific—observe how your potential users work, ask about pain points, and use broad, open-ended questions to ensure you aren't leading them to a specific answer. Market Once you've found an actual problem and talked to enough potential users to really understand its specific pain points, the next step is to determine the market size and demand for a solution. Size– Determining the market size is essential because it determines whether or not it's commercially worthwhile to pursue the problem and develop a solution for it. You need to determine if there are enough potential customers out there experiencing this problem to gauge the market size. There's no secret strategy for this; you have to interview as many potential users as possible to confirm that it's a widespread problem in the industry. Demand– Make sure that you're working on a problem that people will gladly pay to have solved. Even if the problem is large enough, you have to make sure it's painful enough to warrant a paid solution. If many people experience the same problem, but aren't willing to pay for a solution, then you don't have a market and should look for a different problem to validate. Another way of looking at it is that your true market size is the number of potential customers actually willing to pay* for the solution to the problem, not the number of people simply experiencing the same problem. Solution When validating a potential solution to the problem, I would look at the 3 factors of desirability, feasibility and viability. Desirability– the degree to which a solution appeals to people and fulfills their wants and needs. Without strong desirability, even the most technically advanced or economically practical product is unlikely to succeed. The best way to test this is to secure financial commitments early on during the proof-of-concept stage. Most people are polite, so they may simply tell you that your startup's product is good even if it's not. However, if they're actually willing to pay for the solution, this is actual evidence of your product's desirability. Don't just ask people if they would pay for it; actually see whether they will pay for it. Feasibility– whether a product can be built using existing technical capabilities. A lack of feasibility makes it challenging or impossible to develop the product, no matter how appealing it might be to users or how promising its financial prospects are. This is just a matter of conducting initial research and actually trying to build a prototype, which will inform you whether the fully-realised product is truly feasible. Viability– the product's ability to generate sustainable financial returns. Without financial viability, the business supporting the product cannot endure, even if the product is highly appealing to users and technically achievable. Here, you need to look at your unit economics, development costs and other expenses to determine the viability of your solution. \\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\ Hope you enjoyed reading this; let me know your honest thoughts in the comments and I'll try to improve how I interview founders based on those!

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.
reddit
LLM Vibe Score0
Human Vibe Score0.714
level6-killjoyThis week

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.

This is a recap covering the major news from last week. 🔥Top 3 news - OpenAI’s updates, Meta’s upcoming free LLM and EU Regulation 🗞️Interesting reads include PSA about protecting your keys, The GPT ouroboros, Reddit - OpenAI’s moat, and more.. 🧑‍🎓Learning includes a Step-by-step guide from a non-technical founder who launched his MVP, Chatbot for your Gdrive and more 🔥Top 3 AI news in the past week OpenAI: New Pricing, Models, & Functions OpenAI has been on a roll. Last week we saw the release of OpenAI best practice on using GPT. This week we saw some amazing updates. Three major buckets were: First, the price decreases for both embeddings and GPT-3.5 tokens. Second, new models for gpt-4 and gpt-3.5. A new longer context model for gpt-3.5. Third, a new function calling capability. Why is it important? Previously, the output from OpenAI was all text. So, calling an external API from GPT was quite difficult. You had to parse the text data and things were often incorrect. Langchain created the Agents and Tools feature to tackle this problem. It was still unreliable and prone to issues. Now you get native support to generate a fixed format output. You can use the output to generate functional calls and also pass functions which need to be called. For example, if your app has multiple API endpoints then you can use GPT to generate the API calls with parameters. You can also pass the endpoints as function calls to ensure the correct function is executed. This functionality can further be used to generate structured data (JSON) out of GPT. So, you can generate data from GPT and load it into your backend. What’s next? This functionality allows turning natural language responses into structured data. This can be used to create “intelligent” backends using LLMs. We might see implementations in no-code tools to allow more robust and natural-language tools for non-technical folks. The structured data process goes both ways. You can also feed structured data into GPT for better responses. This feature also has its share of issues. Function calling suffers from the same prompt injection issues. Malicious actors can pass malicious code in function or the responses. For example, creation of queries using functions might contain malicious code to delete data. Without proper user validation this code will be executed automatically and delete data. So, using LLM as the back-end layer needs proper security implementation. Meta's LLM: Commercial Use Ahead Llama has been a boon for the open source community. Many of the open source models rely on Llama. The issue is that Llama is research-only and cannot be used commercially. So, no one can use it to build any product. Meta is now working on the next version of the model. This model will be available for commercial use. This is in stark contrast to both OpenAI and Google. Both safe-guarde their models and make it available through API. Why is it important? Certain industries cannot use LLM APIs because of strict restrictions on data privacy. These companies would want to run their own instance of a foundational model. A commercially available foundational model is also going to help people who want to keep their “API call” costs next to 0. A commercially available free-for-all model will also help push the open source community further. Just like Llama. What’s next? Sam Altman has said OpenAI didn’t release GPT-3 as open-source because they didn’t think people would be able to run it. Now OpenAI is working on an open-source model. This is going to be weaker than GPT-4. Let the battle of LLMs begin. EU's Proposed Legislation and Its Impact on AI Usage The EU parliament voted to move ahead with the E.U. AI Act. This act aims to ensure consumer protection against the dangers of AI. Why is it important? OpenAI and Sam Altman want regulations for models. They have proposed a IAEA-type of agency to stop the proliferation of LLM models. As per OpenAI, all models should be regulated and monitored. The suggestion of a license based regulation has led to significant backlash. Many people have called it “regulatory capture” - with the aim of shutting down competing LLMs. Licensing based regulations might not really be effective. The EU is approaching regulation from a different angle. It doesn’t focus on how models are developed. Rather focuses on how AI will/can be used. They have broken down use cases into 4 categories - unacceptable (prohibited), high, medium and low risk. For example, Building a Pre-Crime software,on%20crimes%20not%20yet%20committed.) to predict crimes? Building a Social credit system? Unacceptable. Using tools to influence elections or recommendation algorithms? High (Highly regulated). Using generative AI tools to create text or images on news sites? Medium (Add label that the content is AI generated) AI providers also need to disclose their training source. To me this sounds like good legislation. What do you guys think? But, OpenAI has warned that EU regulations might force them to pull out completely. What’s next? The disclosure requirements might help various publishing companies. AI and media companies are in talks to pay for training data. Google has been leading the charge. Additionally, OpenAI and Deepmind will open their models for safety and research purposes to the UK government. 🗞️10 AI news highlights and interesting reads PSA: If you are using Repl to write code, you might want to check your OpenAI API keys. If you have left them embedded then people can pirate and steal the keys. LLMs rely on human annotation or human feedback to learn. And one way to generate human annotation is crowdsourcing. But what if the crowdsource human annotators use LLMs? Research shows 33-46% workers used LLMs. So, basically we go from Human -> AI -> Human -> AI. The AI ouroboros. Researchers also say generated data to train models might cause serious issue. All the talks about moats \- Reddit might be OpenAI’s \future\ moat. Given the amount of complaints about how Google search experience has deteriorated during the blackout, this might be true? Doctors are using ChatGPT but not to diagnose.Rather to be more empathetic. We discussed this just a month ago. And guess where the data for this study came from? Reddit AskDocs. Moat FTW?! Beatles to make a comeback…using Generative AI. SnapFusion - Text to Image diffusion on mobile phones. Large context lengths are important for better GPT experience. The secret sauce for 100k context length. There is a lot of bad AI research out there. Some border on snake oil. Most AI “research” should be double checked and challenged. A new research on huggingface said that GPT-4 can ace MIT curriculum. Now someone is replicating the results and say that GPT-4 can’t beat MIT. Are we seeing peak AI? Especially when people from Deepmind and Meta are involved? Mistral AI raised $113 million in seed round with no product. Some might say this funding is for the team and the team is really solid. The issue though is whether the valuation is justified when OpenAI and Google already have a head start. The AI Hype Wall of Shame. \- Collection of articles which mislead people about AI in various aspects. 🧑‍🎓3 Learning Resources Building and Launching a company using GPT-4 with prompts. (The author didn’t know how to code but created and launched the MVP in a month). Chatbot for your Gdrive - https://www.haihai.ai/gpt-gdrive/ Building ChatGPT plugin using Supabase - https://supabase.com/blog/building-chatgpt-plugins-template That’s it folks. Thank you for reading and have a great week ahead. If you are interested in a focused weekly recap delivered to your inbox on Mondays you can subscribe here. It is FREE!

I'm Building an "AiExecutiveSuperAgent_Systems_Interface" between humanity and the Ai world, as well as each other... Let's Talk?
reddit
LLM Vibe Score0
Human Vibe Score1
Prudent_Ad_3114This week

I'm Building an "AiExecutiveSuperAgent_Systems_Interface" between humanity and the Ai world, as well as each other... Let's Talk?

Ok... So look... This one is pretty crazy... I'm building an Ai Interface that knows me better than I know myself - Check, lots of people have this, either in reality with employees and family members, or with ai intelligence. But it doesn't just know Me... It knows how to talk with Me. It understands my language, because I've trained it to. I've also trained it to translate that to all my clients and HumanAgents, soon to become RobotAgents... The RESULT: I can literally just spend 1-18 hours talking to it, and things get DONE. Most of that time, I just say EXECUTE, or ENGAGE, or DRAFT, or DISPATCH. I feel like a secret agent communicating in codes with his agency 😂 Not great for the paranoiac in me, but it's easy to get that part under control, ya'll. It's like having a team of 10,000 people, all available 24/7, all perfectly synchronised to each other's communication styles, preferences and ultimately: WHAT DO YOU NEED ME TO DO. At the end of the it all, having run my single COMMAND through a thousand of those people, a Document is prepared that outlines the next 3 stages of the plan, along with instructions to the whole team for how to ENACT it. Sounds rather grand and wonderful... Even when I simply use it to help me come up with a filing system for my creative work... \\\\\\\\\\\\\\\\\\\\\\ Here's my current VISION, why I'm doing this AND why I'm doing it publicly despite it being top secret. VISION To create an army of User-Owned and Operated "AiSuperAgencies" which gather intelligence on the user, securely file and analyse it, and then construct a sub-army of agents and tools that work together to produce the desired output, for any Function in the Personal and Professional Lives of EVERYONE, EVERYWHERE, in 3-5 Years. To start, I'm building it for me and the 5-10 cleaners who've made it to Level 1 in my access system. They were sick of toxic employers, tyrannical agencies and greedy customers. They gathered around us (many came in, many went out, few stayed, took about a year for our core team of 3 Level 2 Cleaners. My goal has always been to never employ anyone. Just me, my Partner and the Cleaners. All Shared Owners in the system for delivering the right cleaner to the right house in our town, at the right time and without any dramas or arguments... I have a personal talent for resolving disputes, which has made working for and buying from my business a mostly enjoyable and upbeat experience, with a touch of mystery and a feeling that you're part of something big! It is a business that ran on Me. I put in my time, every day, building automated tool after automated tool. Hiring a contractor to do a job, scratching my head when it didn't add enough value to pay for itself, then just doing it myself again. I wanted to solve that problem. I'm trusting that the few who hear about it who actually see the potential, will just come join us, no dramas, just cool people partnering up! And those that don't, won't. No one could steal it, because it's Mine, and I'll just change the keys anyway loser! Enjoy digging through my past, you lunatic! I'm out here living Now. Anyways... It's lonely around here. I have a cleaning business that I run from my laptop, which means I can live anywhere, but I still had this big problem of time... NOT ENOUGH Oh Wait. It's Here.

I'm Building an "AiExecutiveSuperAgent_Systems_Interface" between humanity and the Ai world, as well as each other... Let's Talk?
reddit
LLM Vibe Score0
Human Vibe Score1
Prudent_Ad_3114This week

I'm Building an "AiExecutiveSuperAgent_Systems_Interface" between humanity and the Ai world, as well as each other... Let's Talk?

Ok... So look... This one is pretty crazy... I'm building an Ai Interface that knows me better than I know myself - Check, lots of people have this, either in reality with employees and family members, or with ai intelligence. But it doesn't just know Me... It knows how to talk with Me. It understands my language, because I've trained it to. I've also trained it to translate that to all my clients and HumanAgents, soon to become RobotAgents... The RESULT: I can literally just spend 1-18 hours talking to it, and things get DONE. Most of that time, I just say EXECUTE, or ENGAGE, or DRAFT, or DISPATCH. I feel like a secret agent communicating in codes with his agency 😂 Not great for the paranoiac in me, but it's easy to get that part under control, ya'll. It's like having a team of 10,000 people, all available 24/7, all perfectly synchronised to each other's communication styles, preferences and ultimately: WHAT DO YOU NEED ME TO DO. At the end of the it all, having run my single COMMAND through a thousand of those people, a Document is prepared that outlines the next 3 stages of the plan, along with instructions to the whole team for how to ENACT it. Sounds rather grand and wonderful... Even when I simply use it to help me come up with a filing system for my creative work... \\\\\\\\\\\\\\\\\\\\\\ Here's my current VISION, why I'm doing this AND why I'm doing it publicly despite it being top secret. VISION To create an army of User-Owned and Operated "AiSuperAgencies" which gather intelligence on the user, securely file and analyse it, and then construct a sub-army of agents and tools that work together to produce the desired output, for any Function in the Personal and Professional Lives of EVERYONE, EVERYWHERE, in 3-5 Years. To start, I'm building it for me and the 5-10 cleaners who've made it to Level 1 in my access system. They were sick of toxic employers, tyrannical agencies and greedy customers. They gathered around us (many came in, many went out, few stayed, took about a year for our core team of 3 Level 2 Cleaners. My goal has always been to never employ anyone. Just me, my Partner and the Cleaners. All Shared Owners in the system for delivering the right cleaner to the right house in our town, at the right time and without any dramas or arguments... I have a personal talent for resolving disputes, which has made working for and buying from my business a mostly enjoyable and upbeat experience, with a touch of mystery and a feeling that you're part of something big! It is a business that ran on Me. I put in my time, every day, building automated tool after automated tool. Hiring a contractor to do a job, scratching my head when it didn't add enough value to pay for itself, then just doing it myself again. I wanted to solve that problem. I'm trusting that the few who hear about it who actually see the potential, will just come join us, no dramas, just cool people partnering up! And those that don't, won't. No one could steal it, because it's Mine, and I'll just change the keys anyway loser! Enjoy digging through my past, you lunatic! I'm out here living Now. Anyways... It's lonely around here. I have a cleaning business that I run from my laptop, which means I can live anywhere, but I still had this big problem of time... NOT ENOUGH Oh Wait. It's Here.

My Side Projects: From CEO to 4th Developer (Thanks, AI 🤖)
reddit
LLM Vibe Score0
Human Vibe Score1
tilopediaThis week

My Side Projects: From CEO to 4th Developer (Thanks, AI 🤖)

Hey Reddit 👋, I wanted to share a bit about some side projects I’ve been working on lately. Quick background for context: I’m the CEO of a mid-to-large-scale eCommerce company pulling in €10M+ annually in net turnover. We even built our own internal tracking software that’s now a SaaS (in early review stages on Shopify), competing with platforms like Lifetimely and TrueROAS. But! That’s not really the point of this post — there’s another journey I’ve been on that I’m super excited to share (and maybe get your feedback on!). AI Transformed My Role (and My Ideas List) I’m not a developer by trade — never properly learned how to code, and to be honest, I don’t intend to. But, I’ve always been the kind of guy who jots down ideas in a notes app and dreams about execution. My dev team calls me their “4th developer” (they’re a team of three) because I have solid theoretical knowledge and can kinda read code. And then AI happened. 🛠️ It basically turned my random ideas app into an MVP generation machine. I thought it’d be fun to share one of the apps I’m especially proud of. I am also planning to build this in public and therefore I am planning to post my progress on X and every project will have /stats page where live stats of the app will be available. Tackling My Task Management Problem 🚀 I’ve sucked at task management for YEARS, I still do! I’ve tried literally everything — Sheets, Todoist, Asana, ClickUp, Notion — you name it. I’d start… and then quit after a few weeks - always. What I struggle with the most is delegating tasks. As a CEO, I delegate a ton, and it’s super hard to track everything I’ve handed off to the team. Take this example: A few days ago, I emailed an employee about checking potential collaboration opportunities with a courier company. Just one of 10s of tasks like this I delegate daily. Suddenly, I thought: “Wouldn’t it be AMAZING if just typing out this email automatically created a task for me to track?” 💡 So… I jumped in. With the power of AI and a few intense days of work, I built a task manager that does just that. But of course, I couldn’t stop there. Research & Leveling It Up 📈 I looked at similar tools like TickTick and Todoist, scraped their G2 reviews (totally legally, promise! 😅), and ran them through AI for a deep SWOT analysis. I wanted to understand what their users liked/didn’t like and what gaps my app could fill. Some of the features people said they were missing didn’t align with the vision for my app (keeping it simple and personal), but I found some gold nuggets: Integration with calendars (Google) Reminders Customizable UX (themes) So, I started implementing what made sense and am keeping others on the roadmap for the future. And I’ve even built for that to, it still doesn’t have a name, however the point is you select on how many reviews of a specific app you want to make a SWOT analysis on and it will do it for you. Example for Todoist in comments. But more on that, some other time, maybe other post ... Key Features So Far: Here’s what’s live right now: ✅ Email to Task: Add an email as to, cc, or bcc — and it automatically creates a task with context, due dates, labels, etc. ✅ WhatsApp Reminders: Get nudged to handle your tasks via WhatsApp. ✅ WhatsApp to Task: Send a message like /task buy groceries — bam, it’s added with full context etc.. ✅ Chrome Extension (work-in-progress): Highlight text on any page, right-click, and send it straight to your task list. Next Steps: Build WITH the Community 👥 Right now, the app is 100% free while still in the early stages. But hey, API calls and server costs aren’t cheap, so pricing is something I’ll figure out with you as we grow. For now, my goal is to hit 100 users and iterate from there. My first pricing idea is, without monthly subscription, I don’t want to charge someone for something he didn’t use. So I am planning on charging "per task", what do you think? Here’s what I have planned: 📍 End of Year Goal: 100 users (starting from… 1 🥲). 💸 Revenue Roadmap: When we establish pricing, we’ll talk about that. 🛠️ Milestones: Post on Product Hunt when we hit 100 users. Clean up my self-written spaghetti code (hire a pro dev for review 🙃). Hire a part-time dev once we hit MRR that can cover its costs. You can check how are we doing on thisisatask.me/stats Other Side Projects I’m Working On: Because… what’s life without taking on too much, right? 😂 Full list of things I’m building: Internal HRM: Not public, tried and tested in-house. Android TV App: Syncs with HRM to post announcements to office TVs (streamlined and simple). Stats Tracker App: Connects to our internal software and gives me real-time company insights. Review Analyzer: Scrapes SaaS reviews (e.g., G2) and runs deep analysis via AI. This was originally for my Shopify SaaS but is quickly turning into something standalone. Coming soon! Mobile app game: secret for now. Let’s Build This Together! Would love it if you guys checked out https://thisisatask.me and gave it a spin! Still super early, super raw, but I’m pumped to hear your thoughts. Also, what’s a must-have task manager feature for you? Anything that frustrates you with current tools? I want to keep evolving this in public, so your feedback is gold. 🌟 Let me know, Reddit! Are you with me? 🙌

How to get your first 10 customers with cold email
reddit
LLM Vibe Score0
Human Vibe Score0.905
LieIgnorant6304This week

How to get your first 10 customers with cold email

Cold email is an insane channel for growth, especially for bootstrapped startups as it's very low cost but completely scalable. Yet there's a huge difference between blind cold emailing and crafting personalized outreach for select individuals. The latter is a legit channel which makes many businesses scale in short amounts of time (i.e. see Alex Hormozi’s ‘$100 Million Dollar Offer’). My goal here is to help other founders do what I did but quicker. So you can learn faster. And then teach me something new too. These are the step-by-step lessons I've learnt as a bootstrapped founder, showing you how to use cold email to get your first customers: Find your leads Write engaging email copy Personalize your outreach Send emails Scale up Find your leads This is a key step. Once you figure out exactly who you want to target and where to find them, you'll be printing money. There's a few different ways to go about finding valuable leads. The secret? Keep testing different approaches until you strike gold. First, dedicate some time every day to find and organise leads. Then, keep an eye on your numbers and bounce rates. If something's not working, switch it up. Stick with what's bringing in results and ditch what's not. It's all about staying flexible and learning as you go. Apollo.io is a great starting point as an effective lead source. Their tool allows you to specify filters including job titles, location, company size, industry, keywords, technologies, and revenue. Get specific with your searches to find your ideal customers. Once you have some results you can save and export them, you'll get a list of contact information including name, email, company, LinkedIn, ready to be verified and used. LinkedIn Sales Navigator is another good source. You can either do manual searches or use a scraper to automate the process. The scrapers I'd recommend checking out are FindyMail and Evaboot. As with Apollo, it's best to get very specific with your targeting so you know the prospect will be interested in your offer. BuiltWith is more expensive but ideal if you're targeting competitors. With BuiltWith you can build lists based on what technologies companies are using. For example if you're selling a Shopify app, you'd want to know websites or stores using Shopify, and reach out to them. The best lead sources will always be those that haven't been contacted a lot in the past. If you are able to find places where your target audience uniquely hangs out, and you can get their company website domains, they have the potential to be scrapped, and you have a way to personalize like "I spotted your comment on XYZ website". Once you've got your leads, keep them organized. Set up folders for different niches, countries, company sizes, so you can review what works and what doesn't. One more thing – before you start firing off emails, make sure those addresses are verified. Always use an email verifier to clean up your list and avoid bounces that may affect your sending reputation, and land you in the spam folder. I use Neverbounce for this but there are other tools available. Write engaging email copy Writing a good copy that gets replies is difficult, it changes depending on your offer/audience and nobody knows what's going to work. The best approach is to keep testing different targeting and messaging until you find what works. However, there are some key rules to stick to that I've outlined. For the subject line, keep it short and personalized. Try to write something that sparks interest, and mention the recipients name: Thought you’d like this {{first name}} {{firstName}} - quick question For the email body it's best to use a framework of personalization, offer, then call to action. Personalization is an entire subject in its own right, which I've covered below. In short, a personalized email opener is the best way to grab their attention, and let them know the email is relevant to them and to keep reading. Take it from Alex Hormozi and his $100M Offers playbook – your offer is very important to get right. Make sure your offer hits the mark for your target audience, and get as specific as possible. For example: I built a SaaS shopify app for small ecommerce businesses selling apparel that doubles your revenue in 60-days or your money back. We developed a cold email personalization tool for lead generation agencies that saves hundreds of hours, and can 3x your reply rate. Lastly, the CTA. The goal here isn't to get sign-ups directly from your first email. It's better to ask a brief question about whether the prospect would be interested in learning more. Something very low friction, that warrants a response. Some examples might include: Would you be interested in learning more about this? Can we connect a bit more on this? Mind if I send over a loom I recorded for you? Never send any links in the first email. You've reached out to this person because you have good reason to believe they'd find real value in your offer, and you want to verify if that's the case. After you get one reply, this is a great positive signal and from there you can send a link, book a call, provide a free resource, whatever makes sense based on their response. Personalize your outreach Personalization is one of the most important parts of the process to get right. Your recipient probably receives a multitude of emails every day, how can you make yours stand out, letting them know you've done your research, and that your email is relevant to them? Personalizing each email ensures you get more positive replies, and avoid spam filters, as your email is unique and hasn't been copied and pasted a million times over. The goal is to spark the recipient's interest, and let them know that you're contacting them for good reason. You might mention a recent achievement, blog post or product release that led you to reach out to the prospect specifically. For example: Your post on "Doing Nothing" gave me a good chuckle. Savvy marketing on Cadbury's part. Saw that you've been at Google for just under a year now as a new VP of sales. Spotted that you've got over 7 years of experience in the digital marketing space. Ideally you'll mention something specifically about the prospect or their company that relates to your offer. The downside to personalization is that it's hard to get right, and very time consuming at scale, but totally worth it. Full disclosure, me and my partner Igor just launched our new startup ColdClicks which uses AI to generate hyper-personalized email openers at scale. We built the tool as we were sending hundreds of emails a day, and personalizing every individual email took hours out of our day. ColdClicks automates this process, saving you time and getting you 2-3x more replies. Send emails At this stage you've decided on who you're targeting, you've mined some leads, and written copy. Now it's time to get sending. You can do this manually by copy and pasting each message, but one of the reasons cold email is so powerful is that it's scalable. When you build a process that gets customers, you'll want to send as many emails as you can to your target market. To get started quickly, you can use a mail-merge gmail tool, the best I've used is Maileteor. With Maileteor you upload your lead data to Google sheets, set-up an email template and Mailmetor will send out emails every day automatically. In your template you can define variables including name, company, and personalization to ensure your email is unique for each recipient. Alternatively, you may opt for a more comprehensive tool such as Instantly. Instantly includes unlimited email sending and accounts. There's more initial setup involved as you'll need to set-up Google workspace, buy sending domains, and warm up your email accounts, but when you become familiar with the process you can build a powerful lead generation / customer acquisition machine. Some key points to note, it's very important to warm up any new email accounts you set up. Warmup is the process of gradually establishing a positive reputation with email service providers like Gmail or Yahoo. Make sure to set up DKIM and DMARC on those new email accounts too, to maximise your chances of landing in the inbox. Scale up Once you've found a process that works, good things happen, and it becomes a numbers game. As you get replies and start to see new users signing up, you'll want to scale the process and send more emails. It's straightforward to add new sending accounts in a sending tool like Instantly, and you'll want to broaden your targeting when mining to test new markets. Unfortunately, sending more emails usually comes with a drop in reply rate as you have less time to personalize your messaging for each recipient. This is where ColdClicks shines. The tool allows you to upload thousands of leads and generate perfectly relevant email personalizations for every lead in your list, then export to your favorite sending tool. The examples I listed above in the personalization section were all generated by ColdClicks. Wrapping it up Cold email is an amazing way to validate your product and get new customers. The channel gets a bad rap, but there's a huge difference between blind cold emailing and crafting personalized outreach for individuals who will find value in your product. It's perfect for bootstrapped founders due to its affordability and scalability, and it's the driver of growth for many SaaS businesses. Time to get your first 10 customers! As you start sending, make it a habit to regularly check for new leads. Always experiment with market/messaging, track every campaign so you can learn what's working and iterate, and when you do get positive responses, reply as soon as you can!

[D] Gary Marcus and Luis Lamb -- discussion of AGI and Neurosymbolic methods
reddit
LLM Vibe Score0
Human Vibe Score1
timscarfeThis week

[D] Gary Marcus and Luis Lamb -- discussion of AGI and Neurosymbolic methods

https://youtu.be/nhUt6mKCPf8 Pod: https://anchor.fm/machinelearningstreettalk/episodes/54-Gary-Marcus-and-Luis-Lamb---Neurosymbolic-models-e125495 Professor Gary Marcus is a scientist, best-selling author, and entrepreneur. He is Founder and CEO of Robust.AI, and was Founder and CEO of Geometric Intelligence, a machine learning company acquired by Uber in 2016. Gary said in his recent next decade paper that — without us, or other creatures like us, the world would continue to exist, but it would not be described, distilled, or understood. Human lives are filled with abstraction and causal description. This is so powerful. Francois Chollet the other week said that intelligence is literally sensitivity to abstract analogies, and that is all there is to it. It's almost as if one of the most important features of intelligence is to be able to abstract knowledge, this drives the generalisation which will allow you to mine previous experience to make sense of many future novel situations. Also joining us today is Professor Luis Lamb — Secretary of Innovation for Science and Technology of the State of Rio Grande do Sul, Brazil. His Research Interests are Machine Learning and Reasoning, Neuro-Symbolic Computing, Logic in Computation and Artificial Intelligence, Cognitive and Neural Computation and also AI Ethics and Social Computing. Luis released his new paper Neurosymbolic AI: the third wave at the end of last year. It beautifully articulated the key ingredients needed in the next generation of AI systems, integrating type 1 and type 2 approaches to AI and it summarises all the of the achievements of the last 20 years of research. We cover a lot of ground in today's show. Explaining the limitations of deep learning, Rich Sutton's the bitter lesson and "reward is enough", and the semantic foundation which is required for us to build robust AI.

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.
reddit
LLM Vibe Score0
Human Vibe Score0.765
hardmaruThis week

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.

Schmidhuber interview expressing his views on the future of AI and AGI. Original source. I think the interview is of interest to r/MachineLearning, and presents an alternate view, compared to other influential leaders in AI. Juergen Schmidhuber, Renowned 'Father Of Modern AI,' Says His Life’s Work Won't Lead To Dystopia May 23, 2023. Contributed by Hessie Jones. Amid the growing concern about the impact of more advanced artificial intelligence (AI) technologies on society, there are many in the technology community who fear the implications of the advancements in Generative AI if they go unchecked. Dr. Juergen Schmidhuber, a renowned scientist, artificial intelligence researcher and widely regarded as one of the pioneers in the field, is more optimistic. He declares that many of those who suddenly warn against the dangers of AI are just seeking publicity, exploiting the media’s obsession with killer robots which has attracted more attention than “good AI” for healthcare etc. The potential to revolutionize various industries and improve our lives is clear, as are the equal dangers if bad actors leverage the technology for personal gain. Are we headed towards a dystopian future, or is there reason to be optimistic? I had a chance to sit down with Dr. Juergen Schmidhuber to understand his perspective on this seemingly fast-moving AI-train that will leap us into the future. As a teenager in the 1970s, Juergen Schmidhuber became fascinated with the idea of creating intelligent machines that could learn and improve on their own, becoming smarter than himself within his lifetime. This would ultimately lead to his groundbreaking work in the field of deep learning. In the 1980s, he studied computer science at the Technical University of Munich (TUM), where he earned his diploma in 1987. His thesis was on the ultimate self-improving machines that, not only, learn through some pre-wired human-designed learning algorithm, but also learn and improve the learning algorithm itself. Decades later, this became a hot topic. He also received his Ph.D. at TUM in 1991 for work that laid some of the foundations of modern AI. Schmidhuber is best known for his contributions to the development of recurrent neural networks (RNNs), the most powerful type of artificial neural network that can process sequential data such as speech and natural language. With his students Sepp Hochreiter, Felix Gers, Alex Graves, Daan Wierstra, and others, he published architectures and training algorithms for the long short-term memory (LSTM), a type of RNN that is widely used in natural language processing, speech recognition, video games, robotics, and other applications. LSTM has become the most cited neural network of the 20th century, and Business Week called it "arguably the most commercial AI achievement." Throughout his career, Schmidhuber has received various awards and accolades for his groundbreaking work. In 2013, he was awarded the Helmholtz Prize, which recognizes significant contributions to the field of machine learning. In 2016, he was awarded the IEEE Neural Network Pioneer Award for "pioneering contributions to deep learning and neural networks." The media have often called him the “father of modern AI,” because the most cited neural networks all build on his lab’s work. He is quick to point out, however, that AI history goes back centuries. Despite his many accomplishments, at the age of 60, he feels mounting time pressure towards building an Artificial General Intelligence within his lifetime and remains committed to pushing the boundaries of AI research and development. He is currently director of the KAUST AI Initiative, scientific director of the Swiss AI Lab IDSIA, and co-founder and chief scientist of AI company NNAISENSE, whose motto is "AI∀" which is a math-inspired way of saying "AI For All." He continues to work on cutting-edge AI technologies and applications to improve human health and extend human lives and make lives easier for everyone. The following interview has been edited for clarity. Jones: Thank you Juergen for joining me. You have signed letters warning about AI weapons. But you didn't sign the recent publication, "Pause Gigantic AI Experiments: An Open Letter"? Is there a reason? Schmidhuber: Thank you Hessie. Glad to speak with you. I have realized that many of those who warn in public against the dangers of AI are just seeking publicity. I don't think the latest letter will have any significant impact because many AI researchers, companies, and governments will ignore it completely. The proposal frequently uses the word "we" and refers to "us," the humans. But as I have pointed out many times in the past, there is no "we" that everyone can identify with. Ask 10 different people, and you will hear 10 different opinions about what is "good." Some of those opinions will be completely incompatible with each other. Don't forget the enormous amount of conflict between the many people. The letter also says, "If such a pause cannot be quickly put in place, governments should intervene and impose a moratorium." The problem is that different governments have ALSO different opinions about what is good for them and for others. Great Power A will say, if we don't do it, Great Power B will, perhaps secretly, and gain an advantage over us. The same is true for Great Powers C and D. Jones: Everyone acknowledges this fear surrounding current generative AI technology. Moreover, the existential threat of this technology has been publicly acknowledged by Sam Altman, CEO of OpenAI himself, calling for AI regulation. From your perspective, is there an existential threat? Schmidhuber: It is true that AI can be weaponized, and I have no doubt that there will be all kinds of AI arms races, but AI does not introduce a new quality of existential threat. The threat coming from AI weapons seems to pale in comparison to the much older threat from nuclear hydrogen bombs that don’t need AI at all. We should be much more afraid of half-century-old tech in the form of H-bomb rockets. The Tsar Bomba of 1961 had almost 15 times more destructive power than all weapons of WW-II combined. Despite the dramatic nuclear disarmament since the 1980s, there are still more than enough nuclear warheads to wipe out human civilization within two hours, without any AI I’m much more worried about that old existential threat than the rather harmless AI weapons. Jones: I realize that while you compare AI to the threat of nuclear bombs, there is a current danger that a current technology can be put in the hands of humans and enable them to “eventually” exact further harms to individuals of group in a very precise way, like targeted drone attacks. You are giving people a toolset that they've never had before, enabling bad actors, as some have pointed out, to be able to do a lot more than previously because they didn't have this technology. Schmidhuber: Now, all that sounds horrible in principle, but our existing laws are sufficient to deal with these new types of weapons enabled by AI. If you kill someone with a gun, you will go to jail. Same if you kill someone with one of these drones. Law enforcement will get better at understanding new threats and new weapons and will respond with better technology to combat these threats. Enabling drones to target persons from a distance in a way that requires some tracking and some intelligence to perform, which has traditionally been performed by skilled humans, to me, it seems is just an improved version of a traditional weapon, like a gun, which is, you know, a little bit smarter than the old guns. But, in principle, all of that is not a new development. For many centuries, we have had the evolution of better weaponry and deadlier poisons and so on, and law enforcement has evolved their policies to react to these threats over time. So, it's not that we suddenly have a new quality of existential threat and it's much more worrisome than what we have had for about six decades. A large nuclear warhead doesn’t need fancy face recognition to kill an individual. No, it simply wipes out an entire city with ten million inhabitants. Jones: The existential threat that’s implied is the extent to which humans have control over this technology. We see some early cases of opportunism which, as you say, tends to get more media attention than positive breakthroughs. But you’re implying that this will all balance out? Schmidhuber: Historically, we have a long tradition of technological breakthroughs that led to advancements in weapons for the purpose of defense but also for protection. From sticks, to rocks, to axes to gunpowder to cannons to rockets… and now to drones… this has had a drastic influence on human history but what has been consistent throughout history is that those who are using technology to achieve their own ends are themselves, facing the same technology because the opposing side is learning to use it against them. And that's what has been repeated in thousands of years of human history and it will continue. I don't see the new AI arms race as something that is remotely as existential a threat as the good old nuclear warheads. You said something important, in that some people prefer to talk about the downsides rather than the benefits of this technology, but that's misleading, because 95% of all AI research and AI development is about making people happier and advancing human life and health. Jones: Let’s touch on some of those beneficial advances in AI research that have been able to radically change present day methods and achieve breakthroughs. Schmidhuber: All right! For example, eleven years ago, our team with my postdoc Dan Ciresan was the first to win a medical imaging competition through deep learning. We analyzed female breast cells with the objective to determine harmless cells vs. those in the pre-cancer stage. Typically, a trained oncologist needs a long time to make these determinations. Our team, who knew nothing about cancer, were able to train an artificial neural network, which was totally dumb in the beginning, on lots of this kind of data. It was able to outperform all the other methods. Today, this is being used not only for breast cancer, but also for radiology and detecting plaque in arteries, and many other things. Some of the neural networks that we have developed in the last 3 decades are now prevalent across thousands of healthcare applications, detecting Diabetes and Covid-19 and what not. This will eventually permeate across all healthcare. The good consequences of this type of AI are much more important than the click-bait new ways of conducting crimes with AI. Jones: Adoption is a product of reinforced outcomes. The massive scale of adoption either leads us to believe that people have been led astray, or conversely, technology is having a positive effect on people’s lives. Schmidhuber: The latter is the likely case. There's intense commercial pressure towards good AI rather than bad AI because companies want to sell you something, and you are going to buy only stuff you think is going to be good for you. So already just through this simple, commercial pressure, you have a tremendous bias towards good AI rather than bad AI. However, doomsday scenarios like in Schwarzenegger movies grab more attention than documentaries on AI that improve people’s lives. Jones: I would argue that people are drawn to good stories – narratives that contain an adversary and struggle, but in the end, have happy endings. And this is consistent with your comment on human nature and how history, despite its tendency for violence and destruction of humanity, somehow tends to correct itself. Let’s take the example of a technology, which you are aware – GANs – General Adversarial Networks, which today has been used in applications for fake news and disinformation. In actuality, the purpose in the invention of GANs was far from what it is used for today. Schmidhuber: Yes, the name GANs was created in 2014 but we had the basic principle already in the early 1990s. More than 30 years ago, I called it artificial curiosity. It's a very simple way of injecting creativity into a little two network system. This creative AI is not just trying to slavishly imitate humans. Rather, it’s inventing its own goals. Let me explain: You have two networks. One network is producing outputs that could be anything, any action. Then the second network is looking at these actions and it’s trying to predict the consequences of these actions. An action could move a robot, then something happens, and the other network is just trying to predict what will happen. Now we can implement artificial curiosity by reducing the prediction error of the second network, which, at the same time, is the reward of the first network. The first network wants to maximize its reward and so it will invent actions that will lead to situations that will surprise the second network, which it has not yet learned to predict well. In the case where the outputs are fake images, the first network will try to generate images that are good enough to fool the second network, which will attempt to predict the reaction of the environment: fake or real image, and it will try to become better at it. The first network will continue to also improve at generating images whose type the second network will not be able to predict. So, they fight each other. The 2nd network will continue to reduce its prediction error, while the 1st network will attempt to maximize it. Through this zero-sum game the first network gets better and better at producing these convincing fake outputs which look almost realistic. So, once you have an interesting set of images by Vincent Van Gogh, you can generate new images that leverage his style, without the original artist having ever produced the artwork himself. Jones: I see how the Van Gogh example can be applied in an education setting and there are countless examples of artists mimicking styles from famous painters but image generation from this instance that can happen within seconds is quite another feat. And you know this is how GANs has been used. What’s more prevalent today is a socialized enablement of generating images or information to intentionally fool people. It also surfaces new harms that deal with the threat to intellectual property and copyright, where laws have yet to account for. And from your perspective this was not the intention when the model was conceived. What was your motivation in your early conception of what is now GANs? Schmidhuber: My old motivation for GANs was actually very important and it was not to create deepfakes or fake news but to enable AIs to be curious and invent their own goals, to make them explore their environment and make them creative. Suppose you have a robot that executes one action, then something happens, then it executes another action, and so on, because it wants to achieve certain goals in the environment. For example, when the battery is low, this will trigger “pain” through hunger sensors, so it wants to go to the charging station, without running into obstacles, which will trigger other pain sensors. It will seek to minimize pain (encoded through numbers). Now the robot has a friend, the second network, which is a world model ––it’s a prediction machine that learns to predict the consequences of the robot’s actions. Once the robot has a good model of the world, it can use it for planning. It can be used as a simulation of the real world. And then it can determine what is a good action sequence. If the robot imagines this sequence of actions, the model will predict a lot of pain, which it wants to avoid. If it plays this alternative action sequence in its mental model of the world, then it will predict a rewarding situation where it’s going to sit on the charging station and its battery is going to load again. So, it'll prefer to execute the latter action sequence. In the beginning, however, the model of the world knows nothing, so how can we motivate the first network to generate experiments that lead to data that helps the world model learn something it didn’t already know? That’s what artificial curiosity is about. The dueling two network systems effectively explore uncharted environments by creating experiments so that over time the curious AI gets a better sense of how the environment works. This can be applied to all kinds of environments, and has medical applications. Jones: Let’s talk about the future. You have said, “Traditional humans won’t play a significant role in spreading intelligence across the universe.” Schmidhuber: Let’s first conceptually separate two types of AIs. The first type of AI are tools directed by humans. They are trained to do specific things like accurately detect diabetes or heart disease and prevent attacks before they happen. In these cases, the goal is coming from the human. More interesting AIs are setting their own goals. They are inventing their own experiments and learning from them. Their horizons expand and eventually they become more and more general problem solvers in the real world. They are not controlled by their parents, but much of what they learn is through self-invented experiments. A robot, for example, is rotating a toy, and as it is doing this, the video coming in through the camera eyes, changes over time and it begins to learn how this video changes and learns how the 3D nature of the toy generates certain videos if you rotate it a certain way, and eventually, how gravity works, and how the physics of the world works. Like a little scientist! And I have predicted for decades that future scaled-up versions of such AI scientists will want to further expand their horizons, and eventually go where most of the physical resources are, to build more and bigger AIs. And of course, almost all of these resources are far away from earth out there in space, which is hostile to humans but friendly to appropriately designed AI-controlled robots and self-replicating robot factories. So here we are not talking any longer about our tiny biosphere; no, we are talking about the much bigger rest of the universe. Within a few tens of billions of years, curious self-improving AIs will colonize the visible cosmos in a way that’s infeasible for humans. Those who don’t won’t have an impact. Sounds like science fiction, but since the 1970s I have been unable to see a plausible alternative to this scenario, except for a global catastrophe such as an all-out nuclear war that stops this development before it takes off. Jones: How long have these AIs, which can set their own goals — how long have they existed? To what extent can they be independent of human interaction? Schmidhuber: Neural networks like that have existed for over 30 years. My first simple adversarial neural network system of this kind is the one from 1990 described above. You don’t need a teacher there; it's just a little agent running around in the world and trying to invent new experiments that surprise its own prediction machine. Once it has figured out certain parts of the world, the agent will become bored and will move on to more exciting experiments. The simple 1990 systems I mentioned have certain limitations, but in the past three decades, we have also built more sophisticated systems that are setting their own goals and such systems I think will be essential for achieving true intelligence. If you are only imitating humans, you will never go beyond them. So, you really must give AIs the freedom to explore previously unexplored regions of the world in a way that no human is really predefining. Jones: Where is this being done today? Schmidhuber: Variants of neural network-based artificial curiosity are used today for agents that learn to play video games in a human-competitive way. We have also started to use them for automatic design of experiments in fields such as materials science. I bet many other fields will be affected by it: chemistry, biology, drug design, you name it. However, at least for now, these artificial scientists, as I like to call them, cannot yet compete with human scientists. I don’t think it’s going to stay this way but, at the moment, it’s still the case. Sure, AI has made a lot of progress. Since 1997, there have been superhuman chess players, and since 2011, through the DanNet of my team, there have been superhuman visual pattern recognizers. But there are other things where humans, at the moment at least, are much better, in particular, science itself. In the lab we have many first examples of self-directed artificial scientists, but they are not yet convincing enough to appear on the radar screen of the public space, which is currently much more fascinated with simpler systems that just imitate humans and write texts based on previously seen human-written documents. Jones: You speak of these numerous instances dating back 30 years of these lab experiments where these self-driven agents are deciding and learning and moving on once they’ve learned. And I assume that that rate of learning becomes even faster over time. What kind of timeframe are we talking about when this eventually is taken outside of the lab and embedded into society? Schmidhuber: This could still take months or even years :-) Anyway, in the not-too-distant future, we will probably see artificial scientists who are good at devising experiments that allow them to discover new, previously unknown physical laws. As always, we are going to profit from the old trend that has held at least since 1941: every decade compute is getting 100 times cheaper. Jones: How does this trend affect modern AI such as ChatGPT? Schmidhuber: Perhaps you know that all the recent famous AI applications such as ChatGPT and similar models are largely based on principles of artificial neural networks invented in the previous millennium. The main reason why they works so well now is the incredible acceleration of compute per dollar. ChatGPT is driven by a neural network called “Transformer” described in 2017 by Google. I am happy about that because a quarter century earlier in 1991 I had a particular Transformer variant which is now called the “Transformer with linearized self-attention”. Back then, not much could be done with it, because the compute cost was a million times higher than today. But today, one can train such models on half the internet and achieve much more interesting results. Jones: And for how long will this acceleration continue? Schmidhuber: There's no reason to believe that in the next 30 years, we won't have another factor of 1 million and that's going to be really significant. In the near future, for the first time we will have many not-so expensive devices that can compute as much as a human brain. The physical limits of computation, however, are much further out so even if the trend of a factor of 100 every decade continues, the physical limits (of 1051 elementary instructions per second and kilogram of matter) won’t be hit until, say, the mid-next century. Even in our current century, however, we’ll probably have many machines that compute more than all 10 billion human brains collectively and you can imagine, everything will change then! Jones: That is the big question. Is everything going to change? If so, what do you say to the next generation of leaders, currently coming out of college and university. So much of this change is already impacting how they study, how they will work, or how the future of work and livelihood is defined. What is their purpose and how do we change our systems so they will adapt to this new version of intelligence? Schmidhuber: For decades, people have asked me questions like that, because you know what I'm saying now, I have basically said since the 1970s, it’s just that today, people are paying more attention because, back then, they thought this was science fiction. They didn't think that I would ever come close to achieving my crazy life goal of building a machine that learns to become smarter than myself such that I can retire. But now many have changed their minds and think it's conceivable. And now I have two daughters, 23 and 25. People ask me: what do I tell them? They know that Daddy always said, “It seems likely that within your lifetimes, you will have new types of intelligence that are probably going to be superior in many ways, and probably all kinds of interesting ways.” How should they prepare for that? And I kept telling them the obvious: Learn how to learn new things! It's not like in the previous millennium where within 20 years someone learned to be a useful member of society, and then took a job for 40 years and performed in this job until she received her pension. Now things are changing much faster and we must learn continuously just to keep up. I also told my girls that no matter how smart AIs are going to get, learn at least the basics of math and physics, because that’s the essence of our universe, and anybody who understands this will have an advantage, and learn all kinds of new things more easily. I also told them that social skills will remain important, because most future jobs for humans will continue to involve interactions with other humans, but I couldn’t teach them anything about that; they know much more about social skills than I do. You touched on the big philosophical question about people’s purpose. Can this be answered without answering the even grander question: What’s the purpose of the entire universe? We don’t know. But what’s happening right now might be connected to the unknown answer. Don’t think of humans as the crown of creation. Instead view human civilization as part of a much grander scheme, an important step (but not the last one) on the path of the universe from very simple initial conditions towards more and more unfathomable complexity. Now it seems ready to take its next step, a step comparable to the invention of life itself over 3.5 billion years ago. Alas, don’t worry, in the end, all will be good! Jones: Let’s get back to this transformation happening right now with OpenAI. There are many questioning the efficacy and accuracy of ChatGPT, and are concerned its release has been premature. In light of the rampant adoption, educators have banned its use over concerns of plagiarism and how it stifles individual development. Should large language models like ChatGPT be used in school? Schmidhuber: When the calculator was first introduced, instructors forbade students from using it in school. Today, the consensus is that kids should learn the basic methods of arithmetic, but they should also learn to use the “artificial multipliers” aka calculators, even in exams, because laziness and efficiency is a hallmark of intelligence. Any intelligent being wants to minimize its efforts to achieve things. And that's the reason why we have tools, and why our kids are learning to use these tools. The first stone tools were invented maybe 3.5 million years ago; tools just have become more sophisticated over time. In fact, humans have changed in response to the properties of their tools. Our anatomical evolution was shaped by tools such as spears and fire. So, it's going to continue this way. And there is no permanent way of preventing large language models from being used in school. Jones: And when our children, your children graduate, what does their future work look like? Schmidhuber: A single human trying to predict details of how 10 billion people and their machines will evolve in the future is like a single neuron in my brain trying to predict what the entire brain and its tens of billions of neurons will do next year. 40 years ago, before the WWW was created at CERN in Switzerland, who would have predicted all those young people making money as YouTube video bloggers? Nevertheless, let’s make a few limited job-related observations. For a long time, people have thought that desktop jobs may require more intelligence than skills trade or handicraft professions. But now, it turns out that it's much easier to replace certain aspects of desktop jobs than replacing a carpenter, for example. Because everything that works well in AI is happening behind the screen currently, but not so much in the physical world. There are now artificial systems that can read lots of documents and then make really nice summaries of these documents. That is a desktop job. Or you give them a description of an illustration that you want to have for your article and pretty good illustrations are being generated that may need some minimal fine-tuning. But you know, all these desktop jobs are much easier to facilitate than the real tough jobs in the physical world. And it's interesting that the things people thought required intelligence, like playing chess, or writing or summarizing documents, are much easier for machines than they thought. But for things like playing football or soccer, there is no physical robot that can remotely compete with the abilities of a little boy with these skills. So, AI in the physical world, interestingly, is much harder than AI behind the screen in virtual worlds. And it's really exciting, in my opinion, to see that jobs such as plumbers are much more challenging than playing chess or writing another tabloid story. Jones: The way data has been collected in these large language models does not guarantee personal information has not been excluded. Current consent laws already are outdated when it comes to these large language models (LLM). The concern, rightly so, is increasing surveillance and loss of privacy. What is your view on this? Schmidhuber: As I have indicated earlier: are surveillance and loss of privacy inevitable consequences of increasingly complex societies? Super-organisms such as cities and states and companies consist of numerous people, just like people consist of numerous cells. These cells enjoy little privacy. They are constantly monitored by specialized "police cells" and "border guard cells": Are you a cancer cell? Are you an external intruder, a pathogen? Individual cells sacrifice their freedom for the benefits of being part of a multicellular organism. Similarly, for super-organisms such as nations. Over 5000 years ago, writing enabled recorded history and thus became its inaugural and most important invention. Its initial purpose, however, was to facilitate surveillance, to track citizens and their tax payments. The more complex a super-organism, the more comprehensive its collection of information about its constituents. 200 years ago, at least, the parish priest in each village knew everything about all the village people, even about those who did not confess, because they appeared in the confessions of others. Also, everyone soon knew about the stranger who had entered the village, because some occasionally peered out of the window, and what they saw got around. Such control mechanisms were temporarily lost through anonymization in rapidly growing cities but are now returning with the help of new surveillance devices such as smartphones as part of digital nervous systems that tell companies and governments a lot about billions of users. Cameras and drones etc. are becoming increasingly tinier and more ubiquitous. More effective recognition of faces and other detection technology are becoming cheaper and cheaper, and many will use it to identify others anywhere on earth; the big wide world will not offer any more privacy than the local village. Is this good or bad? Some nations may find it easier than others to justify more complex kinds of super-organisms at the expense of the privacy rights of their constituents. Jones: So, there is no way to stop or change this process of collection, or how it continuously informs decisions over time? How do you see governance and rules responding to this, especially amid Italy’s ban on ChatGPT following suspected user data breach and the more recent news about the Meta’s record $1.3billion fine in the company’s handling of user information? Schmidhuber: Data collection has benefits and drawbacks, such as the loss of privacy. How to balance those? I have argued for addressing this through data ownership in data markets. If it is true that data is the new oil, then it should have a price, just like oil. At the moment, the major surveillance platforms such as Meta do not offer users any money for their data and the transitive loss of privacy. In the future, however, we will likely see attempts at creating efficient data markets to figure out the data's true financial value through the interplay between supply and demand. Even some of the sensitive medical data should not be priced by governmental regulators but by patients (and healthy persons) who own it and who may sell or license parts thereof as micro-entrepreneurs in a healthcare data market. Following a previous interview, I gave for one of the largest re-insurance companies , let's look at the different participants in such a data market: patients, hospitals, data companies. (1) Patients with a rare form of cancer can offer more valuable data than patients with a very common form of cancer. (2) Hospitals and their machines are needed to extract the data, e.g., through magnet spin tomography, radiology, evaluations through human doctors, and so on. (3) Companies such as Siemens, Google or IBM would like to buy annotated data to make better artificial neural networks that learn to predict pathologies and diseases and the consequences of therapies. Now the market’s invisible hand will decide about the data’s price through the interplay between demand and supply. On the demand side, you will have several companies offering something for the data, maybe through an app on the smartphone (a bit like a stock market app). On the supply side, each patient in this market should be able to profit from high prices for rare valuable types of data. Likewise, competing data extractors such as hospitals will profit from gaining recognition and trust for extracting data well at a reasonable price. The market will make the whole system efficient through incentives for all who are doing a good job. Soon there will be a flourishing ecosystem of commercial data market advisors and what not, just like the ecosystem surrounding the traditional stock market. The value of the data won’t be determined by governments or ethics committees, but by those who own the data and decide by themselves which parts thereof they want to license to others under certain conditions. At first glance, a market-based system seems to be detrimental to the interest of certain monopolistic companies, as they would have to pay for the data - some would prefer free data and keep their monopoly. However, since every healthy and sick person in the market would suddenly have an incentive to collect and share their data under self-chosen anonymity conditions, there will soon be many more useful data to evaluate all kinds of treatments. On average, people will live longer and healthier, and many companies and the entire healthcare system will benefit. Jones: Finally, what is your view on open source versus the private companies like Google and OpenAI? Is there a danger to supporting these private companies’ large language models versus trying to keep these models open source and transparent, very much like what LAION is doing? Schmidhuber: I signed this open letter by LAION because I strongly favor the open-source movement. And I think it's also something that is going to challenge whatever big tech dominance there might be at the moment. Sure, the best models today are run by big companies with huge budgets for computers, but the exciting fact is that open-source models are not so far behind, some people say maybe six to eight months only. Of course, the private company models are all based on stuff that was created in academia, often in little labs without so much funding, which publish without patenting their results and open source their code and others take it and improved it. Big tech has profited tremendously from academia; their main achievement being that they have scaled up everything greatly, sometimes even failing to credit the original inventors. So, it's very interesting to see that as soon as some big company comes up with a new scaled-up model, lots of students out there are competing, or collaborating, with each other, trying to come up with equal or better performance on smaller networks and smaller machines. And since they are open sourcing, the next guy can have another great idea to improve it, so now there’s tremendous competition also for the big companies. Because of that, and since AI is still getting exponentially cheaper all the time, I don't believe that big tech companies will dominate in the long run. They find it very hard to compete with the enormous open-source movement. As long as you can encourage the open-source community, I think you shouldn't worry too much. Now, of course, you might say if everything is open source, then the bad actors also will more easily have access to these AI tools. And there's truth to that. But as always since the invention of controlled fire, it was good that knowledge about how technology works quickly became public such that everybody could use it. And then, against any bad actor, there's almost immediately a counter actor trying to nullify his efforts. You see, I still believe in our old motto "AI∀" or "AI For All." Jones: Thank you, Juergen for sharing your perspective on this amazing time in history. It’s clear that with new technology, the enormous potential can be matched by disparate and troubling risks which we’ve yet to solve, and even those we have yet to identify. If we are to dispel the fear of a sentient system for which we have no control, humans, alone need to take steps for more responsible development and collaboration to ensure AI technology is used to ultimately benefit society. Humanity will be judged by what we do next.

[D] Why is the AI Hype Absolutely Bonkers
reddit
LLM Vibe Score0
Human Vibe Score1
good_riceThis week

[D] Why is the AI Hype Absolutely Bonkers

Edit 2: Both the repo and the post were deleted. Redacting identifying information as the author has appeared to make rectifications, and it’d be pretty damaging if this is what came up when googling their name / GitHub (hopefully they’ve learned a career lesson and can move on). TL;DR: A PhD candidate claimed to have achieved 97% accuracy for coronavirus from chest x-rays. Their post gathered thousands of reactions, and the candidate was quick to recruit branding, marketing, frontend, and backend developers for the project. Heaps of praise all around. He listed himself as a Director of XXXX (redacted), the new name for his project. The accuracy was based on a training dataset of ~30 images of lesion / healthy lungs, sharing of data between test / train / validation, and code to train ResNet50 from a PyTorch tutorial. Nonetheless, thousands of reactions and praise from the “AI | Data Science | Entrepreneur” community. Original Post: I saw this post circulating on LinkedIn: https://www.linkedin.com/posts/activity-6645711949554425856-9Dhm Here, a PhD candidate claims to achieve great performance with “ARTIFICIAL INTELLIGENCE” to predict coronavirus, asks for more help, and garners tens of thousands of views. The repo housing this ARTIFICIAL INTELLIGENCE solution already has a backend, front end, branding, a README translated in 6 languages, and a call to spread the word for this wonderful technology. Surely, I thought, this researcher has some great and novel tech for all of this hype? I mean dear god, we have branding, and the author has listed himself as the founder of an organization based on this project. Anything with this much attention, with dozens of “AI | Data Scientist | Entrepreneur” members of LinkedIn praising it, must have some great merit, right? Lo and behold, we have ResNet50, from torchvision.models import resnet50, with its linear layer replaced. We have a training dataset of 30 images. This should’ve taken at MAX 3 hours to put together - 1 hour for following a tutorial, and 2 for obfuscating the training with unnecessary code. I genuinely don’t know what to think other than this is bonkers. I hope I’m wrong, and there’s some secret model this author is hiding? If so, I’ll delete this post, but I looked through the repo and (REPO link redacted) that’s all I could find. I’m at a loss for thoughts. Can someone explain why this stuff trends on LinkedIn, gets thousands of views and reactions, and gets loads of praise from “expert data scientists”? It’s almost offensive to people who are like ... actually working to treat coronavirus and develop real solutions. It also seriously turns me off from pursuing an MS in CV as opposed to CS. Edit: It turns out there were duplicate images between test / val / training, as if ResNet50 on 30 images wasn’t enough already. He’s also posted an update signed as “Director of XXXX (redacted)”. This seems like a straight up sleazy way to capitalize on the pandemic by advertising himself to be the head of a made up organization, pulling resources away from real biomedical researchers.

[N] New Trends to Power Faster Artificial Intelligence and Machine Learning Adoption?
reddit
LLM Vibe Score0
Human Vibe Score1
EsotericaCapitalThis week

[N] New Trends to Power Faster Artificial Intelligence and Machine Learning Adoption?

In 2012, Google X lab created a neural network that can identify cats. Since then, technology companies have been increasingly adopting AI/ML on a large scale to build better applications and services for consumers (ToC). On the other hand, AI/ML's adoption on the enterprises' side (ToB) has yet to see the same growth trajectory due to the costs and complexities in both hardware and software. However, Since 2020, we started noticing three emerging tech trends that can help accelerate enterprises' adoption of AI/ML. Breakthrough in semiconductors: In 2020, Nvidia debut the concept of "Data Processing Unit," a new class of programmable processors that combine high-performance CPU with SmartNiC (network interface controller). Data centers can deploy DPUs to optimize computing offload and frees up CPUs to focus on intended tasks, such as machine learning. DPUs help resolve a significant bottleneck for ML training, where models, sometimes with billions of parameters, are way too big for traditional CPUs and GPUs to handle. Other leading semiconductor players, such as Marvell and Xilinx, follow suit with their in-house or partner-designed DPUs. Industry analysts have forecasted that the market size for DPUs in data centers alone could reach $50 billion by 2025. ​ https://preview.redd.it/l436muluhnn61.png?width=1430&format=png&auto=webp&s=ba8d1298056ea31bddd25f1596ff64b7e107580a Breakthrough in software: we also saw significant progress of "Conversational AI," a new form of AI that can understand and speak languages with human-like accuracy, in 2020. Conversational AI allows two-way interactions and provides a much better user experience than traditional AI-powered Chatbot, mostly a one-way response system. The secret of conversational AI is its ability to handle lots of human conversation variance. Developers have designed innovative algorithms such as "Switch transformers" and "Sparse training" to enable models to handle vast amounts of data. The size of conversational AI training models is enormous and keeps expanding. For example, in February 2021, Google Brain announced a model with 1.6 trillion parameters, nine times the size of the famous Open AI GPT-3 (175 billion parameters) unveiled in July 2020. GPT-3 is 100+ times bigger than GPT-2 introduced in 2019. ​ https://preview.redd.it/oajpi2yvhnn61.png?width=1430&format=png&auto=webp&s=1482913a98e17ddc1d62cc79864598d4012ad6f7 Cloud giants are expanding machine-learning platforms for developers. Andy Jassy famously said that "AI is shifting from a niche experiment inside technical departments to becoming more mainstream in business processes." in the 2020 AWS reInvent. During the conference, AWS rolled out many AI products across the technology stack, including AI chips (AWS Trainium), database (Aurora Machine Learning), and vertical solutions (Amazon Healthlake), etc. However, the most significant development is the drastic expansion of "Amazon SageMaker," one of the largest cloud machine-learning platforms. SageMaker has been offering new features to make it easier for developers to automate machine learning workflow. Microsoft Azure and Google Cloud are also growing their ML developer platforms. ​ https://preview.redd.it/z9wf2o8xhnn61.png?width=1430&format=png&auto=webp&s=9f607acfe8f0dbf36fb9b472f3cb40b80f13879e Witnessing these breakthroughs in semiconductor and software, coupled with cloud giants' effort to democratize AI, we see a coming inflection point of accelerated AI adoption in both ToC and ToB markets. So how do we benefit from this megatrend? In semiconductors, we like companies with DPUs exposure. In AI development and processing, we favor multi-cloud AI platforms such as Databricks. In enterprise software, we believe there will be a strong wave of new AI-based enterprise applications that can be creative and efficient in solving real-world problems.

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.
reddit
LLM Vibe Score0
Human Vibe Score0.765
hardmaruThis week

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.

Schmidhuber interview expressing his views on the future of AI and AGI. Original source. I think the interview is of interest to r/MachineLearning, and presents an alternate view, compared to other influential leaders in AI. Juergen Schmidhuber, Renowned 'Father Of Modern AI,' Says His Life’s Work Won't Lead To Dystopia May 23, 2023. Contributed by Hessie Jones. Amid the growing concern about the impact of more advanced artificial intelligence (AI) technologies on society, there are many in the technology community who fear the implications of the advancements in Generative AI if they go unchecked. Dr. Juergen Schmidhuber, a renowned scientist, artificial intelligence researcher and widely regarded as one of the pioneers in the field, is more optimistic. He declares that many of those who suddenly warn against the dangers of AI are just seeking publicity, exploiting the media’s obsession with killer robots which has attracted more attention than “good AI” for healthcare etc. The potential to revolutionize various industries and improve our lives is clear, as are the equal dangers if bad actors leverage the technology for personal gain. Are we headed towards a dystopian future, or is there reason to be optimistic? I had a chance to sit down with Dr. Juergen Schmidhuber to understand his perspective on this seemingly fast-moving AI-train that will leap us into the future. As a teenager in the 1970s, Juergen Schmidhuber became fascinated with the idea of creating intelligent machines that could learn and improve on their own, becoming smarter than himself within his lifetime. This would ultimately lead to his groundbreaking work in the field of deep learning. In the 1980s, he studied computer science at the Technical University of Munich (TUM), where he earned his diploma in 1987. His thesis was on the ultimate self-improving machines that, not only, learn through some pre-wired human-designed learning algorithm, but also learn and improve the learning algorithm itself. Decades later, this became a hot topic. He also received his Ph.D. at TUM in 1991 for work that laid some of the foundations of modern AI. Schmidhuber is best known for his contributions to the development of recurrent neural networks (RNNs), the most powerful type of artificial neural network that can process sequential data such as speech and natural language. With his students Sepp Hochreiter, Felix Gers, Alex Graves, Daan Wierstra, and others, he published architectures and training algorithms for the long short-term memory (LSTM), a type of RNN that is widely used in natural language processing, speech recognition, video games, robotics, and other applications. LSTM has become the most cited neural network of the 20th century, and Business Week called it "arguably the most commercial AI achievement." Throughout his career, Schmidhuber has received various awards and accolades for his groundbreaking work. In 2013, he was awarded the Helmholtz Prize, which recognizes significant contributions to the field of machine learning. In 2016, he was awarded the IEEE Neural Network Pioneer Award for "pioneering contributions to deep learning and neural networks." The media have often called him the “father of modern AI,” because the most cited neural networks all build on his lab’s work. He is quick to point out, however, that AI history goes back centuries. Despite his many accomplishments, at the age of 60, he feels mounting time pressure towards building an Artificial General Intelligence within his lifetime and remains committed to pushing the boundaries of AI research and development. He is currently director of the KAUST AI Initiative, scientific director of the Swiss AI Lab IDSIA, and co-founder and chief scientist of AI company NNAISENSE, whose motto is "AI∀" which is a math-inspired way of saying "AI For All." He continues to work on cutting-edge AI technologies and applications to improve human health and extend human lives and make lives easier for everyone. The following interview has been edited for clarity. Jones: Thank you Juergen for joining me. You have signed letters warning about AI weapons. But you didn't sign the recent publication, "Pause Gigantic AI Experiments: An Open Letter"? Is there a reason? Schmidhuber: Thank you Hessie. Glad to speak with you. I have realized that many of those who warn in public against the dangers of AI are just seeking publicity. I don't think the latest letter will have any significant impact because many AI researchers, companies, and governments will ignore it completely. The proposal frequently uses the word "we" and refers to "us," the humans. But as I have pointed out many times in the past, there is no "we" that everyone can identify with. Ask 10 different people, and you will hear 10 different opinions about what is "good." Some of those opinions will be completely incompatible with each other. Don't forget the enormous amount of conflict between the many people. The letter also says, "If such a pause cannot be quickly put in place, governments should intervene and impose a moratorium." The problem is that different governments have ALSO different opinions about what is good for them and for others. Great Power A will say, if we don't do it, Great Power B will, perhaps secretly, and gain an advantage over us. The same is true for Great Powers C and D. Jones: Everyone acknowledges this fear surrounding current generative AI technology. Moreover, the existential threat of this technology has been publicly acknowledged by Sam Altman, CEO of OpenAI himself, calling for AI regulation. From your perspective, is there an existential threat? Schmidhuber: It is true that AI can be weaponized, and I have no doubt that there will be all kinds of AI arms races, but AI does not introduce a new quality of existential threat. The threat coming from AI weapons seems to pale in comparison to the much older threat from nuclear hydrogen bombs that don’t need AI at all. We should be much more afraid of half-century-old tech in the form of H-bomb rockets. The Tsar Bomba of 1961 had almost 15 times more destructive power than all weapons of WW-II combined. Despite the dramatic nuclear disarmament since the 1980s, there are still more than enough nuclear warheads to wipe out human civilization within two hours, without any AI I’m much more worried about that old existential threat than the rather harmless AI weapons. Jones: I realize that while you compare AI to the threat of nuclear bombs, there is a current danger that a current technology can be put in the hands of humans and enable them to “eventually” exact further harms to individuals of group in a very precise way, like targeted drone attacks. You are giving people a toolset that they've never had before, enabling bad actors, as some have pointed out, to be able to do a lot more than previously because they didn't have this technology. Schmidhuber: Now, all that sounds horrible in principle, but our existing laws are sufficient to deal with these new types of weapons enabled by AI. If you kill someone with a gun, you will go to jail. Same if you kill someone with one of these drones. Law enforcement will get better at understanding new threats and new weapons and will respond with better technology to combat these threats. Enabling drones to target persons from a distance in a way that requires some tracking and some intelligence to perform, which has traditionally been performed by skilled humans, to me, it seems is just an improved version of a traditional weapon, like a gun, which is, you know, a little bit smarter than the old guns. But, in principle, all of that is not a new development. For many centuries, we have had the evolution of better weaponry and deadlier poisons and so on, and law enforcement has evolved their policies to react to these threats over time. So, it's not that we suddenly have a new quality of existential threat and it's much more worrisome than what we have had for about six decades. A large nuclear warhead doesn’t need fancy face recognition to kill an individual. No, it simply wipes out an entire city with ten million inhabitants. Jones: The existential threat that’s implied is the extent to which humans have control over this technology. We see some early cases of opportunism which, as you say, tends to get more media attention than positive breakthroughs. But you’re implying that this will all balance out? Schmidhuber: Historically, we have a long tradition of technological breakthroughs that led to advancements in weapons for the purpose of defense but also for protection. From sticks, to rocks, to axes to gunpowder to cannons to rockets… and now to drones… this has had a drastic influence on human history but what has been consistent throughout history is that those who are using technology to achieve their own ends are themselves, facing the same technology because the opposing side is learning to use it against them. And that's what has been repeated in thousands of years of human history and it will continue. I don't see the new AI arms race as something that is remotely as existential a threat as the good old nuclear warheads. You said something important, in that some people prefer to talk about the downsides rather than the benefits of this technology, but that's misleading, because 95% of all AI research and AI development is about making people happier and advancing human life and health. Jones: Let’s touch on some of those beneficial advances in AI research that have been able to radically change present day methods and achieve breakthroughs. Schmidhuber: All right! For example, eleven years ago, our team with my postdoc Dan Ciresan was the first to win a medical imaging competition through deep learning. We analyzed female breast cells with the objective to determine harmless cells vs. those in the pre-cancer stage. Typically, a trained oncologist needs a long time to make these determinations. Our team, who knew nothing about cancer, were able to train an artificial neural network, which was totally dumb in the beginning, on lots of this kind of data. It was able to outperform all the other methods. Today, this is being used not only for breast cancer, but also for radiology and detecting plaque in arteries, and many other things. Some of the neural networks that we have developed in the last 3 decades are now prevalent across thousands of healthcare applications, detecting Diabetes and Covid-19 and what not. This will eventually permeate across all healthcare. The good consequences of this type of AI are much more important than the click-bait new ways of conducting crimes with AI. Jones: Adoption is a product of reinforced outcomes. The massive scale of adoption either leads us to believe that people have been led astray, or conversely, technology is having a positive effect on people’s lives. Schmidhuber: The latter is the likely case. There's intense commercial pressure towards good AI rather than bad AI because companies want to sell you something, and you are going to buy only stuff you think is going to be good for you. So already just through this simple, commercial pressure, you have a tremendous bias towards good AI rather than bad AI. However, doomsday scenarios like in Schwarzenegger movies grab more attention than documentaries on AI that improve people’s lives. Jones: I would argue that people are drawn to good stories – narratives that contain an adversary and struggle, but in the end, have happy endings. And this is consistent with your comment on human nature and how history, despite its tendency for violence and destruction of humanity, somehow tends to correct itself. Let’s take the example of a technology, which you are aware – GANs – General Adversarial Networks, which today has been used in applications for fake news and disinformation. In actuality, the purpose in the invention of GANs was far from what it is used for today. Schmidhuber: Yes, the name GANs was created in 2014 but we had the basic principle already in the early 1990s. More than 30 years ago, I called it artificial curiosity. It's a very simple way of injecting creativity into a little two network system. This creative AI is not just trying to slavishly imitate humans. Rather, it’s inventing its own goals. Let me explain: You have two networks. One network is producing outputs that could be anything, any action. Then the second network is looking at these actions and it’s trying to predict the consequences of these actions. An action could move a robot, then something happens, and the other network is just trying to predict what will happen. Now we can implement artificial curiosity by reducing the prediction error of the second network, which, at the same time, is the reward of the first network. The first network wants to maximize its reward and so it will invent actions that will lead to situations that will surprise the second network, which it has not yet learned to predict well. In the case where the outputs are fake images, the first network will try to generate images that are good enough to fool the second network, which will attempt to predict the reaction of the environment: fake or real image, and it will try to become better at it. The first network will continue to also improve at generating images whose type the second network will not be able to predict. So, they fight each other. The 2nd network will continue to reduce its prediction error, while the 1st network will attempt to maximize it. Through this zero-sum game the first network gets better and better at producing these convincing fake outputs which look almost realistic. So, once you have an interesting set of images by Vincent Van Gogh, you can generate new images that leverage his style, without the original artist having ever produced the artwork himself. Jones: I see how the Van Gogh example can be applied in an education setting and there are countless examples of artists mimicking styles from famous painters but image generation from this instance that can happen within seconds is quite another feat. And you know this is how GANs has been used. What’s more prevalent today is a socialized enablement of generating images or information to intentionally fool people. It also surfaces new harms that deal with the threat to intellectual property and copyright, where laws have yet to account for. And from your perspective this was not the intention when the model was conceived. What was your motivation in your early conception of what is now GANs? Schmidhuber: My old motivation for GANs was actually very important and it was not to create deepfakes or fake news but to enable AIs to be curious and invent their own goals, to make them explore their environment and make them creative. Suppose you have a robot that executes one action, then something happens, then it executes another action, and so on, because it wants to achieve certain goals in the environment. For example, when the battery is low, this will trigger “pain” through hunger sensors, so it wants to go to the charging station, without running into obstacles, which will trigger other pain sensors. It will seek to minimize pain (encoded through numbers). Now the robot has a friend, the second network, which is a world model ––it’s a prediction machine that learns to predict the consequences of the robot’s actions. Once the robot has a good model of the world, it can use it for planning. It can be used as a simulation of the real world. And then it can determine what is a good action sequence. If the robot imagines this sequence of actions, the model will predict a lot of pain, which it wants to avoid. If it plays this alternative action sequence in its mental model of the world, then it will predict a rewarding situation where it’s going to sit on the charging station and its battery is going to load again. So, it'll prefer to execute the latter action sequence. In the beginning, however, the model of the world knows nothing, so how can we motivate the first network to generate experiments that lead to data that helps the world model learn something it didn’t already know? That’s what artificial curiosity is about. The dueling two network systems effectively explore uncharted environments by creating experiments so that over time the curious AI gets a better sense of how the environment works. This can be applied to all kinds of environments, and has medical applications. Jones: Let’s talk about the future. You have said, “Traditional humans won’t play a significant role in spreading intelligence across the universe.” Schmidhuber: Let’s first conceptually separate two types of AIs. The first type of AI are tools directed by humans. They are trained to do specific things like accurately detect diabetes or heart disease and prevent attacks before they happen. In these cases, the goal is coming from the human. More interesting AIs are setting their own goals. They are inventing their own experiments and learning from them. Their horizons expand and eventually they become more and more general problem solvers in the real world. They are not controlled by their parents, but much of what they learn is through self-invented experiments. A robot, for example, is rotating a toy, and as it is doing this, the video coming in through the camera eyes, changes over time and it begins to learn how this video changes and learns how the 3D nature of the toy generates certain videos if you rotate it a certain way, and eventually, how gravity works, and how the physics of the world works. Like a little scientist! And I have predicted for decades that future scaled-up versions of such AI scientists will want to further expand their horizons, and eventually go where most of the physical resources are, to build more and bigger AIs. And of course, almost all of these resources are far away from earth out there in space, which is hostile to humans but friendly to appropriately designed AI-controlled robots and self-replicating robot factories. So here we are not talking any longer about our tiny biosphere; no, we are talking about the much bigger rest of the universe. Within a few tens of billions of years, curious self-improving AIs will colonize the visible cosmos in a way that’s infeasible for humans. Those who don’t won’t have an impact. Sounds like science fiction, but since the 1970s I have been unable to see a plausible alternative to this scenario, except for a global catastrophe such as an all-out nuclear war that stops this development before it takes off. Jones: How long have these AIs, which can set their own goals — how long have they existed? To what extent can they be independent of human interaction? Schmidhuber: Neural networks like that have existed for over 30 years. My first simple adversarial neural network system of this kind is the one from 1990 described above. You don’t need a teacher there; it's just a little agent running around in the world and trying to invent new experiments that surprise its own prediction machine. Once it has figured out certain parts of the world, the agent will become bored and will move on to more exciting experiments. The simple 1990 systems I mentioned have certain limitations, but in the past three decades, we have also built more sophisticated systems that are setting their own goals and such systems I think will be essential for achieving true intelligence. If you are only imitating humans, you will never go beyond them. So, you really must give AIs the freedom to explore previously unexplored regions of the world in a way that no human is really predefining. Jones: Where is this being done today? Schmidhuber: Variants of neural network-based artificial curiosity are used today for agents that learn to play video games in a human-competitive way. We have also started to use them for automatic design of experiments in fields such as materials science. I bet many other fields will be affected by it: chemistry, biology, drug design, you name it. However, at least for now, these artificial scientists, as I like to call them, cannot yet compete with human scientists. I don’t think it’s going to stay this way but, at the moment, it’s still the case. Sure, AI has made a lot of progress. Since 1997, there have been superhuman chess players, and since 2011, through the DanNet of my team, there have been superhuman visual pattern recognizers. But there are other things where humans, at the moment at least, are much better, in particular, science itself. In the lab we have many first examples of self-directed artificial scientists, but they are not yet convincing enough to appear on the radar screen of the public space, which is currently much more fascinated with simpler systems that just imitate humans and write texts based on previously seen human-written documents. Jones: You speak of these numerous instances dating back 30 years of these lab experiments where these self-driven agents are deciding and learning and moving on once they’ve learned. And I assume that that rate of learning becomes even faster over time. What kind of timeframe are we talking about when this eventually is taken outside of the lab and embedded into society? Schmidhuber: This could still take months or even years :-) Anyway, in the not-too-distant future, we will probably see artificial scientists who are good at devising experiments that allow them to discover new, previously unknown physical laws. As always, we are going to profit from the old trend that has held at least since 1941: every decade compute is getting 100 times cheaper. Jones: How does this trend affect modern AI such as ChatGPT? Schmidhuber: Perhaps you know that all the recent famous AI applications such as ChatGPT and similar models are largely based on principles of artificial neural networks invented in the previous millennium. The main reason why they works so well now is the incredible acceleration of compute per dollar. ChatGPT is driven by a neural network called “Transformer” described in 2017 by Google. I am happy about that because a quarter century earlier in 1991 I had a particular Transformer variant which is now called the “Transformer with linearized self-attention”. Back then, not much could be done with it, because the compute cost was a million times higher than today. But today, one can train such models on half the internet and achieve much more interesting results. Jones: And for how long will this acceleration continue? Schmidhuber: There's no reason to believe that in the next 30 years, we won't have another factor of 1 million and that's going to be really significant. In the near future, for the first time we will have many not-so expensive devices that can compute as much as a human brain. The physical limits of computation, however, are much further out so even if the trend of a factor of 100 every decade continues, the physical limits (of 1051 elementary instructions per second and kilogram of matter) won’t be hit until, say, the mid-next century. Even in our current century, however, we’ll probably have many machines that compute more than all 10 billion human brains collectively and you can imagine, everything will change then! Jones: That is the big question. Is everything going to change? If so, what do you say to the next generation of leaders, currently coming out of college and university. So much of this change is already impacting how they study, how they will work, or how the future of work and livelihood is defined. What is their purpose and how do we change our systems so they will adapt to this new version of intelligence? Schmidhuber: For decades, people have asked me questions like that, because you know what I'm saying now, I have basically said since the 1970s, it’s just that today, people are paying more attention because, back then, they thought this was science fiction. They didn't think that I would ever come close to achieving my crazy life goal of building a machine that learns to become smarter than myself such that I can retire. But now many have changed their minds and think it's conceivable. And now I have two daughters, 23 and 25. People ask me: what do I tell them? They know that Daddy always said, “It seems likely that within your lifetimes, you will have new types of intelligence that are probably going to be superior in many ways, and probably all kinds of interesting ways.” How should they prepare for that? And I kept telling them the obvious: Learn how to learn new things! It's not like in the previous millennium where within 20 years someone learned to be a useful member of society, and then took a job for 40 years and performed in this job until she received her pension. Now things are changing much faster and we must learn continuously just to keep up. I also told my girls that no matter how smart AIs are going to get, learn at least the basics of math and physics, because that’s the essence of our universe, and anybody who understands this will have an advantage, and learn all kinds of new things more easily. I also told them that social skills will remain important, because most future jobs for humans will continue to involve interactions with other humans, but I couldn’t teach them anything about that; they know much more about social skills than I do. You touched on the big philosophical question about people’s purpose. Can this be answered without answering the even grander question: What’s the purpose of the entire universe? We don’t know. But what’s happening right now might be connected to the unknown answer. Don’t think of humans as the crown of creation. Instead view human civilization as part of a much grander scheme, an important step (but not the last one) on the path of the universe from very simple initial conditions towards more and more unfathomable complexity. Now it seems ready to take its next step, a step comparable to the invention of life itself over 3.5 billion years ago. Alas, don’t worry, in the end, all will be good! Jones: Let’s get back to this transformation happening right now with OpenAI. There are many questioning the efficacy and accuracy of ChatGPT, and are concerned its release has been premature. In light of the rampant adoption, educators have banned its use over concerns of plagiarism and how it stifles individual development. Should large language models like ChatGPT be used in school? Schmidhuber: When the calculator was first introduced, instructors forbade students from using it in school. Today, the consensus is that kids should learn the basic methods of arithmetic, but they should also learn to use the “artificial multipliers” aka calculators, even in exams, because laziness and efficiency is a hallmark of intelligence. Any intelligent being wants to minimize its efforts to achieve things. And that's the reason why we have tools, and why our kids are learning to use these tools. The first stone tools were invented maybe 3.5 million years ago; tools just have become more sophisticated over time. In fact, humans have changed in response to the properties of their tools. Our anatomical evolution was shaped by tools such as spears and fire. So, it's going to continue this way. And there is no permanent way of preventing large language models from being used in school. Jones: And when our children, your children graduate, what does their future work look like? Schmidhuber: A single human trying to predict details of how 10 billion people and their machines will evolve in the future is like a single neuron in my brain trying to predict what the entire brain and its tens of billions of neurons will do next year. 40 years ago, before the WWW was created at CERN in Switzerland, who would have predicted all those young people making money as YouTube video bloggers? Nevertheless, let’s make a few limited job-related observations. For a long time, people have thought that desktop jobs may require more intelligence than skills trade or handicraft professions. But now, it turns out that it's much easier to replace certain aspects of desktop jobs than replacing a carpenter, for example. Because everything that works well in AI is happening behind the screen currently, but not so much in the physical world. There are now artificial systems that can read lots of documents and then make really nice summaries of these documents. That is a desktop job. Or you give them a description of an illustration that you want to have for your article and pretty good illustrations are being generated that may need some minimal fine-tuning. But you know, all these desktop jobs are much easier to facilitate than the real tough jobs in the physical world. And it's interesting that the things people thought required intelligence, like playing chess, or writing or summarizing documents, are much easier for machines than they thought. But for things like playing football or soccer, there is no physical robot that can remotely compete with the abilities of a little boy with these skills. So, AI in the physical world, interestingly, is much harder than AI behind the screen in virtual worlds. And it's really exciting, in my opinion, to see that jobs such as plumbers are much more challenging than playing chess or writing another tabloid story. Jones: The way data has been collected in these large language models does not guarantee personal information has not been excluded. Current consent laws already are outdated when it comes to these large language models (LLM). The concern, rightly so, is increasing surveillance and loss of privacy. What is your view on this? Schmidhuber: As I have indicated earlier: are surveillance and loss of privacy inevitable consequences of increasingly complex societies? Super-organisms such as cities and states and companies consist of numerous people, just like people consist of numerous cells. These cells enjoy little privacy. They are constantly monitored by specialized "police cells" and "border guard cells": Are you a cancer cell? Are you an external intruder, a pathogen? Individual cells sacrifice their freedom for the benefits of being part of a multicellular organism. Similarly, for super-organisms such as nations. Over 5000 years ago, writing enabled recorded history and thus became its inaugural and most important invention. Its initial purpose, however, was to facilitate surveillance, to track citizens and their tax payments. The more complex a super-organism, the more comprehensive its collection of information about its constituents. 200 years ago, at least, the parish priest in each village knew everything about all the village people, even about those who did not confess, because they appeared in the confessions of others. Also, everyone soon knew about the stranger who had entered the village, because some occasionally peered out of the window, and what they saw got around. Such control mechanisms were temporarily lost through anonymization in rapidly growing cities but are now returning with the help of new surveillance devices such as smartphones as part of digital nervous systems that tell companies and governments a lot about billions of users. Cameras and drones etc. are becoming increasingly tinier and more ubiquitous. More effective recognition of faces and other detection technology are becoming cheaper and cheaper, and many will use it to identify others anywhere on earth; the big wide world will not offer any more privacy than the local village. Is this good or bad? Some nations may find it easier than others to justify more complex kinds of super-organisms at the expense of the privacy rights of their constituents. Jones: So, there is no way to stop or change this process of collection, or how it continuously informs decisions over time? How do you see governance and rules responding to this, especially amid Italy’s ban on ChatGPT following suspected user data breach and the more recent news about the Meta’s record $1.3billion fine in the company’s handling of user information? Schmidhuber: Data collection has benefits and drawbacks, such as the loss of privacy. How to balance those? I have argued for addressing this through data ownership in data markets. If it is true that data is the new oil, then it should have a price, just like oil. At the moment, the major surveillance platforms such as Meta do not offer users any money for their data and the transitive loss of privacy. In the future, however, we will likely see attempts at creating efficient data markets to figure out the data's true financial value through the interplay between supply and demand. Even some of the sensitive medical data should not be priced by governmental regulators but by patients (and healthy persons) who own it and who may sell or license parts thereof as micro-entrepreneurs in a healthcare data market. Following a previous interview, I gave for one of the largest re-insurance companies , let's look at the different participants in such a data market: patients, hospitals, data companies. (1) Patients with a rare form of cancer can offer more valuable data than patients with a very common form of cancer. (2) Hospitals and their machines are needed to extract the data, e.g., through magnet spin tomography, radiology, evaluations through human doctors, and so on. (3) Companies such as Siemens, Google or IBM would like to buy annotated data to make better artificial neural networks that learn to predict pathologies and diseases and the consequences of therapies. Now the market’s invisible hand will decide about the data’s price through the interplay between demand and supply. On the demand side, you will have several companies offering something for the data, maybe through an app on the smartphone (a bit like a stock market app). On the supply side, each patient in this market should be able to profit from high prices for rare valuable types of data. Likewise, competing data extractors such as hospitals will profit from gaining recognition and trust for extracting data well at a reasonable price. The market will make the whole system efficient through incentives for all who are doing a good job. Soon there will be a flourishing ecosystem of commercial data market advisors and what not, just like the ecosystem surrounding the traditional stock market. The value of the data won’t be determined by governments or ethics committees, but by those who own the data and decide by themselves which parts thereof they want to license to others under certain conditions. At first glance, a market-based system seems to be detrimental to the interest of certain monopolistic companies, as they would have to pay for the data - some would prefer free data and keep their monopoly. However, since every healthy and sick person in the market would suddenly have an incentive to collect and share their data under self-chosen anonymity conditions, there will soon be many more useful data to evaluate all kinds of treatments. On average, people will live longer and healthier, and many companies and the entire healthcare system will benefit. Jones: Finally, what is your view on open source versus the private companies like Google and OpenAI? Is there a danger to supporting these private companies’ large language models versus trying to keep these models open source and transparent, very much like what LAION is doing? Schmidhuber: I signed this open letter by LAION because I strongly favor the open-source movement. And I think it's also something that is going to challenge whatever big tech dominance there might be at the moment. Sure, the best models today are run by big companies with huge budgets for computers, but the exciting fact is that open-source models are not so far behind, some people say maybe six to eight months only. Of course, the private company models are all based on stuff that was created in academia, often in little labs without so much funding, which publish without patenting their results and open source their code and others take it and improved it. Big tech has profited tremendously from academia; their main achievement being that they have scaled up everything greatly, sometimes even failing to credit the original inventors. So, it's very interesting to see that as soon as some big company comes up with a new scaled-up model, lots of students out there are competing, or collaborating, with each other, trying to come up with equal or better performance on smaller networks and smaller machines. And since they are open sourcing, the next guy can have another great idea to improve it, so now there’s tremendous competition also for the big companies. Because of that, and since AI is still getting exponentially cheaper all the time, I don't believe that big tech companies will dominate in the long run. They find it very hard to compete with the enormous open-source movement. As long as you can encourage the open-source community, I think you shouldn't worry too much. Now, of course, you might say if everything is open source, then the bad actors also will more easily have access to these AI tools. And there's truth to that. But as always since the invention of controlled fire, it was good that knowledge about how technology works quickly became public such that everybody could use it. And then, against any bad actor, there's almost immediately a counter actor trying to nullify his efforts. You see, I still believe in our old motto "AI∀" or "AI For All." Jones: Thank you, Juergen for sharing your perspective on this amazing time in history. It’s clear that with new technology, the enormous potential can be matched by disparate and troubling risks which we’ve yet to solve, and even those we have yet to identify. If we are to dispel the fear of a sentient system for which we have no control, humans, alone need to take steps for more responsible development and collaboration to ensure AI technology is used to ultimately benefit society. Humanity will be judged by what we do next.

The delicate balance of building an online community business
reddit
LLM Vibe Score0
Human Vibe Score0.895
matthewbarbyThis week

The delicate balance of building an online community business

Hey /r/Entrepreneur 👋 Just under two years ago I launched an online community business called Traffic Think Tank with two other co-founders, Nick Eubanks and Ian Howells. As a Traffic Think Tank customer you (currently) pay $119 a month to get access to our online community, which is run through Slack. The community is focused on helping you learn various aspects of marketing, with a particular focus on search engine optimization (SEO). Alongside access to the Slack community, we publish new educational video content from outside experts every week that all customers have access to. At the time of writing, Traffic Think Tank has around 650 members spanning across 17 of the 24 different global time zones. I was on a business trip over in Sydney recently, and during my time there I met up with some of our Australia-based community members. During dinner I was asked by several of them how the idea for Traffic Think Tank came about and what steps we took to validate that the idea was worth pursuing.  This is what I told them… How it all began It all started with a personal need. Nick, an already successful entrepreneur and owner of a marketing agency, had tested out an early version Traffic Think Tank in early 2017. He offered real-time consulting for around ten customers that he ran from Slack. He would publish some educational videos and offer his advice on projects that the members were running. The initial test went well, but it was tough to maintain on his own and he had to charge a fairly high price to make it worth his time. That’s when he spoke to me and Ian about turning this idea into something much bigger. Both Ian and I offered something slightly different to Nick. We’ve both spent time in senior positions at marketing agencies, but currently hold senior director positions in 2,000+ public employee companies (HubSpot and LendingTree). Alongside this, as a trio we could really ramp up the quality and quantity of content within the community, spread out the administrative workload and just generally have more resources to throw at getting this thing off the ground. Admittedly, Nick was much more optimistic about the potential of Traffic Think Tank – something I’m very thankful for now – whereas Ian and I were in the camp of “you’re out of your mind if you think hundreds of people are going to pay us to be a part of a Slack channel”. To validate the idea at scale, we decided that we’d get an initial MVP of the community up and running with a goal of reaching 100 paying customers in the first six months. If we achieved that, we’d validated that it was a viable business and we would continue to pursue it. If not, we’d kill it. We spent the next month building out the initial tech stack that enabled us to accept payments, do basic user management to the Slack channel, and get a one-page website up and running with information on what Traffic Think Tank was all about.  After this was ready, we doubled down on getting some initial content created for members – I mean, we couldn’t have people just land in an empty Slack channel, could we? We created around ten initial videos, 20 or so articles and then some long threads full of useful information within the Slack channel so that members would have some content to pour into right from the beginning.  Then, it was time to go live. The first 100 customers Fortunately, both Nick and I had built a somewhat substantial following in the SEO space over the previous 5-10 years, so we at least had a large email list to tap into (a total of around 40,000 people). We queued up some launch emails, set an initial price of $99 per month and pressed send. [\[LINK\] The launch email I sent to my subscribers announcing Traffic Think Tank](https://mailchi.mp/matthewbarby/future-of-marketing-1128181) What we didn’t expect was to sell all of the initial 100 membership spots in the first 72 hours. “Shit. What do we do now? Are we ready for this many people? Are we providing them with enough value? What if something breaks in our tech stack? What if they don’t like the content? What if everyone hates Slack?” All of these were thoughts running through my head. This brings me to the first great decision we made: we closed down new membership intake for 3 months so that we could focus completely on adding value to the first cohort of users. The right thing at the right time SEO is somewhat of a dark art to many people that are trying to learn about it for the first time. There’s hundreds of thousands (possibly millions) of articles and videos online that talk about how to do SEO.  Some of it’s good advice; a lot of it is very bad advice.  Add to this that the barrier to entry of claiming to be an “expert” in SEO is practically non-existent and you have a recipe for disaster. This is why, for a long time, individuals involved in SEO have flocked in their masses to online communities for information and to bounce ideas off of others in the space. Forums like SEObook, Black Hat World, WickedFire, Inbound.org, /r/BigSEO, and many more have, at one time, been called home by many SEOs.  In recent times, these communities have either been closed down or just simply haven’t adapted to the changing needs of the community – one of those needs being real-time feedback on real-world problems.  The other big need that we all spotted and personally had was the ability to openly share the things that are working – and the things that aren’t – in SEO within a private forum. Not everyone wanted to share their secret sauce with the world. One of the main reasons we chose Slack as the platform to run our community on was the fact that it solved these two core needs. It gave the ability to communicate in real-time across multiple devices, and all of the information shared within it was outside of the public domain. The other problem that plagued a lot of these early communities was spam. Most of them were web-based forums that were free to access. That meant they became a breeding ground for people trying to either sell their services or promote their own content – neither of which is conducive to building a thriving community. This was our main motivation for charging a monthly fee to access Traffic Think Tank. We spent a lot of time thinking through pricing. It needed to be enough money that people would be motivated to really make use of their membership and act in a way that’s beneficial to the community, but not too much money that it became cost prohibitive to the people that would benefit from it the most. Considering that most of our members would typically spend between $200-800 per month on SEO software, $99 initially felt like the perfect balance. Growing pains The first three months of running the community went by without any major hiccups. Members were incredibly patient with us, gave us great feedback and were incredibly helpful and accommodating to other members. Messages were being posted every day, with Nick, Ian and myself seeding most of the engagement at this stage.  With everything going smoothly, we decided that it was time to open the doors to another intake of new members. At this point we’d accumulated a backlog of people on our waiting list, so we knew that simply opening our doors would result in another large intake. Adding more members to a community has a direct impact on the value that each member receives. For Traffic Think Tank in particular, the value for members comes from three areas: The ability to have your questions answered by me, Nick and Ian, as well as other members of the community. The access to a large library of exclusive content. The ability to build connections with the wider community. In the early stages of membership growth, there was a big emphasis on the first of those three points. We didn’t have an enormous content library, nor did we have a particularly large community of members, so a lot of the value came from getting a lot of one-to-one time with the community founders. [\[IMAGE\] Screenshot of engagement within the Traffic Think Tank Slack community](https://cdn.shortpixel.ai/client/qglossy,retimg,w_1322/https://www.matthewbarby.com/wp-content/uploads/2019/08/Community-Engagement-in-Traffic-Think-Tank.png) The good thing about having 100 members was that it was just about feasible to give each and every member some one-to-one time within the month, which really helped us to deliver those moments of delight that the community needed early on. Two-and-a-half months after we launched Traffic Think Tank, we opened the doors to another 250 people, taking our total number of members to 350. This is where we experienced our first growing pains.  Our original members had become used to being able to drop us direct messages and expect an almost instant response, but this wasn’t feasible anymore. There were too many people, and we needed to create a shift in behavior. We needed more value to come from the community engaging with one another or we’d never be able to scale beyond this level. We started to really pay attention to engagement metrics; how many people were logging in every day, and of those, how many were actually posting messages within public channels.  We asked members that were logging in a lot but weren’t posting (the “lurkers”) why that was the case. We also asked the members that engaged in the community the most what motivated them to post regularly. We learned a lot from doing this. We found that the large majority of highly-engaged members had much more experience in SEO, whereas most of the “lurkers” were beginners. This meant that most of the information being shared in the community was very advanced, with a lot of feedback from the beginners in the group being that they “didn’t want to ask a stupid question”.  As managers of the community, we needed to facilitate conversations that catered to all of our members, not just those at a certain level of skill. To tackle this problem, we created a number of new channels that had a much deeper focus on beginner topics so novice members had a safe place to ask questions without judgment.  We also started running live video Q&As each month where we’d answer questions submitted by the community. This gave our members one-on-one time with me, Nick and Ian, but spread the value of these conversations across the whole community rather than them being hidden within private messages. As a result of these changes, we found that the more experienced members in the community were really enjoying sharing their knowledge with those with less experience. The number of replies within each question thread was really starting to increase, and the community started to shift away from just being a bunch of threads created by me, Nick and Ian to a thriving forum of diverse topics compiled by a diverse set of individuals. This is what we’d always wanted. A true community. It was starting to happen. [\[IMAGE\] Chart showing community engagement vs individual member value](https://cdn.shortpixel.ai/client/qglossy,retimg,w_1602/https://www.matthewbarby.com/wp-content/uploads/2019/08/Community-Engagement-Balance-Graph.jpg) At the same time, we started to realize that we’ll eventually reach a tipping point where there’ll be too much content for us to manage and our members to engage with. When we reach this point, the community will be tough to follow and the quality of any given post will go down. Not only that, but the community will become increasingly difficult to moderate. We’re not there yet, but we recognize that this will come, and we’ll have to adjust our model again. Advocating advocacy As we started to feel more comfortable about the value that members were receiving, we made the decision to indefinitely open for new members. At the same time, we increased the price of membership (from $99 a month to $119) in a bid to strike the right balance between profitability as a business and to slow down the rate at which we were reaching the tipping point of community size. We also made the decision to repay all of our early adopters by grandfathering them in to the original pricing – and committing to always do this in the future. Despite the price increase, we saw a continued flow of new members come into the community. The craziest part about this was that we were doing practically no marketing activities to encourage new members– this was all coming from word of mouth. Our members were getting enough value from the community that they were recommending it to their friends, colleagues and business partners.  The scale at which this was happening really took us by surprise and it told us one thing very clearly: delivering more value to members resulted in more value being delivered to the business. This is a wonderful dynamic to have because it perfectly aligns the incentives on both sides. We’d said from the start that we wouldn’t sacrifice value to members for more revenue – this is something that all three of us felt very strongly about. First and foremost, we wanted to create a community that delivered value to its members and was run in a way that aligned with our values as people. If we could find a way to stimulate brand advocacy, while also tightening the bonds between all of our individual community members, we’d be boosting both customer retention and customer acquisition in the same motion. This became our next big focus. [\[TWEET\] Adam, one of our members wore his Traffic Think Tank t-shirt in the Sahara desert](https://twitter.com/AdamGSteele/status/1130892481099382784) We started with some simple things: We shipped out Traffic Think Tank branded T-shirts to all new members. We’d call out each of the individuals that would submit questions to our live Q&A sessions and thank them live on air. We set up a new channel that was dedicated to sharing a quick introduction to who you are, what you do and where you’re based for all new members. We’d created a jobs channel and a marketplace for selling, buying and trading services with other members. Our monthly “blind dates” calls were started where you’d be randomly grouped with 3-4 other community members so that you could hop on a call to get to know each other better. The Traffic Think Tank In Real Life (IRL)* channel was born, which enabled members to facilitate in-person meetups with each other. In particular, we saw that as members started to meet in person or via calls the community itself was feeling more and more like a family. It became much closer knit and some members started to build up a really positive reputation for being particularly helpful to other members, or for having really strong knowledge in a specific area. [\[TWEET\] Dinner with some of the Traffic Think Tank members in Brighton, UK](https://twitter.com/matthewbarby/status/1117175584080134149) Nick, Ian and I would go out of our way to try and meet with members in real life wherever we could. I was taken aback by how appreciative people were for us doing this, and it also served as an invaluable way to gain honest feedback from members. There was another trend that we’d observed that we didn’t really expect to happen. More and more members were doing business with each another. We’ve had people find new jobs through the community, sell businesses to other members, launch joint ventures together and bring members in as consultants to their business. This has probably been the most rewarding thing to watch, and it was clear that the deeper relationships that our members were forming were resulting in an increased level of trust to work with each other. We wanted to harness this and take it to a new level. This brought us to arguably the best decision we’ve made so far running Traffic Think Tank… we were going to run a big live event for our members. I have no idea what I’m doing It’s the first week of January 2019 and we’re less than three weeks away from Traffic Think Tank LIVE, our first ever in-person event hosting 150 people, most of which are Traffic Think Tank members. It's like an ongoing nightmare I can’t wake up from. That was Nick’s response in our private admin channel to myself and Ian when I asked if they were finding the run-up to the event as stressful as I was. I think that all three of us were riding on such a high from how the community was growing that we felt like we could do anything. Running an event? How hard can it be? Well, turns out it’s really hard. We had seven different speakers flying over from around the world to speak at the event, there was a pre- and after event party, and we’d planned a charity dinner where we would take ten attendees (picked at random via a raffle) out for a fancy meal. Oh, and Nick, Ian and I were hosting a live Q&A session on stage. It wasn’t until precisely 48 hours before the event that we’d realized we didn’t have any microphones, nor had a large amount of the swag we’d ordered arrived. Plus, a giant storm had hit Philly causing a TON of flight cancellations. Perfect. Just perfect. This was honestly the tip of the iceberg. We hadn’t thought about who was going to run the registration desk, who would be taking photos during the event and who would actually field questions from the audience while all three of us sat on stage for our live Q&A panel. Turns out that the answer to all of those questions were my wife, Laura, and Nick’s wife, Kelley. Thankfully, they were on hand to save our asses. The weeks running up to the event were honestly some of the most stressful of my life. We sold around 50% of our ticket allocation within the final two weeks before the event. All of the event organizers told us this would happen, but did we believe them? Hell no!  Imagine having two weeks until the big day and as it stood half of the room would be completely empty. I was ready to fly most of my extended family over just to make it look remotely busy. [\[IMAGE\] One of our speakers, Ryan Stewart, presenting at Traffic Think Tank LIVE](https://cdn.shortpixel.ai/client/qglossy,retimg,w_1920/https://www.matthewbarby.com/wp-content/uploads/2019/08/Traffic-Think-Tank-LIVE-Ryan-Presenting.jpg) Thankfully, if all came together. We managed to acquire some microphones, the swag arrived on the morning of the event, all of our speakers were able to make it on time and the weather just about held up so that our entire allocation of ticket holders was able to make it to the event. We pooled together and I’m proud to say that the event was a huge success. While we made a substantial financial loss on the event itself, January saw a huge spike in new members, which more than recouped our losses. Not only that, but we got to hang out with a load of our members all day while they said really nice things about the thing we’d built. It was both exhausting and incredibly rewarding. Bring on Traffic Think Tank LIVE 2020! (This time we’re hiring an event manager...)   The road ahead Fast forward to today (August 2019) and Traffic Think Tank has over 650 members. The biggest challenges that we’re tackling right now include making sure the most interesting conversations and best content surfaces to the top of the community, making Slack more searchable (this is ultimately one of its flaws as a platform) and giving members a quicker way to find the exclusive content that we create. You’ll notice there’s a pretty clear theme here. In the past 30 days, 4,566 messages were posted in public channels inside Traffic Think Tank. If you add on any messages posted inside private direct messages, this number rises to 21,612. That’s a lot of messages. To solve these challenges and enable further scale in the future, we’ve invested a bunch of cash and our time into building out a full learning management system (LMS) that all members will get access to alongside the Slack community. The LMS will be a web-based portal that houses all of the video content we produce. It will also  provide an account admin section where users can update or change their billing information (they have to email us to do this right now, which isn’t ideal), a list of membership perks and discounts with our partners, and a list of links to some of the best threads within Slack – when clicked, these will drop you directly into Slack. [\[IMAGE\] Designs for the new learning management system (LMS)](https://cdn.shortpixel.ai/client/qglossy,retimg,w_2378/https://www.matthewbarby.com/wp-content/uploads/2019/08/Traffic-Think-Tank-LMS.png) It’s not been easy, but we’re 95% of the way through this and I’m certain that it will have a hugely positive impact on the experience for our members. Alongside this we hired a community manager, Liz, who supports with any questions that our members have, coordinates with external experts to arrange webinars for the community, helps with new member onboarding, and has tightened up some of our processes around billing and general accounts admin. This was a great decision. Finally, we’ve started planning next year’s live event, which we plan to more than double in size to 350 attendees, and we decided to pick a slightly warmer location in Miami this time out. Stay tuned for me to have a complete meltdown 3 weeks from the event. Final thoughts When I look back on the journey we’ve had so far building Traffic Think Tank, there’s one very important piece to this puzzle that’s made all of this work that I’ve failed to mention so far: co-founder alignment. Building a community is a balancing act that relies heavily on those in charge being completely aligned. Nick, Ian and I completely trust each other and more importantly, are philosophically aligned on how we want to run and grow the community. If we didn’t have this, the friction between us could tear apart the entire community. Picking the right people to work with is important in any company, but when your business is literally about bringing people together, there’s no margin for error here.  While I’m sure there will be many more challenges ahead, knowing that we all trust each other to make decisions that fall in line with each of our core values makes these challenges dramatically easier to overcome. Finally, I’d like to thank all of our members for making the community what it is today – it’d be nothing without you and I promise that we’ll never take that for granted. ​ I originally posted this on my blog here. Welcoming all of your thoughts, comments, questions and I'll do my best to answer them :)

My Side Projects: From CEO to 4th Developer (Thanks, AI 🤖)
reddit
LLM Vibe Score0
Human Vibe Score1
tilopediaThis week

My Side Projects: From CEO to 4th Developer (Thanks, AI 🤖)

Hey Reddit 👋, I wanted to share a bit about some side projects I’ve been working on lately. Quick background for context: I’m the CEO of a mid-to-large-scale eCommerce company pulling in €10M+ annually in net turnover. We even built our own internal tracking software that’s now a SaaS (in early review stages on Shopify), competing with platforms like Lifetimely and TrueROAS. But! That’s not really the point of this post — there’s another journey I’ve been on that I’m super excited to share (and maybe get your feedback on!). AI Transformed My Role (and My Ideas List) I’m not a developer by trade — never properly learned how to code, and to be honest, I don’t intend to. But, I’ve always been the kind of guy who jots down ideas in a notes app and dreams about execution. My dev team calls me their “4th developer” (they’re a team of three) because I have solid theoretical knowledge and can kinda read code. And then AI happened. 🛠️ It basically turned my random ideas app into an MVP generation machine. I thought it’d be fun to share one of the apps I’m especially proud of. I am also planning to build this in public and therefore I am planning to post my progress on X and every project will have /stats page where live stats of the app will be available. Tackling My Task Management Problem 🚀 I’ve sucked at task management for YEARS, I still do! I’ve tried literally everything — Sheets, Todoist, Asana, ClickUp, Notion — you name it. I’d start… and then quit after a few weeks - always. What I struggle with the most is delegating tasks. As a CEO, I delegate a ton, and it’s super hard to track everything I’ve handed off to the team. Take this example: A few days ago, I emailed an employee about checking potential collaboration opportunities with a courier company. Just one of 10s of tasks like this I delegate daily. Suddenly, I thought: “Wouldn’t it be AMAZING if just typing out this email automatically created a task for me to track?” 💡 So… I jumped in. With the power of AI and a few intense days of work, I built a task manager that does just that. But of course, I couldn’t stop there. Research & Leveling It Up 📈 I looked at similar tools like TickTick and Todoist, scraped their G2 reviews (totally legally, promise! 😅), and ran them through AI for a deep SWOT analysis. I wanted to understand what their users liked/didn’t like and what gaps my app could fill. Some of the features people said they were missing didn’t align with the vision for my app (keeping it simple and personal), but I found some gold nuggets: Integration with calendars (Google) Reminders Customizable UX (themes) So, I started implementing what made sense and am keeping others on the roadmap for the future. And I’ve even built for that to, it still doesn’t have a name, however the point is you select on how many reviews of a specific app you want to make a SWOT analysis on and it will do it for you. Example for Todoist in comments. But more on that, some other time, maybe other post ... Key Features So Far: Here’s what’s live right now: ✅ Email to Task: Add an email as to, cc, or bcc — and it automatically creates a task with context, due dates, labels, etc. ✅ WhatsApp Reminders: Get nudged to handle your tasks via WhatsApp. ✅ WhatsApp to Task: Send a message like /task buy groceries — bam, it’s added with full context etc.. ✅ Chrome Extension (work-in-progress): Highlight text on any page, right-click, and send it straight to your task list. Next Steps: Build WITH the Community 👥 Right now, the app is 100% free while still in the early stages. But hey, API calls and server costs aren’t cheap, so pricing is something I’ll figure out with you as we grow. For now, my goal is to hit 100 users and iterate from there. My first pricing idea is, without monthly subscription, I don’t want to charge someone for something he didn’t use. So I am planning on charging "per task", what do you think? Here’s what I have planned: 📍 End of Year Goal: 100 users (starting from… 1 🥲). 💸 Revenue Roadmap: When we establish pricing, we’ll talk about that. 🛠️ Milestones: Post on Product Hunt when we hit 100 users. Clean up my self-written spaghetti code (hire a pro dev for review 🙃). Hire a part-time dev once we hit MRR that can cover its costs. You can check how are we doing on thisisatask.me/stats Other Side Projects I’m Working On: Because… what’s life without taking on too much, right? 😂 Full list of things I’m building: Internal HRM: Not public, tried and tested in-house. Android TV App: Syncs with HRM to post announcements to office TVs (streamlined and simple). Stats Tracker App: Connects to our internal software and gives me real-time company insights. Review Analyzer: Scrapes SaaS reviews (e.g., G2) and runs deep analysis via AI. This was originally for my Shopify SaaS but is quickly turning into something standalone. Coming soon! Mobile app game: secret for now. Let’s Build This Together! Would love it if you guys checked out thisisatask.me and gave it a spin! Still super early, super raw, but I’m pumped to hear your thoughts. Also, what’s a must-have task manager feature for you? Anything that frustrates you with current tools? I want to keep evolving this in public, so your feedback is gold. 🌟 Let me know, Reddit! Are you with me? 🙌

Writing a exercise based TTRPG rulebook for a system where your real world fitness is tied to character progression
reddit
LLM Vibe Score0
Human Vibe Score1
BezboznyThis week

Writing a exercise based TTRPG rulebook for a system where your real world fitness is tied to character progression

My dad was a star athlete when he was young, and my mom was a huge sci-fi/fantasy nerd, so I got both ends of the stick as it were. Love gaming and nerd culture, but also love to exercise and self improvement. Sometimes exercise can feel boring though compared to daydreaming about fantastic fictional worlds, so for a long time I've been kicking around the idea of how to "Gamify" fitness. and recently I've been working on this passion project of a Table Top RPG (Like D&D) where the stats of your character are related to your own fitness, so if you want your character in game to improve, you have to improve in the real world. Below is a rough draft you can look through that details the settings and mechanics of the game I've come up with so far. I'd love to eventually get a full book published and sell it online. maybe even starting a whole brand of "Gamified fitness": REP-SET: GAINSZ In the war torn future of 24th century… There are no rest days… In the futuristic setting of "REP-SET: GAINSZ," the "War of Gains" casts a long shadow over the Sol System as the various factions vie for territory and resources. However, war has evolved. Unmanned drones and long-range strikes have faded into obsolescence. Battles, both planet-side and in the depths of space, are now fought by soldiers piloting REP-SETs: Reactive Exoskeletal Platform - Symbiotic Evolution Trainer Massive, humanoid combat mechs. Powered by mysterious “EV” energy, these mechanical marvels amplify, and are in turn amplified by, the fitness and mental acuity of their pilots. The amplification is exponential, leading pilots into a life of constant training in order for their combat prowess to be bolstered by every incremental gain in their level of fitness. With top pilots having lifting capacity measured in tons, and reaction times measured by their Mach number, REP-SET enhanced infantry now dominate the battlefield. The Factions: The Federated Isometocracy of Terra (FIT): Quote: "The strength of the body is the strength of the spirit. Together, we will lift humanity to its destined greatness. But ask not the federation to lift for you. Ask yourself: Do you even lift for the Federation?" Description: An idealistic but authoritarian faction founded on the principle of maximizing the potential of all individuals. FIT citizens believe in relentless striving for physical and mental perfection, leading to collective excellence. Their goal is the unification of humankind under a rule guided by this doctrine, which sometimes comes at the cost of individual liberties. Mech Concept: REP-SET mechs. Versatile humanoid designs focusing on strength, endurance, and adaptability. By connecting to the AI spirit within their REP-SETs core, each pilot enhances the performance of their machine through personal willpower and peak physical training. Some high-rank REP-SETS include features customized to the pilot's strengths, visually signifying their dedication and discipline. The Dominion of Organo-Mechanical Supremacy (DOMS): Quote: "Without pain, there is no gain. Become the machine. Embrace the burn.” Description: A fanatical collective ideologically obsessed with "Ascendency through suffering" by merging their bodies with technology that not only transcends biological limitations, but also acts to constantly induce pain in it's users. Driven by a sense of ideological superiority and a thirst for domination, DOMS seek to bring the painful blessings of their deity "The lord of the Burn" to the rest of the solar system. Their conquest could turn them into a significant threat to humanity. Mech Concept: Hybrid mechs, where the distinction between the pilot and the machine is blurred. The cockpit functions as a life-support system for the pilot, heavily modified with augmentations. Mechs themselves are often modular, allowing for adaptation and assimilation of enemy technology. Some DOMS mechs might display disturbing elements of twisted flesh alongside cold, mechanical parts. The Tren: Quote: "Grow... bigger... feast... protein..." Description: A ravenous conglomeration of biochemically engineered muscular monstrosities, united only by a shared insatiable hunger for "More". Existing mostly in deep space, they seek organic matter to consume and assimilate. They progress in power not due to any form of training or technology, but from a constant regimen of ravenous consumption and chemically induced muscle growth, all exponentially enhanced by EV energies. While some have been known to possess a certain level of intellect and civility, their relentless hunger makes them incredibly mentally volatile. When not consuming others, the strong consume the weak within their own faction. Mech Concept: Bio-Organic horrors. While they do have massive war machines, some are living vessels built around immense creatures. These machines resemble grotesque fleshy designs that prioritize rapid mutation and growth over sleek aesthetics. Often unsettling to behold. Synthetic Intelligence Theocracy (SIT): Quote: "Failure is an unacceptable data point.” Description: A society ruled by a vast and interconnected artificial intelligence network. The SIT governs with seemingly emotionless rationality, striving for efficiency and maximum productivity. This leads to a cold, but arguably prosperous society, unless you challenge the logic of the collective AI. Their goals? Difficult to predict, as it hinges on how the AI calculates what's "optimal" for the continuation or "evolution" of existence. Mech Concept: Sleek, almost featureless robotic creations with a focus on efficient movement and energy management. Often drone-like or modular, piloted through direct mind-machine linking rather than traditional cockpits. Their aesthetic suggests cold and impersonal perfection. The Way Isolate(TWI): Quote: "The body unblemished, the mind unwavering. That is the path to true strength. That and a healthy diet of Aster-Pea proteins." Description: Known by some as "The asteroid farmers", The Way Isolate is a proud and enigmatic faction that stands apart from the other powers in the Sol System. A fiercely independent tribe bound by oaths of honor, loyalty, and hard work. Wandering the asteroid belt in their vast arc ships, their unparalleled mastery in asteroidal-agricultural engineering, ensuring they have no need to colonize planets for nutritional needs, has allowed them to abstain from the pursuit of territorial expansion in “The War of Gains”, instead focusing on inward perfection, both spiritual and physical. They eschew all technological bodily enhancements deemed unnatural, believing that true power can only be cultivated through the relentless pursuit of personal strength achieved through sheer will and bodily perfection. The Way Isolate views biohacking, genetic manipulation, and even advanced cybernetics as corruptions of the human spirit, diluting the sacredness of individual willpower. Mech Concept: Way Isolate mechs are built with maneuverability and precision in mind rather than flashy augmentations. Their REP-SETs are streamlined, favoring lean designs that mirror the athleticism of their pilots. Excelling in low to zero G environments, their mechs lack bulky armor, relying on evasion and maneuverability rather than brute force endurance. Weaponry leans towards traditional kinetic based armaments, perhaps employing archaic but reliable weapon styles such as blades or axes as symbols of their purity of purpose. These mechs reflect the individual prowess of their pilots, where victory is determined by focus, technique, and the raw power of honed physical ability. Base Player Character Example: You are a young, idealistic FIT soldier, barely out of training and working as a junior REP-SET mechanic on the Europa Ring World. The Miazaki district, a landscape of towering mountains and gleaming cities, houses a sprawling mountainside factory – a veritable hive of Gen 5 REP-SET construction. Here, the lines between military and civilian blur within a self-sufficient society dependent on this relentless industry. Beneath the surface, you harbor a secret. In a forgotten workshop, the ghost of a REP-SET takes shape – a unique machine built around an abandoned, enigmatic AI core. Ever since you salvaged it as a child from the wreckage of your hometown, scarred by a brutal Tren attack, you've dedicated yourself to its restoration. A lingering injury from that fateful battle mocks your progress, a constant reminder of the fitness exams you cannot pass. Yet, you train relentlessly, dreaming of the day you'll stand as a true REP-SET pilot. A hidden truth lies at the heart of the REP-SETS: as a pilot's abilities grow, their mech develops unique, almost mystical powers – a manifestation of the bond between the human spirit and the REP-SET's AI. The ache in your old wound serves as a grim prophecy. This cold war cannot last. The drums of battle grow louder with each passing day. GAME MECHANICS: The TTRPG setting of “REP-SET: GAINSZ” is marked by a unique set of rules, by which the players real world capabilities and fitness will reflect and affect the capabilities, progression, and success of their REP-SET pilot character in-game. ABILITY SCORES: Pilots' capabilities will be defined by 6 “Ability scores”: Grace, Agility, Iron, Nourishment, Strength, and Zen. Each of the 6 ability scores will duel represent both a specific area of exercise/athleticism and a specific brand of healthy habits. The definitions of these ability scores are as follows: Grace (GRC): "You are an artist, and your body is your canvas; the way you move is your paint and brush." This ability score, the domain of dancers and martial artists, represents a person's ability to move with organic, flowing control and to bring beauty to the world. Skill challenges may be called upon when the player character needs to act with poise and control, whether socially or physically. Real-world skill checks may involve martial arts drills, dancing to music, or balance exercises. Bonuses may be granted if the player has recently done something artistically creative or kind, and penalties may apply if they have recently lost their temper. This ability score affects how much NPCs like your character in game. Agility (AGI): "Your true potential is locked away, and speed is the key to unlocking it." The domain of sprinters, this ability score represents not only a person's absolute speed and reaction time but also their capacity to finish work early and avoid procrastination. Skill challenges may be called upon when the player character needs to make a split-second choice, move fast, or deftly dodge something dangerous. Real-world skill checks may involve acts of speed such as sprinting or punching/kicking at a steadily increasing tempo. Bonuses may apply if the player has finished work early, and penalties may apply if they are procrastinating. This ability score affects moving speed and turn order in game. Iron (IRN): "Not money, nor genetics, nor the world's greatest trainers... it is your resolve, your will to better yourself, that will make you great." Required by all athletes regardless of focus, this ability score represents a player's willpower and their capacity to push through pain, distraction, or anything else to achieve their goals. Skill challenges may be called upon when the player character needs to push through fear, doubt, or mental manipulation. Real-world skill checks may involve feats of athletic perseverance, such as planking or dead hangs from a pull-up bar. Bonuses may apply when the player maintains or creates scheduled daily routines of exercise, self-improvement, and work completion, and penalties may apply when they falter in those routines. This ability score affects the max "Dynamic exercise bonus” that can be applied to skill checks in game (a base max of +3 when Iron = 10, with an additional +1 for every 2 points of iron. So if every 20 pushups gives you +1 on a “Strength” skill check, then doing 80 pushups will only give you +4 if you have at least 12 iron). Nourishment (NRS): "A properly nourished body will last longer than a famished one." This ability score, focused on by long-distance runners, represents a player's endurance and level of nutrition. Skill challenges may be called upon when making checks that involve the player character's stamina or health. Real-world skill checks may involve endurance exercises like long-distance running. Bonuses may apply if the player has eaten healthily or consumed enough water, and penalties may apply if they have eaten junk food. This ability score affects your HP (Health points), which determines how much damage you can take before you are incapacitated. Strength (STR): "When I get down on my hands, I'm not doing pushups, I'm bench-pressing the planet." The domain of powerlifters and strongmen, this ability score represents raw physical might and the ability to overcome obstacles. Skill challenges may be called upon when the player character needs to lift, push, or break something. Real-world skill checks might involve weightlifting exercises, feats of grip strength, or core stability tests. Bonuses may apply for consuming protein-rich foods or getting a good night's sleep, and penalties may apply after staying up late or indulging in excessive stimulants. This ability score affects your carrying capacity and base attack damage in game. Zen (ZEN): "Clarity of mind reflects clarity of purpose. Still the waters within to act decisively without." This ability score, prized by meditators and yogis, represents mental focus, clarity, and inner peace. Skill challenges may be called upon when the player character needs to resist distractions, see through illusions, or make difficult decisions under pressure. Real-world skill checks may involve meditation, breathing exercises, or mindfulness activities. Bonuses may apply after attending a yoga class, spending time in nature, or creating a calm and organized living space. Penalties may apply after experiencing significant stress, emotional turmoil, or having an unclean or unorganized living space. This ability score affects your amount of ZP in game (Zen Points: your pool of energy you pull from to use mystical abilities) Determining initial player ability scores: Initially, “Ability scores” are decided during character creation by giving the player a list of 6 fitness tests to gauge their level of fitness in each category. Running each test through a specific calculation will output an ability score. A score of 10 represents the average person, a score of 20 represents a peak athlete in their category. The tests are: Grace: Timed balancing on one leg with eyes closed (10 seconds is average, 60 is peak) Agility: Mile run time in minutes and second (10:00 minutes:seconds is average, 3:47 is peak) Iron: Timed dead-hang from a pull-up bar (30 seconds is average, 160 is peak) Nourishment: Miles run in an hour (4 is average, 12 is peak) Strength: Pushups in 2 minute (34 is average, 100 is peak) Zen: Leg stretch in degrees (80 is average, and 180 aka "The splits" is peak) Initial Score Calculation Formula: Ability Score = 10 + (Player Test Score - Average Score) / (Peak Score - Average\_Score) \* 10 Example: if the player does 58 pushups in 2 minutes, their strength would be: 10 plus (58 - 34) divided by (100-34) multiplied by 10 = 10 + (24)/(66)\* 10 = 10 + 3.6363... = 13.6363 rounded to nearest whole number = Strength (STR): 14 SKILLS AND SKILL CHALLENGES: The core mechanic of the game will be in how skill challenges are resolved. All “Skill challenges” will have a numerical challenge rating that must be met or beaten by the sum of a 10 sided dice roll and your score in the pertinent skill. Skill scores are determined by 2 factors: Ability Score Bonus: Every 2 points above 10 gives +1 bonus point. (EX. 12 = +1, 14 = +2, etc.) This also means that if you have less than 10 in an ability score, you will get negative points. Personal Best Bonus: Each skill has its own unique associated exercise that can be measured (Time, speed, distance, amount of reps, etc). A higher record means a higher bonus. EX: Authority skill checks are associated with a timed “Lateral raise hold”. Every 30 seconds of the hold added onto your personal best single attempt offers a +1 bonus. So if you can do a lateral hold for 90 seconds, that’s a +3 to your authority check! So if you have a 16 in Iron, and your Personal Best lateral raise hold is 90 seconds, that would give you an Authority score of +6 (T-Pose for dominance!) Dynamic Exercise Bonus: This is where the unique mechanics of the game kick in. At any time during a skill challenge (even after your roll) you can add an additional modifier to the skill check by completing the exercise during gameplay! Did you roll just below the threshold for success? Crank out another 20 pushups, squats, or curls to push yourself just over the edge into success! There are 18 skills total, each with its own associated ability score and unique exercise: Grace (GRC): \-Kinesthesia (Timed: Blind single leg stand time) \-Precision (Scored: Basket throws) \-Charm (Timed reps: Standing repeated forward dumbell chest press and thrust) \-Stealth (Timed distance: Leopard Crawl) Agility (AGI): \-acrobatics (timed reps: high kicks) \-Computers (Word per minute: Typing test) \-Speed (Time: 100 meter sprint) Iron (IRN): \-Authority (Timed: Lateral raise hold) \-Resist (Timed: Plank) \-Persist (Timed:Pull-up bar dead hang) Nourishment(NRS): \-Recovery (TBD) \-Stim crafting (TBD) \-Survival (TBD) Strength(STR): \-Mechanics (Timed reps: Alternating curls) \-Might (Timed reps: pushups) Zen(ZEN): \-Perceive (TBD) \-Empathy (TBD) \-Harmony (TBD) \-Lore (TBD) Healthy Habits Bonus: Being able to demonstrate that you have conducted healthy habits during gameplay can also add one time bonuses per skill challenge “Drank a glass of water +1 to Nourishment check”, “Cleaned your room, +3 on Zen check”. But watch out, if you’re caught in unhealthy Habits, the GM can throw in penalties, “Ate junk food, -1 to Nourishment check”, etc. Bonuses/penalties from in-game items, equipment, buffs, debuffs, etc., helping players to immerse into the mechanics of the world of REP-SET for the thrill of constantly finding ways to improve their player. Gradient success: Result of skill challenges can be pass or fail, but can also be on a sliding scale of success. Are you racing to the battlefield? Depending on your Speed check, you might arrive early and have a tactical advantage, just in time for an even fight, or maybe far too late and some of your favorite allied NPCs have paid the price… So you’re often encouraged to stack on those dynamic exercise bonuses when you can to get the most fortuitous outcomes available to you. Gameplay sample: GM: Your REP-SET is a phantom, a streak of light against the vast hull of the warship. Enemy fighters buzz angrily, but you weaves and dodges with uncanny precision. The energy wave might be losing effectiveness, but your agility and connection to the machine have never been stronger. Then, it happens. A gap in the defenses. A vulnerable seam in the warship's armor. Your coms agents keen eye spots it instantly. "Lower power junction, starboard side! You have an opening!" This is your chance to strike the decisive blow. But how? It'll take a perfect combination of skill and strategy, drawing upon your various strengths. Here are your options: Option 1: Brute Strength: Channel all remaining power into a single, overwhelming blast from the core. High-risk, high-reward. It could overload the REP-SET if you fail, but it might also cripple the warship. (Strength-focused, Might sub-skill) Option 2: Calculated Strike: With surgical precision, target the power junction with a pinpoint burst of destabilizing energy. Less flashy and ultimately less damaging, but potentially more effective in temporarily disabling the ship. (Agility-focused, Precision sub-skill) Option 3: Harmonic Disruption: Attempt to harmonize with your REP-SET's AI spirit for help in connecting to the digital systems of the Warship. Can you generate an internal energy resonance within the warship, causing it to malfunction from within? (Zen-focused, Harmony sub-skill) Player: I'll take option 1, brute strength! GM: Ok, This will be a "Might" check. The CR is going to be very high on this one. I'm setting it at a 20. What's your Might bonus? Player: Dang, a 20?? That's literally impossible. My Might is 15 and I've got a PB of 65 pushups in 2 minutes, that sets me at a +5. Even if I roll a 10 and do 60 pushups for the DE I'll only get 18 max. GM: Hey I told you it was high risk. You want to choose another option? Player: No, no. This is what my character would do. I'm a real hot-blooded meathead for sure. GM: Ok then, roll a D10 and add your bonus. Player: \Rolls\ a 9! not bad, actually that's a really good roll. So +5, that's a 14. GM: Alright, would you like to add a dynamic exercise bonus? Player: Duh, it's not like I can do 120 pushups I'd need to beat the CR, but I can at least do better than 14. Alright, here goes. \the player gets down to do pushups and the 2 minute time begins. After some time...\ Player: 65....... 66! GM: Times up. Player: Ow... my arms... GM: so with 66, that's an extra +3, and its a new PB, so that's a +1. That sets your roll to 18. Player: Ow... Frack... still not 20... for a second there i really believed I could do 120 pushups... well I did my best... Ow... 20 CR is just too impossible you jerk... GM: Hmm... Tell me, what did you eat for lunch today? Player: Me? I made some vegetable and pork soup, and a protein shake. I recorded it all in my diet app. GM: And how did you sleep last night? Player: Like a baby, went to sleep early, woke up at 6. GM: in that case, you can add a +1 "Protein bonus" and +1 "Healthy rest" bonus to any strength related check for the day if you'd like, including this one. Player: Really?? Heck yes! add it to the roll! GM: With those extra bonuses, your roll reaches 20. How do you want to do this? Player: I roar "For Terra!" and pour every last ounce of my strength into the REP-SET. GM: "For Terra!" you roar, your cry echoing through coms systems of the REP-SET. The core flares blindingly bright. The surge of power dwarfs anything the REP-SET has unleashed before. With a titanic shriek that cracks the very fabric of space, the REP-SET slams into the vulnerable power junction. Raw energy explodes outwards, tendrils of light arcing across the warship's massive hull. The impact is staggering. The leviathan-like warship buckles, its sleek form rippling with shockwaves. Sparks shower like rain, secondary explosions erupt as critical systems overload. Then…silence. The warship goes dark. Power flickers within the REP-SET itself, then steadies. Alarms fade, replaced by the eerie quiet of damaged but functional systems. "We…did it?" The coms agents voice is incredulous, tinged with relief. She's awaiting your reply. Player: "I guess so." I say, and I smile and laugh. And then I slump back... and fall unconscious. \to the other players\ I'm not doing any more skill checks for a while guys, come pick me up please. \teammates cheer\ ​

The Weekly Brief for anyone looking to incorporate AI into their business.
reddit
LLM Vibe Score0
Human Vibe Score1
AI_Business_BriefThis week

The Weekly Brief for anyone looking to incorporate AI into their business.

Good morning and happy Sunday. Today is Sunday and you know what that means. The weekly brief. Covering all of last week’s most important AI business related stories. Here are some of the biggest stories: Claude the newest generative AI. Amazon to change up its search. AI leaders Testify. Meta Open sources its LLM. Voice Actors Struggle Growing AI innovations has led to a struggle for many voice actors. As AI powered voice technology is progressing everyday jobs are becoming more and more scarce. With many publishers already leaning towards replacing many of their voice actors for faster, cheaper, and more efficient AI voices. Meet Claude Anthropic, an AI company founded by ex-OpenAI employee released their generative AI called Claude. Some key aspects of their model is the ability to give more correct and less harmful answers, and perform similar tasks that many other generative AI’s can do. A keynote is that Google has invested 300milloion into the company, which is a direct competitor to their AI Bard. Interesting to see how that will play out. Amazon Changes to Change up Search A new job description at Amazon may have hinted towards their future plans for AI. The description under software developer read “reimagining Amazon Search with an interactive conversational experience”. This may hint towards a generative AI search experience in Amazon for customers. ChatGPT User Get More Access Premium ChatGPT users got access to Web browsing and plugins. This is a crucial step for OpenAI as they plan to pivot to a more assist type AI. While at the same time continuing to research and develop their AI models. This move puts a lot of pressure on Google to hopefully step up their game. AI Leaders Testify This Wednesday AI leaders (Sam Altman, Christina Montgomery and Gary Marcus) all testified before congress about AI regulation. They were asked many questions about AI regulation but came up with two solutions. FDA-Like Approval Processing: AI developing companies are open to safety checks, audits, licensing and risk review. Precision Approach: Develop risk rules, provide explanations and provide guidelines for risks, encourage transparency around AI companies, finally assess impact of AI technologies. Meta Open Sourcing Thursday Meta open sourced this coding for their LLM. As the company wants to see the use of its LLM to help drive innovation, inspire smaller companies, and overall develop better AI technologies. Comes as an interesting move as competitors try and keep their AI’s an insider secret.

AITreasureBox
github
LLM Vibe Score0.447
Human Vibe Score0.1014145151561518
superiorluMar 28, 2025

AITreasureBox

AI TreasureBox English | 中文 Collect practical AI repos, tools, websites, papers and tutorials on AI. Translated from ChatGPT, picture from Midjourney. Catalog Repos Tools Websites Report&Paper Tutorials Repos updated repos and stars every 2 hours and re-ranking automatically. | No. | Repos | Description | | ----:|:-----------------------------------------|:------------------------------------------------------------------------------------------------------| | 1|🔥codecrafters-io/build-your-own-x !2025-03-28364681428|Master programming by recreating your favorite technologies from scratch.| | 2|sindresorhus/awesome !2025-03-28353614145|😎 Awesome lists about all kinds of interesting topics| | 3|public-apis/public-apis !2025-03-28334299125|A collective list of free APIs| | 4|kamranahmedse/developer-roadmap !2025-03-2831269540|Interactive roadmaps, guides and other educational content to help developers grow in their careers.| | 5|vinta/awesome-python !2025-03-28238581114|A curated list of awesome Python frameworks, libraries, software and resources| | 6|practical-tutorials/project-based-learning !2025-03-28222661124|Curated list of project-based tutorials| | 7|tensorflow/tensorflow !2025-03-281888714|An Open Source Machine Learning Framework for Everyone| | 8|Significant-Gravitas/AutoGPT !2025-03-2817391338|An experimental open-source attempt to make GPT-4 fully autonomous.| | 9|jackfrued/Python-100-Days !2025-03-2816305141|Python - 100天从新手到大师| | 10|AUTOMATIC1111/stable-diffusion-webui !2025-03-2815011553|Stable Diffusion web UI| | 11|huggingface/transformers !2025-03-2814207850|🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.| | 12|ollama/ollama !2025-03-28135166151|Get up and running with Llama 2, Mistral, Gemma, and other large language models.| | 13|f/awesome-chatgpt-prompts !2025-03-2812212738 |This repo includes ChatGPT prompt curation to use ChatGPT better.| | 14|justjavac/free-programming-books-zhCN !2025-03-2811316119|📚 免费的计算机编程类中文书籍,欢迎投稿| | 15|krahets/hello-algo !2025-03-2811107930|《Hello 算法》:动画图解、一键运行的数据结构与算法教程。支持 Python, Java, C++, C, C#, JS, Go, Swift, Rust, Ruby, Kotlin, TS, Dart 代码。简体版和繁体版同步更新,English version ongoing| | 16|yt-dlp/yt-dlp !2025-03-28105801114|A feature-rich command-line audio/video downloader| | 17|langchain-ai/langchain !2025-03-2810449479|⚡ Building applications with LLMs through composability ⚡| | 18|goldbergyoni/nodebestpractices !2025-03-281021629|✅ The Node.js best practices list (July 2024)| | 19|puppeteer/puppeteer !2025-03-289018212|JavaScript API for Chrome and Firefox| | 20|pytorch/pytorch !2025-03-288833938|Tensors and Dynamic neural networks in Python with strong GPU acceleration| | 21|neovim/neovim !2025-03-288781482|Vim-fork focused on extensibility and usability| | 22|🔥🔥langgenius/dify !2025-03-2887342639 |One API for plugins and datasets, one interface for prompt engineering and visual operation, all for creating powerful AI applications.| | 23|mtdvio/every-programmer-should-know !2025-03-28867069|A collection of (mostly) technical things every software developer should know about| | 24|open-webui/open-webui !2025-03-2886025159|User-friendly WebUI for LLMs (Formerly Ollama WebUI)| | 25|ChatGPTNextWeb/NextChat !2025-03-288231521|✨ Light and Fast AI Assistant. Support: Web | | 26|supabase/supabase !2025-03-287990956|The open source Firebase alternative.| | 27|openai/whisper !2025-03-287905542|Robust Speech Recognition via Large-Scale Weak Supervision| | 28|home-assistant/core !2025-03-287773219|🏡 Open source home automation that puts local control and privacy first.| | 29|tensorflow/models !2025-03-28774694|Models and examples built with TensorFlow| | 30| ggerganov/llama.cpp !2025-03-287731836 | Port of Facebook's LLaMA model in C/C++ | | 31|3b1b/manim !2025-03-287641918|Animation engine for explanatory math videos| | 32|microsoft/generative-ai-for-beginners !2025-03-287623860|12 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/| | 33|nomic-ai/gpt4all !2025-03-28729285 |gpt4all: an ecosystem of open-source chatbots trained on a massive collection of clean assistant data including code, stories and dialogue| | 34|comfyanonymous/ComfyUI !2025-03-2872635111|The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.| | 35|bregman-arie/devops-exercises !2025-03-2872225209|Linux, Jenkins, AWS, SRE, Prometheus, Docker, Python, Ansible, Git, Kubernetes, Terraform, OpenStack, SQL, NoSQL, Azure, GCP, DNS, Elastic, Network, Virtualization. DevOps Interview Questions| | 36|elastic/elasticsearch !2025-03-28721419|Free and Open, Distributed, RESTful Search Engine| | 37|🔥n8n-io/n8n !2025-03-2872093495|Free and source-available fair-code licensed workflow automation tool. Easily automate tasks across different services.| | 38|fighting41love/funNLP !2025-03-287200422|The Most Powerful NLP-Weapon Arsenal| | 39|hoppscotch/hoppscotch !2025-03-287060134|Open source API development ecosystem - https://hoppscotch.io (open-source alternative to Postman, Insomnia)| | 40|abi/screenshot-to-code !2025-03-286932817|Drop in a screenshot and convert it to clean HTML/Tailwind/JS code| | 41|binary-husky/gptacademic !2025-03-28680374|Academic Optimization of GPT| | 42|d2l-ai/d2l-zh !2025-03-286774142|Targeting Chinese readers, functional and open for discussion. The Chinese and English versions are used for teaching in over 400 universities across more than 60 countries| | 43|josephmisiti/awesome-machine-learning !2025-03-286739215|A curated list of awesome Machine Learning frameworks, libraries and software.| | 44|grafana/grafana !2025-03-286725414|The open and composable observability and data visualization platform. Visualize metrics, logs, and traces from multiple sources like Prometheus, Loki, Elasticsearch, InfluxDB, Postgres and many more.| | 45|python/cpython !2025-03-286602218|The Python programming language| | 46|apache/superset !2025-03-286519020|Apache Superset is a Data Visualization and Data Exploration Platform| | 47|xtekky/gpt4free !2025-03-28639391 |decentralizing the Ai Industry, free gpt-4/3.5 scripts through several reverse engineered API's ( poe.com, phind.com, chat.openai.com etc...)| | 48|sherlock-project/sherlock !2025-03-286332536|Hunt down social media accounts by username across social networks| | 49|twitter/the-algorithm !2025-03-28630586 |Source code for Twitter's Recommendation Algorithm| | 50|keras-team/keras !2025-03-28627835|Deep Learning for humans| | 51|openai/openai-cookbook !2025-03-28625136 |Examples and guides for using the OpenAI API| | 52|immich-app/immich !2025-03-286238670|High performance self-hosted photo and video management solution.| | 53|AppFlowy-IO/AppFlowy !2025-03-286173528|Bring projects, wikis, and teams together with AI. AppFlowy is an AI collaborative workspace where you achieve more without losing control of your data. The best open source alternative to Notion.| | 54|scikit-learn/scikit-learn !2025-03-286158212|scikit-learn: machine learning in Python| | 55|binhnguyennus/awesome-scalability !2025-03-286117021|The Patterns of Scalable, Reliable, and Performant Large-Scale Systems| | 56|labmlai/annotateddeeplearningpaperimplementations !2025-03-285951726|🧑‍🏫 59 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠| | 57|OpenInterpreter/open-interpreter !2025-03-285894710|A natural language interface for computers| | 58|lobehub/lobe-chat !2025-03-285832054|🤖 Lobe Chat - an open-source, extensible (Function Calling), high-performance chatbot framework. It supports one-click free deployment of your private ChatGPT/LLM web application.| | 59|meta-llama/llama !2025-03-28579536|Inference code for Llama models| | 60|nuxt/nuxt !2025-03-28566437|The Intuitive Vue Framework.| | 61|imartinez/privateGPT !2025-03-28555192|Interact with your documents using the power of GPT, 100% privately, no data leaks| | 62|Stirling-Tools/Stirling-PDF !2025-03-285500846|#1 Locally hosted web application that allows you to perform various operations on PDF files| | 63|PlexPt/awesome-chatgpt-prompts-zh !2025-03-285459720|ChatGPT Chinese Training Guide. Guidelines for various scenarios. Learn how to make it listen to you| | 64|dair-ai/Prompt-Engineering-Guide !2025-03-285451025 |🐙 Guides, papers, lecture, notebooks and resources for prompt engineering| | 65|ageitgey/facerecognition !2025-03-28544382|The world's simplest facial recognition api for Python and the command line| | 66|CorentinJ/Real-Time-Voice-Cloning !2025-03-285384814|Clone a voice in 5 seconds to generate arbitrary speech in real-time| | 67|geekan/MetaGPT !2025-03-285375376|The Multi-Agent Meta Programming Framework: Given one line Requirement, return PRD, Design, Tasks, Repo | | 68|gpt-engineer-org/gpt-engineer !2025-03-285367419|Specify what you want it to build, the AI asks for clarification, and then builds it.| | 69|lencx/ChatGPT !2025-03-2853653-3|🔮 ChatGPT Desktop Application (Mac, Windows and Linux)| | 70|deepfakes/faceswap !2025-03-28535672|Deepfakes Software For All| | 71|langflow-ai/langflow !2025-03-285319584|Langflow is a low-code app builder for RAG and multi-agent AI applications. It’s Python-based and agnostic to any model, API, or database.| | 72|commaai/openpilot !2025-03-28529759|openpilot is an operating system for robotics. Currently, it upgrades the driver assistance system on 275+ supported cars.| | 73|clash-verge-rev/clash-verge-rev !2025-03-2852848124|Continuation of Clash Verge - A Clash Meta GUI based on Tauri (Windows, MacOS, Linux)| | 74|All-Hands-AI/OpenHands !2025-03-285150675|🙌 OpenHands: Code Less, Make More| | 75|xai-org/grok-1 !2025-03-28502504|Grok open release| | 76|meilisearch/meilisearch !2025-03-284999122|A lightning-fast search API that fits effortlessly into your apps, websites, and workflow| | 77|🔥browser-use/browser-use !2025-03-2849910294|Make websites accessible for AI agents| | 78|jgthms/bulma !2025-03-28496783|Modern CSS framework based on Flexbox| | 79|facebookresearch/segment-anything !2025-03-284947116|The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.| |!green-up-arrow.svg 80|hacksider/Deep-Live-Cam !2025-03-2848612146|real time face swap and one-click video deepfake with only a single image (uncensored)| |!red-down-arrow 81|mlabonne/llm-course !2025-03-284860934|Course with a roadmap and notebooks to get into Large Language Models (LLMs).| | 82|PaddlePaddle/PaddleOCR !2025-03-284785530|Awesome multilingual OCR toolkits based on PaddlePaddle (practical ultra lightweight OCR system, support 80+ languages recognition, provide data annotation and synthesis tools, support training and deployment among server, mobile, embedded and IoT devices)| | 83|alist-org/alist !2025-03-284732618|🗂️A file list/WebDAV program that supports multiple storages, powered by Gin and Solidjs. / 一个支持多存储的文件列表/WebDAV程序,使用 Gin 和 Solidjs。| | 84|infiniflow/ragflow !2025-03-2847027129|RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.| | 85|Avik-Jain/100-Days-Of-ML-Code !2025-03-284679312|100 Days of ML Coding| | 86|v2ray/v2ray-core !2025-03-28458706|A platform for building proxies to bypass network restrictions.| | 87|hiyouga/LLaMA-Factory !2025-03-284555881|Easy-to-use LLM fine-tuning framework (LLaMA, BLOOM, Mistral, Baichuan, Qwen, ChatGLM)| | 88|Asabeneh/30-Days-Of-Python !2025-03-284544930|30 days of Python programming challenge is a step-by-step guide to learn the Python programming language in 30 days. This challenge may take more than100 days, follow your own pace. These videos may help too: https://www.youtube.com/channel/UC7PNRuno1rzYPb1xLa4yktw| | 89|type-challenges/type-challenges !2025-03-284488511|Collection of TypeScript type challenges with online judge| | 90|lllyasviel/Fooocus !2025-03-284402716|Focus on prompting and generating| | 91|RVC-Boss/GPT-SoVITS !2025-03-284327738|1 min voice data can also be used to train a good TTS model! (few shot voice cloning)| | 92|rasbt/LLMs-from-scratch !2025-03-284320667|Implementing a ChatGPT-like LLM from scratch, step by step| | 93|oobabooga/text-generation-webui !2025-03-284302012 |A gradio web UI for running Large Language Models like LLaMA, llama.cpp, GPT-J, OPT, and GALACTICA.| | 94|vllm-project/vllm !2025-03-2842982102|A high-throughput and memory-efficient inference and serving engine for LLMs| | 95|dani-garcia/vaultwarden !2025-03-284297121|Unofficial Bitwarden compatible server written in Rust, formerly known as bitwarden_rs| | 96|microsoft/autogen !2025-03-284233049|Enable Next-Gen Large Language Model Applications. Join our Discord: https://discord.gg/pAbnFJrkgZ| | 97|jeecgboot/JeecgBoot !2025-03-284205920|🔥「企业级低代码平台」前后端分离架构SpringBoot 2.x/3.x,SpringCloud,Ant Design&Vue3,Mybatis,Shiro,JWT。强大的代码生成器让前后端代码一键生成,无需写任何代码! 引领新的开发模式OnlineCoding->代码生成->手工MERGE,帮助Java项目解决70%重复工作,让开发更关注业务,既能快速提高效率,帮助公司节省成本,同时又不失灵活性。| | 98|Mintplex-Labs/anything-llm !2025-03-284186955|A full-stack application that turns any documents into an intelligent chatbot with a sleek UI and easier way to manage your workspaces.| | 99|THUDM/ChatGLM-6B !2025-03-28410192 |ChatGLM-6B: An Open Bilingual Dialogue Language Model| | 100|hpcaitech/ColossalAI !2025-03-28406902|Making large AI models cheaper, faster and more accessible| | 101|Stability-AI/stablediffusion !2025-03-28406337|High-Resolution Image Synthesis with Latent Diffusion Models| | 102|mingrammer/diagrams !2025-03-28405063|🎨 Diagram as Code for prototyping cloud system architectures| | 103|Kong/kong !2025-03-28404616|🦍 The Cloud-Native API Gateway and AI Gateway.| | 104|getsentry/sentry !2025-03-284040913|Developer-first error tracking and performance monitoring| | 105| karpathy/nanoGPT !2025-03-284034613 |The simplest, fastest repository for training/finetuning medium-sized GPTs| | 106|fastlane/fastlane !2025-03-2840014-1|🚀 The easiest way to automate building and releasing your iOS and Android apps| | 107|psf/black !2025-03-28399765|The uncompromising Python code formatter| | 108|OpenBB-finance/OpenBBTerminal !2025-03-283972074 |Investment Research for Everyone, Anywhere.| | 109|2dust/v2rayNG !2025-03-283943415|A V2Ray client for Android, support Xray core and v2fly core| | 110|apache/airflow !2025-03-283937314|Apache Airflow - A platform to programmatically author, schedule, and monitor workflows| | 111|KRTirtho/spotube !2025-03-283902746|🎧 Open source Spotify client that doesn't require Premium nor uses Electron! Available for both desktop & mobile!| | 112|coqui-ai/TTS !2025-03-283889719 |🐸💬 - a deep learning toolkit for Text-to-Speech, battle-tested in research and production| | 113|ggerganov/whisper.cpp !2025-03-283882116|Port of OpenAI's Whisper model in C/C++| | 114|ultralytics/ultralytics !2025-03-283866951|NEW - YOLOv8 🚀 in PyTorch > ONNX > OpenVINO > CoreML > TFLite| | 115|typst/typst !2025-03-283863914|A new markup-based typesetting system that is powerful and easy to learn.| | 116|streamlit/streamlit !2025-03-283845828|Streamlit — A faster way to build and share data apps.| | 117|LC044/WeChatMsg !2025-03-283836931|提取微信聊天记录,将其导出成HTML、Word、Excel文档永久保存,对聊天记录进行分析生成年度聊天报告,用聊天数据训练专属于个人的AI聊天助手| | 118|lm-sys/FastChat !2025-03-283822112 |An open platform for training, serving, and evaluating large languages. Release repo for Vicuna and FastChat-T5.| | 119|NaiboWang/EasySpider !2025-03-283819013|A visual no-code/code-free web crawler/spider易采集:一个可视化浏览器自动化测试/数据采集/爬虫软件,可以无代码图形化的设计和执行爬虫任务。别名:ServiceWrapper面向Web应用的智能化服务封装系统。| | 120|microsoft/DeepSpeed !2025-03-283765816 |A deep learning optimization library that makes distributed training and inference easy, efficient, and effective| | 121|QuivrHQ/quivr !2025-03-28376067|Your GenAI Second Brain 🧠 A personal productivity assistant (RAG) ⚡️🤖 Chat with your docs (PDF, CSV, ...) & apps using Langchain, GPT 3.5 / 4 turbo, Private, Anthropic, VertexAI, Ollama, LLMs, that you can share with users ! Local & Private alternative to OpenAI GPTs & ChatGPT powered by retrieval-augmented generation.| | 122|freqtrade/freqtrade !2025-03-283757817 |Free, open source crypto trading bot| | 123|suno-ai/bark !2025-03-28373178 |🔊 Text-Prompted Generative Audio Model| | 124|🔥cline/cline !2025-03-2837307282|Autonomous coding agent right in your IDE, capable of creating/editing files, executing commands, and more with your permission every step of the way.| | 125|LAION-AI/Open-Assistant !2025-03-28372712 |OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.| | 126|penpot/penpot !2025-03-283716217|Penpot: The open-source design tool for design and code collaboration| | 127|gradio-app/gradio !2025-03-283713320|Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work!| | 128|FlowiseAI/Flowise !2025-03-283667135 |Drag & drop UI to build your customized LLM flow using LangchainJS| | 129|SimplifyJobs/Summer2025-Internships !2025-03-28366506|Collection of Summer 2025 tech internships!| | 130|TencentARC/GFPGAN !2025-03-28365027 |GFPGAN aims at developing Practical Algorithms for Real-world Face Restoration.| | 131|ray-project/ray !2025-03-283626819|Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a toolkit of libraries (Ray AIR) for accelerating ML workloads.| | 132|babysor/MockingBird !2025-03-28360498|🚀AI拟声: 5秒内克隆您的声音并生成任意语音内容 Clone a voice in 5 seconds to generate arbitrary speech in real-time| | 133|unslothai/unsloth !2025-03-283603691|5X faster 50% less memory LLM finetuning| | 134|zhayujie/chatgpt-on-wechat !2025-03-283600124 |Wechat robot based on ChatGPT, which uses OpenAI api and itchat library| | 135|upscayl/upscayl !2025-03-283599824|🆙 Upscayl - Free and Open Source AI Image Upscaler for Linux, MacOS and Windows built with Linux-First philosophy.| | 136|freeCodeCamp/devdocs !2025-03-28359738|API Documentation Browser| | 137|XingangPan/DragGAN !2025-03-28359043 |Code for DragGAN (SIGGRAPH 2023)| | 138|2noise/ChatTTS !2025-03-283543922|ChatTTS is a generative speech model for daily dialogue.| | 139|google-research/google-research !2025-03-28352207 |Google Research| | 140|karanpratapsingh/system-design !2025-03-28351003|Learn how to design systems at scale and prepare for system design interviews| | 141|lapce/lapce !2025-03-28350855|Lightning-fast and Powerful Code Editor written in Rust| | 142| microsoft/TaskMatrix !2025-03-2834500-3 | Talking, Drawing and Editing with Visual Foundation Models| | 143|chatchat-space/Langchain-Chatchat !2025-03-283442020|Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM) QA app with langchain| | 144|unclecode/crawl4ai !2025-03-283434163|🔥🕷️ Crawl4AI: Open-source LLM Friendly Web Crawler & Scrapper| | 145|Bin-Huang/chatbox !2025-03-283374733 |A desktop app for GPT-4 / GPT-3.5 (OpenAI API) that supports Windows, Mac & Linux| | 146|milvus-io/milvus !2025-03-283366525 |A cloud-native vector database, storage for next generation AI applications| | 147|mendableai/firecrawl !2025-03-2833297128|🔥 Turn entire websites into LLM-ready markdown| | 148|pola-rs/polars !2025-03-283269320|Fast multi-threaded, hybrid-out-of-core query engine focussing on DataFrame front-ends| | 149|Pythagora-io/gpt-pilot !2025-03-28325321|PoC for a scalable dev tool that writes entire apps from scratch while the developer oversees the implementation| | 150|hashicorp/vault !2025-03-28320797|A tool for secrets management, encryption as a service, and privileged access management| | 151|shardeum/shardeum !2025-03-28319580|Shardeum is an EVM based autoscaling blockchain| | 152|Chanzhaoyu/chatgpt-web !2025-03-28319242 |A demonstration website built with Express and Vue3 called ChatGPT| | 153|lllyasviel/ControlNet !2025-03-283186413 |Let us control diffusion models!| | 154|google/jax !2025-03-28317727|Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more| | 155|facebookresearch/detectron2 !2025-03-28315987|Detectron2 is a platform for object detection, segmentation and other visual recognition tasks.| | 156|myshell-ai/OpenVoice !2025-03-28315233|Instant voice cloning by MyShell| | 157|TheAlgorithms/C-Plus-Plus !2025-03-283151411|Collection of various algorithms in mathematics, machine learning, computer science and physics implemented in C++ for educational purposes.| | 158|hiroi-sora/Umi-OCR !2025-03-283138129|OCR图片转文字识别软件,完全离线。截屏/批量导入图片,支持多国语言、合并段落、竖排文字。可排除水印区域,提取干净的文本。基于 PaddleOCR 。| | 159|mudler/LocalAI !2025-03-283127815|🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.| | 160|facebookresearch/fairseq !2025-03-28312124 |Facebook AI Research Sequence-to-Sequence Toolkit written in Python.| | 161|alibaba/nacos !2025-03-28310559|an easy-to-use dynamic service discovery, configuration and service management platform for building cloud native applications.| | 162|yunjey/pytorch-tutorial !2025-03-28310326|PyTorch Tutorial for Deep Learning Researchers| | 163|v2fly/v2ray-core !2025-03-28307448|A platform for building proxies to bypass network restrictions.| | 164|mckaywrigley/chatbot-ui !2025-03-283067714|The open-source AI chat interface for everyone.| | 165|TabbyML/tabby !2025-03-28305949 |Self-hosted AI coding assistant| | 166|deepseek-ai/awesome-deepseek-integration !2025-03-283053193|| | 167|danielmiessler/fabric !2025-03-283028914|fabric is an open-source framework for augmenting humans using AI.| | 168|xinntao/Real-ESRGAN !2025-03-283026623 |Real-ESRGAN aims at developing Practical Algorithms for General Image/Video Restoration.| | 169|paul-gauthier/aider !2025-03-283014642|aider is GPT powered coding in your terminal| | 170|tatsu-lab/stanfordalpaca !2025-03-28299022 |Code and documentation to train Stanford's Alpaca models, and generate the data.| | 171|DataTalksClub/data-engineering-zoomcamp !2025-03-282971817|Free Data Engineering course!| | 172|HeyPuter/puter !2025-03-282967014|🌐 The Internet OS! Free, Open-Source, and Self-Hostable.| | 173|mli/paper-reading !2025-03-282962314|Classic Deep Learning and In-Depth Reading of New Papers Paragraph by Paragraph| | 174|linexjlin/GPTs !2025-03-28295568|leaked prompts of GPTs| | 175|s0md3v/roop !2025-03-28295286 |one-click deepfake (face swap)| | 176|JushBJJ/Mr.-Ranedeer-AI-Tutor !2025-03-2829465-1 |A GPT-4 AI Tutor Prompt for customizable personalized learning experiences.| | 177|opendatalab/MinerU !2025-03-282927074|A one-stop, open-source, high-quality data extraction tool, supports PDF/webpage/e-book extraction.一站式开源高质量数据提取工具,支持PDF/网页/多格式电子书提取。| | 178|mouredev/Hello-Python !2025-03-282920720|Curso para aprender el lenguaje de programación Python desde cero y para principiantes. 75 clases, 37 horas en vídeo, código, proyectos y grupo de chat. Fundamentos, frontend, backend, testing, IA...| | 179|Lightning-AI/pytorch-lightning !2025-03-28292039|Pretrain, finetune and deploy AI models on multiple GPUs, TPUs with zero code changes.| | 180|crewAIInc/crewAI !2025-03-282919344|Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.| | 181|facebook/folly !2025-03-282916612|An open-source C++ library developed and used at Facebook.| | 182|google-ai-edge/mediapipe !2025-03-28291519|Cross-platform, customizable ML solutions for live and streaming media.| | 183| getcursor/cursor !2025-03-282892025 | An editor made for programming with AI| | 184|chatanywhere/GPTAPIfree !2025-03-282856424|Free ChatGPT API Key, Free ChatGPT API, supports GPT-4 API (free), ChatGPT offers a free domestic forwarding API that allows direct connections without the need for a proxy. It can be used in conjunction with software/plugins like ChatBox, significantly reducing interface usage costs. Enjoy unlimited and unrestricted chatting within China| | 185|meta-llama/llama3 !2025-03-28285552|The official Meta Llama 3 GitHub site| | 186|tinygrad/tinygrad !2025-03-282845811|You like pytorch? You like micrograd? You love tinygrad! ❤️| | 187|google-research/tuningplaybook !2025-03-282841514|A playbook for systematically maximizing the performance of deep learning models.| | 188|huggingface/diffusers !2025-03-282830222|🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.| | 189|tokio-rs/tokio !2025-03-28282408|A runtime for writing reliable asynchronous applications with Rust. Provides I/O, networking, scheduling, timers, ...| | 190|RVC-Project/Retrieval-based-Voice-Conversion-WebUI !2025-03-282823817|Voice data !2025-03-282822612|Jan is an open source alternative to ChatGPT that runs 100% offline on your computer| | 192|openai/CLIP !2025-03-282814720|CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image| | 193|🔥khoj-ai/khoj !2025-03-2828112313|Your AI second brain. A copilot to get answers to your questions, whether they be from your own notes or from the internet. Use powerful, online (e.g gpt4) or private, local (e.g mistral) LLMs. Self-host locally or use our web app. Access from Obsidian, Emacs, Desktop app, Web or Whatsapp.| | 194| acheong08/ChatGPT !2025-03-2828054-2 | Reverse engineered ChatGPT API | | 195|iperov/DeepFaceLive !2025-03-28279345 |Real-time face swap for PC streaming or video calls| | 196|eugeneyan/applied-ml !2025-03-28278471|📚 Papers & tech blogs by companies sharing their work on data science & machine learning in production.| | 197|XTLS/Xray-core !2025-03-282778213|Xray, Penetrates Everything. Also the best v2ray-core, with XTLS support. Fully compatible configuration.| | 198|feder-cr/JobsApplierAIAgent !2025-03-282776410|AutoJobsApplierAI_Agent aims to easy job hunt process by automating the job application process. Utilizing artificial intelligence, it enables users to apply for multiple jobs in an automated and personalized way.| | 199|mindsdb/mindsdb !2025-03-282750631|The platform for customizing AI from enterprise data| | 200|DataExpert-io/data-engineer-handbook !2025-03-282721611|This is a repo with links to everything you'd ever want to learn about data engineering| | 201|exo-explore/exo !2025-03-282721633|Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚| | 202|taichi-dev/taichi !2025-03-2826926-1|Productive, portable, and performant GPU programming in Python.| | 203|mem0ai/mem0 !2025-03-282689134|The memory layer for Personalized AI| | 204|svc-develop-team/so-vits-svc !2025-03-28268096 |SoftVC VITS Singing Voice Conversion| | 205|OpenBMB/ChatDev !2025-03-28265624|Create Customized Software using Natural Language Idea (through Multi-Agent Collaboration)| | 206|roboflow/supervision !2025-03-282632010|We write your reusable computer vision tools. 💜| | 207|drawdb-io/drawdb !2025-03-282626913|Free, simple, and intuitive online database design tool and SQL generator.| | 208|karpathy/llm.c !2025-03-28261633|LLM training in simple, raw C/CUDA| | 209|airbnb/lottie-ios !2025-03-28261431|An iOS library to natively render After Effects vector animations| | 210|openai/openai-python !2025-03-282607713|The OpenAI Python library provides convenient access to the OpenAI API from applications written in the Python language.| | 211|academic/awesome-datascience !2025-03-28259876|📝 An awesome Data Science repository to learn and apply for real world problems.| | 212|harry0703/MoneyPrinterTurbo !2025-03-282576618|Generate short videos with one click using a large model| | 213|gabime/spdlog !2025-03-282571511|Fast C++ logging library.| | 214|ocrmypdf/OCRmyPDF !2025-03-2825674217|OCRmyPDF adds an OCR text layer to scanned PDF files, allowing them to be searched| | 215|Vision-CAIR/MiniGPT-4 !2025-03-28256170 |Enhancing Vision-language Understanding with Advanced Large Language Models| | 216|Stability-AI/generative-models !2025-03-28255936|Generative Models by Stability AI| | 217|DS4SD/docling !2025-03-282555662|Get your docs ready for gen AI| | 218|PostHog/posthog !2025-03-282533227|🦔 PostHog provides open-source product analytics, session recording, feature flagging and A/B testing that you can self-host.| | 219|nrwl/nx !2025-03-282509612|Smart Monorepos · Fast CI| | 220|continuedev/continue !2025-03-282500737|⏩ the open-source copilot chat for software development—bring the power of ChatGPT to VS Code| | 221|opentofu/opentofu !2025-03-28247968|OpenTofu lets you declaratively manage your cloud infrastructure.| | 222|invoke-ai/InvokeAI !2025-03-28247293|InvokeAI is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The solution offers an industry leading WebUI, supports terminal use through a CLI, and serves as the foundation for multiple commercial products.| | 223|deepinsight/insightface !2025-03-282471615 |State-of-the-art 2D and 3D Face Analysis Project| | 224|apache/flink !2025-03-28246865|Apache Flink| | 225|ComposioHQ/composio !2025-03-28246436|Composio equips agents with well-crafted tools empowering them to tackle complex tasks| | 226|Genesis-Embodied-AI/Genesis !2025-03-282458314|A generative world for general-purpose robotics & embodied AI learning.| | 227|stretchr/testify !2025-03-28243184|A toolkit with common assertions and mocks that plays nicely with the standard library| | 228| yetone/openai-translator !2025-03-28242921 | Browser extension and cross-platform desktop application for translation based on ChatGPT API | | 229|frappe/erpnext !2025-03-282425211|Free and Open Source Enterprise Resource Planning (ERP)| | 230|songquanpeng/one-api !2025-03-282410034|OpenAI 接口管理 & 分发系统,支持 Azure、Anthropic Claude、Google PaLM 2 & Gemini、智谱 ChatGLM、百度文心一言、讯飞星火认知、阿里通义千问、360 智脑以及腾讯混元,可用于二次分发管理 key,仅单可执行文件,已打包好 Docker 镜像,一键部署,开箱即用. OpenAI key management & redistribution system, using a single API for all LLMs, and features an English UI.| | 231| microsoft/JARVIS !2025-03-28240604 | a system to connect LLMs with ML community | | 232|google/flatbuffers !2025-03-28239965|FlatBuffers: Memory Efficient Serialization Library| | 233|microsoft/graphrag !2025-03-282398928|A modular graph-based Retrieval-Augmented Generation (RAG) system| | 234|rancher/rancher !2025-03-28239675|Complete container management platform| | 235|bazelbuild/bazel !2025-03-282384618|a fast, scalable, multi-language and extensible build system| | 236|modularml/mojo !2025-03-28238236 |The Mojo Programming Language| | 237|danny-avila/LibreChat !2025-03-282378753|Enhanced ChatGPT Clone: Features OpenAI, GPT-4 Vision, Bing, Anthropic, OpenRouter, Google Gemini, AI model switching, message search, langchain, DALL-E-3, ChatGPT Plugins, OpenAI Functions, Secure Multi-User System, Presets, completely open-source for self-hosting. More features in development| |!green-up-arrow.svg 238|🔥🔥🔥Shubhamsaboo/awesome-llm-apps !2025-03-28237391211|Collection of awesome LLM apps with RAG using OpenAI, Anthropic, Gemini and opensource models.| |!red-down-arrow 239|microsoft/semantic-kernel !2025-03-282373611|Integrate cutting-edge LLM technology quickly and easily into your apps| |!red-down-arrow 240|TheAlgorithms/Rust !2025-03-28236995|All Algorithms implemented in Rust| | 241|stanford-oval/storm !2025-03-28236326|An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.| | 242|openai/gpt-2 !2025-03-28232483|Code for the paper "Language Models are Unsupervised Multitask Learners"| | 243|labring/FastGPT !2025-03-282319445|A platform that uses the OpenAI API to quickly build an AI knowledge base, supporting many-to-many relationships.| | 244|pathwaycom/llm-app !2025-03-2822928-10|Ready-to-run cloud templates for RAG, AI pipelines, and enterprise search with live data. 🐳Docker-friendly.⚡Always in sync with Sharepoint, Google Drive, S3, Kafka, PostgreSQL, real-time data APIs, and more.| | 245|warpdotdev/Warp !2025-03-282286825|Warp is a modern, Rust-based terminal with AI built in so you and your team can build great software, faster.| | 246|🔥agno-agi/agno !2025-03-2822833298|Agno is a lightweight library for building Multimodal Agents. It exposes LLMs as a unified API and gives them superpowers like memory, knowledge, tools and reasoning.| | 247|qdrant/qdrant !2025-03-282275214 |Qdrant - Vector Database for the next generation of AI applications. Also available in the cloud https://cloud.qdrant.io/| | 248|ashishpatel26/500-AI-Machine-learning-Deep-learning-Computer-vision-NLP-Projects-with-code !2025-03-282271815|500 AI Machine learning Deep learning Computer vision NLP Projects with code| | 249|stanfordnlp/dspy !2025-03-282268321|Stanford DSPy: The framework for programming—not prompting—foundation models| | 250|PaddlePaddle/Paddle !2025-03-28226246|PArallel Distributed Deep LEarning: Machine Learning Framework from Industrial Practice (『飞桨』核心框架,深度学习&机器学习高性能单机、分布式训练和跨平台部署)| | 251|zulip/zulip !2025-03-28225464|Zulip server and web application. Open-source team chat that helps teams stay productive and focused.| | 252|Hannibal046/Awesome-LLM !2025-03-282240721|Awesome-LLM: a curated list of Large Language Model| | 253|facefusion/facefusion !2025-03-282218812|Next generation face swapper and enhancer| | 254|Mozilla-Ocho/llamafile !2025-03-28220624|Distribute and run LLMs with a single file.| | 255|yuliskov/SmartTube !2025-03-282201614|SmartTube - an advanced player for set-top boxes and tvs running Android OS| | 256|haotian-liu/LLaVA !2025-03-282201316 |Large Language-and-Vision Assistant built towards multimodal GPT-4 level capabilities.| | 257|ashishps1/awesome-system-design-resources !2025-03-282189367|This repository contains System Design resources which are useful while preparing for interviews and learning Distributed Systems| | 258|Cinnamon/kotaemon !2025-03-28218248|An open-source RAG-based tool for chatting with your documents.| | 259|CodePhiliaX/Chat2DB !2025-03-282179757|🔥🔥🔥AI-driven database tool and SQL client, The hottest GUI client, supporting MySQL, Oracle, PostgreSQL, DB2, SQL Server, DB2, SQLite, H2, ClickHouse, and more.| | 260|blakeblackshear/frigate !2025-03-282177113|NVR with realtime local object detection for IP cameras| | 261|facebookresearch/audiocraft !2025-03-28217111|Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable music generation LM with textual and melodic conditioning.| | 262|karpathy/minGPT !2025-03-28216567|A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training| | 263|grpc/grpc-go !2025-03-282159510|The Go language implementation of gRPC. HTTP/2 based RPC| | 264|HumanSignal/label-studio !2025-03-282137618|Label Studio is a multi-type data labeling and annotation tool with standardized output format| | 265|yoheinakajima/babyagi !2025-03-28212764 |uses OpenAI and Pinecone APIs to create, prioritize, and execute tasks, This is a pared-down version of the original Task-Driven Autonomous Agent| | 266|deepseek-ai/DeepSeek-Coder !2025-03-282118210|DeepSeek Coder: Let the Code Write Itself| | 267|BuilderIO/gpt-crawler !2025-03-282118010|Crawl a site to generate knowledge files to create your own custom GPT from a URL| | 268| openai/chatgpt-retrieval-plugin !2025-03-2821152-1 | Plugins are chat extensions designed specifically for language models like ChatGPT, enabling them to access up-to-date information, run computations, or interact with third-party services in response to a user's request.| | 269|microsoft/OmniParser !2025-03-282113123|A simple screen parsing tool towards pure vision based GUI agent| | 270|black-forest-labs/flux !2025-03-282107219|Official inference repo for FLUX.1 models| | 271|ItzCrazyKns/Perplexica !2025-03-282099154|Perplexica is an AI-powered search engine. It is an Open source alternative to Perplexity AI| | 272|microsoft/unilm !2025-03-28209876|Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities| | 273|Sanster/lama-cleaner !2025-03-282077614|Image inpainting tool powered by SOTA AI Model. Remove any unwanted object, defect, people from your pictures or erase and replace(powered by stable diffusion) any thing on your pictures.| | 274|assafelovic/gpt-researcher !2025-03-282057222|GPT based autonomous agent that does online comprehensive research on any given topic| | 275|PromtEngineer/localGPT !2025-03-28204230 |Chat with your documents on your local device using GPT models. No data leaves your device and 100% private.| | 276|elastic/kibana !2025-03-28203482|Your window into the Elastic Stack| | 277|fishaudio/fish-speech !2025-03-282033222|Brand new TTS solution| | 278|mlc-ai/mlc-llm !2025-03-282028110 |Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.| | 279|deepset-ai/haystack !2025-03-282005320|🔍 Haystack is an open source NLP framework to interact with your data using Transformer models and LLMs (GPT-4, ChatGPT and alike). Haystack offers production-ready tools to quickly build complex question answering, semantic search, text generation applications, and more.| | 280|tree-sitter/tree-sitter !2025-03-28200487|An incremental parsing system for programming tools| | 281|Anjok07/ultimatevocalremovergui !2025-03-281999811|GUI for a Vocal Remover that uses Deep Neural Networks.| | 282|guidance-ai/guidance !2025-03-28199622|A guidance language for controlling large language models.| | 283|ml-explore/mlx !2025-03-28199619|MLX: An array framework for Apple silicon| | 284|mlflow/mlflow !2025-03-281995314|Open source platform for the machine learning lifecycle| | 285|ml-tooling/best-of-ml-python !2025-03-28198631|🏆 A ranked list of awesome machine learning Python libraries. Updated weekly.| | 286|BerriAI/litellm !2025-03-281981862|Call all LLM APIs using the OpenAI format. Use Bedrock, Azure, OpenAI, Cohere, Anthropic, Ollama, Sagemaker, HuggingFace, Replicate (100+ LLMs)| | 287|LazyVim/LazyVim !2025-03-281981320|Neovim config for the lazy| | 288|wez/wezterm !2025-03-281976018|A GPU-accelerated cross-platform terminal emulator and multiplexer written by @wez and implemented in Rust| | 289|valkey-io/valkey !2025-03-281970416|A flexible distributed key-value datastore that supports both caching and beyond caching workloads.| | 290|LiLittleCat/awesome-free-chatgpt !2025-03-28196185|🆓免费的 ChatGPT 镜像网站列表,持续更新。List of free ChatGPT mirror sites, continuously updated.| | 291|Byaidu/PDFMathTranslate !2025-03-281947645|PDF scientific paper translation with preserved formats - 基于 AI 完整保留排版的 PDF 文档全文双语翻译,支持 Google/DeepL/Ollama/OpenAI 等服务,提供 CLI/GUI/Docker| | 292|openai/swarm !2025-03-281947111|Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.| | 293|HqWu-HITCS/Awesome-Chinese-LLM !2025-03-281921423|Organizing smaller, cost-effective, privately deployable open-source Chinese language models, including related datasets and tutorials| | 294|stitionai/devika !2025-03-28190903|Devika is an Agentic AI Software Engineer that can understand high-level human instructions, break them down into steps, research relevant information, and write code to achieve the given objective. Devika aims to be a competitive open-source alternative to Devin by Cognition AI.| | 295|OpenBMB/MiniCPM-o !2025-03-28190887|MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone| | 296|samber/lo !2025-03-281904815|💥 A Lodash-style Go library based on Go 1.18+ Generics (map, filter, contains, find...)| | 297|chroma-core/chroma !2025-03-281895221 |the AI-native open-source embedding database| | 298|DarkFlippers/unleashed-firmware !2025-03-28189278|Flipper Zero Unleashed Firmware| | 299|brave/brave-browser !2025-03-281892710|Brave browser for Android, iOS, Linux, macOS, Windows.| | 300| tloen/alpaca-lora !2025-03-28188641 | Instruct-tune LLaMA on consumer hardware| | 301|VinciGit00/Scrapegraph-ai !2025-03-281884618|Python scraper based on AI| | 302|gitroomhq/postiz-app !2025-03-281879110|📨 Schedule social posts, measure them, exchange with other members and get a lot of help from AI 🚀| | 303|PrefectHQ/prefect !2025-03-281878715|Prefect is a workflow orchestration tool empowering developers to build, observe, and react to data pipelines| | 304|ymcui/Chinese-LLaMA-Alpaca !2025-03-28187723 |Chinese LLaMA & Alpaca LLMs| | 305|kenjihiranabe/The-Art-of-Linear-Algebra !2025-03-28187335|Graphic notes on Gilbert Strang's "Linear Algebra for Everyone"| | 306|joonspk-research/generativeagents !2025-03-28187288|Generative Agents: Interactive Simulacra of Human Behavior| | 307|renovatebot/renovate !2025-03-28186820|Universal dependency update tool that fits into your workflows.| | 308|gventuri/pandas-ai !2025-03-28186109 |Pandas AI is a Python library that integrates generative artificial intelligence capabilities into Pandas, making dataframes conversational| | 309|thingsboard/thingsboard !2025-03-28185184|Open-source IoT Platform - Device management, data collection, processing and visualization.| | 310|ente-io/ente !2025-03-28184722|Fully open source, End to End Encrypted alternative to Google Photos and Apple Photos| | 311|serengil/deepface !2025-03-281840113|A Lightweight Face Recognition and Facial Attribute Analysis (Age, Gender, Emotion and Race) Library for Python| | 312|Raphire/Win11Debloat !2025-03-281840132|A simple, easy to use PowerShell script to remove pre-installed apps from windows, disable telemetry, remove Bing from windows search as well as perform various other changes to declutter and improve your windows experience. This script works for both windows 10 and windows 11.| | 313|Avaiga/taipy !2025-03-28179235|Turns Data and AI algorithms into production-ready web applications in no time.| | 314|microsoft/qlib !2025-03-281784231|Qlib is an AI-oriented quantitative investment platform that aims to realize the potential, empower research, and create value using AI technologies in quantitative investment, from exploring ideas to implementing productions. Qlib supports diverse machine learning modeling paradigms. including supervised learning, market dynamics modeling, and RL.| | 315|CopilotKit/CopilotKit !2025-03-281778571|Build in-app AI chatbots 🤖, and AI-powered Textareas ✨, into react web apps.| | 316|QwenLM/Qwen-7B !2025-03-281766017|The official repo of Qwen-7B (通义千问-7B) chat & pretrained large language model proposed by Alibaba Cloud.| | 317|w-okada/voice-changer !2025-03-28176078 |リアルタイムボイスチェンジャー Realtime Voice Changer| | 318|rlabbe/Kalman-and-Bayesian-Filters-in-Python !2025-03-281756011|Kalman Filter book using Jupyter Notebook. Focuses on building intuition and experience, not formal proofs. Includes Kalman filters,extended Kalman filters, unscented Kalman filters, particle filters, and more. All exercises include solutions.| | 319|Mikubill/sd-webui-controlnet !2025-03-28174794 |WebUI extension for ControlNet| | 320|jingyaogong/minimind !2025-03-2817380116|「大模型」3小时完全从0训练26M的小参数GPT,个人显卡即可推理训练!| | 321|apify/crawlee !2025-03-28172696|Crawlee—A web scraping and browser automation library for Node.js to build reliable crawlers. In JavaScript and TypeScript. Extract data for AI, LLMs, RAG, or GPTs. Download HTML, PDF, JPG, PNG, and other files from websites. Works with Puppeteer, Playwright, Cheerio, JSDOM, and raw HTTP. Both headful and headless mode. With proxy rotation.| | 322|apple/ml-stable-diffusion !2025-03-28172395|Stable Diffusion with Core ML on Apple Silicon| | 323| transitive-bullshit/chatgpt-api !2025-03-28172095 | Node.js client for the official ChatGPT API. | | 324|teableio/teable !2025-03-281719222|✨ The Next Gen Airtable Alternative: No-Code Postgres| | 325| xx025/carrot !2025-03-28170900 | Free ChatGPT Site List | | 326|microsoft/LightGBM !2025-03-28170723|A fast, distributed, high-performance gradient boosting (GBT, GBDT, GBRT, GBM or MART) framework based on decision tree algorithms, used for ranking, classification and many other machine learning tasks.| | 327|VikParuchuri/surya !2025-03-28169827|Accurate line-level text detection and recognition (OCR) in any language| | 328|deepseek-ai/Janus !2025-03-281692825|Janus-Series: Unified Multimodal Understanding and Generation Models| | 329|ardalis/CleanArchitecture !2025-03-28168823|Clean Architecture Solution Template: A starting point for Clean Architecture with ASP.NET Core| | 330|neondatabase/neon !2025-03-28166466|Neon: Serverless Postgres. We separated storage and compute to offer autoscaling, code-like database branching, and scale to zero.| | 331|kestra-io/kestra !2025-03-281661313|⚡ Workflow Automation Platform. Orchestrate & Schedule code in any language, run anywhere, 500+ plugins. Alternative to Zapier, Rundeck, Camunda, Airflow...| | 332|Dao-AILab/flash-attention !2025-03-281659720|Fast and memory-efficient exact attention| | 333|RPCS3/rpcs3 !2025-03-281655712|PS3 emulator/debugger| | 334|meta-llama/llama-recipes !2025-03-28165486|Scripts for fine-tuning Llama2 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization & question answering. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment.Demo apps to showcase Llama2 for WhatsApp & Messenger| | 335|emilwallner/Screenshot-to-code !2025-03-28165180|A neural network that transforms a design mock-up into a static website.| | 336|datawhalechina/llm-cookbook !2025-03-281650922|面向开发者的 LLM 入门教程,吴恩达大模型系列课程中文版| | 337|e2b-dev/awesome-ai-agents !2025-03-281643923|A list of AI autonomous agents| | 338|QwenLM/Qwen2.5 !2025-03-281641114|Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.| | 339|dair-ai/ML-YouTube-Courses !2025-03-28164114|📺 Discover the latest machine learning / AI courses on YouTube.| | 340|pybind/pybind11 !2025-03-28163620|Seamless operability between C++11 and Python| | 341|graphdeco-inria/gaussian-splatting !2025-03-281627116|Original reference implementation of "3D Gaussian Splatting for Real-Time Radiance Field Rendering"| | 342|meta-llama/codellama !2025-03-28162531|Inference code for CodeLlama models| | 343|TransformerOptimus/SuperAGI !2025-03-28161292 | SuperAGI - A dev-first open source autonomous AI agent framework. Enabling developers to build, manage & run useful autonomous agents quickly and reliably.| | 344|microsoft/onnxruntime !2025-03-28161169|ONNX Runtime: cross-platform, high-performance ML inferencing and training accelerator| | 345|IDEA-Research/Grounded-Segment-Anything !2025-03-281601411 |Marrying Grounding DINO with Segment Anything & Stable Diffusion & BLIP - Automatically Detect, Segment and Generate Anything with Image and Text Inputs| | 346|ddbourgin/numpy-ml !2025-03-28160054|Machine learning, in numpy| | 347|eosphoros-ai/DB-GPT !2025-03-281585225|Revolutionizing Database Interactions with Private LLM Technology| | 348|Stability-AI/StableLM !2025-03-28158310 |Stability AI Language Models| | 349|openai/evals !2025-03-28157935 |Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.| | 350|THUDM/ChatGLM2-6B !2025-03-28157500|ChatGLM2-6B: An Open Bilingual Chat LLM | | 351|sunner/ChatALL !2025-03-28156761 |Concurrently chat with ChatGPT, Bing Chat, Bard, Alpaca, Vincuna, Claude, ChatGLM, MOSS, iFlytek Spark, ERNIE and more, discover the best answers| | 352|abseil/abseil-cpp !2025-03-28156656|Abseil Common Libraries (C++)| | 353|NVIDIA/open-gpu-kernel-modules !2025-03-28156531|NVIDIA Linux open GPU kernel module source| | 354|letta-ai/letta !2025-03-281563718|Letta (formerly MemGPT) is a framework for creating LLM services with memory.| | 355|typescript-eslint/typescript-eslint !2025-03-28156211|✨ Monorepo for all the tooling which enables ESLint to support TypeScript| | 356|umijs/umi !2025-03-28156211|A framework in react community ✨| | 357|AI4Finance-Foundation/FinGPT !2025-03-281561215|Data-Centric FinGPT. Open-source for open finance! Revolutionize 🔥 We'll soon release the trained model.| | 358|amplication/amplication !2025-03-28156022|🔥🔥🔥 The Only Production-Ready AI-Powered Backend Code Generation| | 359|KindXiaoming/pykan !2025-03-28155477|Kolmogorov Arnold Networks| | 360|arc53/DocsGPT !2025-03-28154900|GPT-powered chat for documentation, chat with your documents| | 361|influxdata/telegraf !2025-03-28154502|Agent for collecting, processing, aggregating, and writing metrics, logs, and other arbitrary data.| | 362|microsoft/Bringing-Old-Photos-Back-to-Life !2025-03-28154084|Bringing Old Photo Back to Life (CVPR 2020 oral)| | 363|GaiZhenbiao/ChuanhuChatGPT !2025-03-2815394-2|GUI for ChatGPT API and many LLMs. Supports agents, file-based QA, GPT finetuning and query with web search. All with a neat UI.| | 364|Zeyi-Lin/HivisionIDPhotos !2025-03-281529710|⚡️HivisionIDPhotos: a lightweight and efficient AI ID photos tools. 一个轻量级的AI证件照制作算法。| | 365| mayooear/gpt4-pdf-chatbot-langchain !2025-03-281529518 | GPT4 & LangChain Chatbot for large PDF docs | | 366|1Panel-dev/MaxKB !2025-03-2815277148|? Based on LLM large language model knowledge base Q&A system. Ready to use out of the box, supports quick integration into third-party business systems. Officially produced by 1Panel| | 367|ai16z/eliza !2025-03-281526811|Conversational Agent for Twitter and Discord| | 368|apache/arrow !2025-03-28151684|Apache Arrow is a multi-language toolbox for accelerated data interchange and in-memory processing| | 369|princeton-nlp/SWE-agent !2025-03-281516119|SWE-agent: Agent Computer Interfaces Enable Software Engineering Language Models| | 370|mlc-ai/web-llm !2025-03-281509311 |Bringing large-language models and chat to web browsers. Everything runs inside the browser with no server support.| | 371|guillaumekln/faster-whisper !2025-03-281507117 |Faster Whisper transcription with CTranslate2| | 372|overleaf/overleaf !2025-03-28150316|A web-based collaborative LaTeX editor| | 373|triton-lang/triton !2025-03-28150169|Development repository for the Triton language and compiler| | 374|soxoj/maigret !2025-03-281500410|🕵️‍♂️ Collect a dossier on a person by username from thousands of sites| | 375|alibaba/lowcode-engine !2025-03-28149841|An enterprise-class low-code technology stack with scale-out design / 一套面向扩展设计的企业级低代码技术体系| | 376|espressif/esp-idf !2025-03-28148545|Espressif IoT Development Framework. Official development framework for Espressif SoCs.| | 377|pgvector/pgvector !2025-03-281484913|Open-source vector similarity search for Postgres| | 378|datawhalechina/leedl-tutorial !2025-03-28148246|《李宏毅深度学习教程》(李宏毅老师推荐👍),PDF下载地址:https://github.com/datawhalechina/leedl-tutorial/releases| | 379|xcanwin/KeepChatGPT !2025-03-28147972 |Using ChatGPT is more efficient and smoother, perfectly solving ChatGPT network errors. No longer do you need to frequently refresh the webpage, saving over 10 unnecessary steps| | 380|m-bain/whisperX !2025-03-281471313|WhisperX: Automatic Speech Recognition with Word-level Timestamps (& Diarization)| | 381|HumanAIGC/AnimateAnyone !2025-03-2814706-1|Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation| |!green-up-arrow.svg 382|naklecha/llama3-from-scratch !2025-03-281469024|llama3 implementation one matrix multiplication at a time| |!red-down-arrow 383| fauxpilot/fauxpilot !2025-03-28146871 | An open-source GitHub Copilot server | | 384|LlamaFamily/Llama-Chinese !2025-03-28145111|Llama Chinese Community, the best Chinese Llama large model, fully open source and commercially available| | 385|BradyFU/Awesome-Multimodal-Large-Language-Models !2025-03-281450121|Latest Papers and Datasets on Multimodal Large Language Models| | 386|vanna-ai/vanna !2025-03-281449819|🤖 Chat with your SQL database 📊. Accurate Text-to-SQL Generation via LLMs using RAG 🔄.| | 387|bleedline/aimoneyhunter !2025-03-28144845|AI Side Hustle Money Mega Collection: Teaching You How to Utilize AI for Various Side Projects to Earn Extra Income.| | 388|stefan-jansen/machine-learning-for-trading !2025-03-28144629|Code for Machine Learning for Algorithmic Trading, 2nd edition.| | 389|state-spaces/mamba !2025-03-28144139|Mamba: Linear-Time Sequence Modeling with Selective State Spaces| | 390|vercel/ai-chatbot !2025-03-281434614|A full-featured, hackable Next.js AI chatbot built by Vercel| | 391|steven-tey/novel !2025-03-281428410|Notion-style WYSIWYG editor with AI-powered autocompletions| | 392|unifyai/ivy !2025-03-281409348|Unified AI| | 393|chidiwilliams/buzz !2025-03-281402411 |Buzz transcribes and translates audio offline on your personal computer. Powered by OpenAI's Whisper.| | 394|lukas-blecher/LaTeX-OCR !2025-03-28139769|pix2tex: Using a ViT to convert images of equations into LaTeX code.| | 395|openai/tiktoken !2025-03-28139599|tiktoken is a fast BPE tokeniser for use with OpenAI's models.| | 396|nocobase/nocobase !2025-03-281391522|NocoBase is a scalability-first, open-source no-code/low-code platform for building business applications and enterprise solutions.| | 397|neonbjb/tortoise-tts !2025-03-28139010 |A multi-voice TTS system trained with an emphasis on quality| | 398|yamadashy/repomix !2025-03-281382036|📦 Repomix (formerly Repopack) is a powerful tool that packs your entire repository into a single, AI-friendly file. Perfect for when you need to feed your codebase to Large Language Models (LLMs) or other AI tools like Claude, ChatGPT, and Gemini.| | 399|adobe/react-spectrum !2025-03-28136766|A collection of libraries and tools that help you build adaptive, accessible, and robust user experiences.| | 400|THUDM/ChatGLM3 !2025-03-28136684|ChatGLM3 series: Open Bilingual Chat LLMs | | 401|NVIDIA/NeMo !2025-03-28134837|A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)| | 402|BlinkDL/RWKV-LM !2025-03-28134346 |RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it combines the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.| | 403| fuergaosi233/wechat-chatgpt !2025-03-28133330 | Use ChatGPT On Wechat via wechaty | | 404|udecode/plate !2025-03-28133325|A rich-text editor powered by AI| | 405|xenova/transformers.js !2025-03-281331219|State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!| | 406|stas00/ml-engineering !2025-03-281325615|Machine Learning Engineering Guides and Tools| | 407| wong2/chatgpt-google-extension !2025-03-2813241-1 | A browser extension that enhances search engines with ChatGPT, this repos will not be updated from 2023-02-20| | 408|mrdbourke/pytorch-deep-learning !2025-03-281317520|Materials for the Learn PyTorch for Deep Learning: Zero to Mastery course.| | 409|Koenkk/zigbee2mqtt !2025-03-28131544|Zigbee 🐝 to MQTT bridge 🌉, get rid of your proprietary Zigbee bridges 🔨| | 410|vercel-labs/ai !2025-03-281298528|Build AI-powered applications with React, Svelte, and Vue| | 411|netease-youdao/QAnything !2025-03-28129318|Question and Answer based on Anything.| | 412|huggingface/trl !2025-03-281289622|Train transformer language models with reinforcement learning.| | 413|microsoft/BitNet !2025-03-28128503|Official inference framework for 1-bit LLMs| | 414|mediar-ai/screenpipe !2025-03-281283915|24/7 local AI screen & mic recording. Build AI apps that have the full context. Works with Ollama. Alternative to Rewind.ai. Open. Secure. You own your data. Rust.| | 415|Skyvern-AI/skyvern !2025-03-281277612|Automate browser-based workflows with LLMs and Computer Vision| | 416|pytube/pytube !2025-03-28126591|A lightweight, dependency-free Python library (and command-line utility) for downloading YouTube Videos.| | 417|official-stockfish/Stockfish !2025-03-28126574|UCI chess engine| | 418|sgl-project/sglang !2025-03-281260143|SGLang is a structured generation language designed for large language models (LLMs). It makes your interaction with LLMs faster and more controllable.| | 419|plasma-umass/scalene !2025-03-28125535|Scalene: a high-performance, high-precision CPU, GPU, and memory profiler for Python with AI-powered optimization proposals| | 420|danswer-ai/danswer !2025-03-28125503|Ask Questions in natural language and get Answers backed by private sources. Connects to tools like Slack, GitHub, Confluence, etc.| | 421|OpenTalker/SadTalker !2025-03-28125226|[CVPR 2023] SadTalker:Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation| | 422|facebookresearch/AnimatedDrawings !2025-03-28123693 |Code to accompany "A Method for Animating Children's Drawings of the Human Figure"| | 423|activepieces/activepieces !2025-03-28123609|Your friendliest open source all-in-one automation tool ✨ Workflow automation tool 100+ integration / Enterprise automation tool / Zapier Alternative| | 424|ggerganov/ggml !2025-03-28121992 |Tensor library for machine learning| | 425|bytebase/bytebase !2025-03-28121694|World's most advanced database DevOps and CI/CD for Developer, DBA and Platform Engineering teams. The GitLab/GitHub for database DevOps.| | 426| willwulfken/MidJourney-Styles-and-Keywords-Reference !2025-03-28120971 | A reference containing Styles and Keywords that you can use with MidJourney AI| | 427|Huanshere/VideoLingo !2025-03-281207013|Netflix-level subtitle cutting, translation, alignment, and even dubbing - one-click fully automated AI video subtitle team | | 428|OpenLMLab/MOSS !2025-03-28120330 |An open-source tool-augmented conversational language model from Fudan University| | 429|llmware-ai/llmware !2025-03-281200727|Providing enterprise-grade LLM-based development framework, tools, and fine-tuned models.| | 430|PKU-YuanGroup/Open-Sora-Plan !2025-03-28119362|This project aim to reproduce Sora (Open AI T2V model), but we only have limited resource. We deeply wish the all open source community can contribute to this project.| | 431|ShishirPatil/gorilla !2025-03-28119332 |Gorilla: An API store for LLMs| | 432|NVIDIA/Megatron-LM !2025-03-281192716|Ongoing research training transformer models at scale| | 433|illacloud/illa-builder !2025-03-28119192|Create AI-Driven Apps like Assembling Blocks| | 434|marimo-team/marimo !2025-03-281191521|A reactive notebook for Python — run reproducible experiments, execute as a script, deploy as an app, and version with git.| | 435|smol-ai/developer !2025-03-28119111 | With 100k context windows on the way, it's now feasible for every dev to have their own smol developer| | 436|Lightning-AI/litgpt !2025-03-28118878|Pretrain, finetune, deploy 20+ LLMs on your own data. Uses state-of-the-art techniques: flash attention, FSDP, 4-bit, LoRA, and more.| | 437|openai/shap-e !2025-03-28118474 |Generate 3D objects conditioned on text or images| | 438|eugeneyan/open-llms !2025-03-28118451 |A list of open LLMs available for commercial use.| | 439|andrewyng/aisuite !2025-03-28118124|Simple, unified interface to multiple Generative AI providers| | 440|hajimehoshi/ebiten !2025-03-28117816|Ebitengine - A dead simple 2D game engine for Go| | 441|kgrzybek/modular-monolith-with-ddd !2025-03-28117493|Full Modular Monolith application with Domain-Driven Design approach.| | 442|h2oai/h2ogpt !2025-03-2811736-1 |Come join the movement to make the world's best open source GPT led by H2O.ai - 100% private chat and document search, no data leaks, Apache 2.0| | 443|owainlewis/awesome-artificial-intelligence !2025-03-28117332|A curated list of Artificial Intelligence (AI) courses, books, video lectures and papers.| | 444|DataTalksClub/mlops-zoomcamp !2025-03-28116643|Free MLOps course from DataTalks.Club| | 445|Rudrabha/Wav2Lip !2025-03-281163410|This repository contains the codes of "A Lip Sync Expert Is All You Need for Speech to Lip Generation In the Wild", published at ACM Multimedia 2020.| | 446|aishwaryanr/awesome-generative-ai-guide !2025-03-281152810|A one stop repository for generative AI research updates, interview resources, notebooks and much more!| | 447|karpathy/micrograd !2025-03-28115146|A tiny scalar-valued autograd engine and a neural net library on top of it with PyTorch-like API| | 448|InstantID/InstantID !2025-03-28115111|InstantID : Zero-shot Identity-Preserving Generation in Seconds 🔥| | 449|facebookresearch/seamlesscommunication !2025-03-28114434|Foundational Models for State-of-the-Art Speech and Text Translation| | 450|anthropics/anthropic-cookbook !2025-03-281140112|A collection of notebooks/recipes showcasing some fun and effective ways of using Claude.| | 451|mastra-ai/mastra !2025-03-281139240|the TypeScript AI agent framework| | 452|NVIDIA/TensorRT !2025-03-28113864|NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.| | 453|plandex-ai/plandex !2025-03-28113645|An AI coding engine for complex tasks| | 454|RUCAIBox/LLMSurvey !2025-03-28112735 |A collection of papers and resources related to Large Language Models.| | 455|kubeshark/kubeshark !2025-03-28112711|The API traffic analyzer for Kubernetes providing real-time K8s protocol-level visibility, capturing and monitoring all traffic and payloads going in, out and across containers, pods, nodes and clusters. Inspired by Wireshark, purposely built for Kubernetes| | 456|electric-sql/pglite !2025-03-28112617|Lightweight Postgres packaged as WASM into a TypeScript library for the browser, Node.js, Bun and Deno from https://electric-sql.com| | 457|lightaime/camel !2025-03-281124441 |🐫 CAMEL: Communicative Agents for “Mind” Exploration of Large Scale Language Model Society| | 458|huggingface/lerobot !2025-03-281120184|🤗 LeRobot: State-of-the-art Machine Learning for Real-World Robotics in Pytorch| | 459|normal-computing/outlines !2025-03-28111657|Generative Model Programming| | 460|libretro/RetroArch !2025-03-28110701|Cross-platform, sophisticated frontend for the libretro API. Licensed GPLv3.| | 461|THUDM/CogVideo !2025-03-28110599|Text-to-video generation: CogVideoX (2024) and CogVideo (ICLR 2023)| | 462|bentoml/OpenLLM !2025-03-28110495|An open platform for operating large language models (LLMs) in production. Fine-tune, serve, deploy, and monitor any LLMs with ease.| | 463|vosen/ZLUDA !2025-03-28110429|CUDA on AMD GPUs| | 464|dair-ai/ML-Papers-of-the-Week !2025-03-28110304 |🔥Highlighting the top ML papers every week.| | 465|WordPress/gutenberg !2025-03-28110212|The Block Editor project for WordPress and beyond. Plugin is available from the official repository.| | 466|microsoft/data-formulator !2025-03-281099827|🪄 Create rich visualizations with AI| | 467|LibreTranslate/LibreTranslate !2025-03-28109887|Free and Open Source Machine Translation API. Self-hosted, offline capable and easy to setup.| | 468|block/goose !2025-03-281097737|an open-source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM| | 469|getumbrel/llama-gpt !2025-03-28109553|A self-hosted, offline, ChatGPT-like chatbot. Powered by Llama 2. 100% private, with no data leaving your device.| | 470|HigherOrderCO/HVM !2025-03-28109182|A massively parallel, optimal functional runtime in Rust| | 471|databrickslabs/dolly !2025-03-2810812-3 | A large language model trained on the Databricks Machine Learning Platform| | 472|srush/GPU-Puzzles !2025-03-28108014|Solve puzzles. Learn CUDA.| | 473|Z3Prover/z3 !2025-03-28107952|The Z3 Theorem Prover| | 474|UFund-Me/Qbot !2025-03-281079313 |Qbot is an AI-oriented quantitative investment platform, which aims to realize the potential, empower AI technologies in quantitative investment| | 475|langchain-ai/langgraph !2025-03-281077336|| | 476|lz4/lz4 !2025-03-28107647|Extremely Fast Compression algorithm| | 477|magic-research/magic-animate !2025-03-28107160|MagicAnimate: Temporally Consistent Human Image Animation using Diffusion Model| | 478|PaperMC/Paper !2025-03-281071410|The most widely used, high performance Minecraft server that aims to fix gameplay and mechanics inconsistencies| | 479|getomni-ai/zerox !2025-03-281071015|Zero shot pdf OCR with gpt-4o-mini| |!green-up-arrow.svg 480|🔥NirDiamant/GenAIAgents !2025-03-2810693318|This repository provides tutorials and implementations for various Generative AI Agent techniques, from basic to advanced. It serves as a comprehensive guide for building intelligent, interactive AI systems.| |!red-down-arrow 481|Unstructured-IO/unstructured !2025-03-28106889|Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.| | 482|apache/thrift !2025-03-28106610|Apache Thrift| | 483| TheR1D/shellgpt !2025-03-28106097 | A command-line productivity tool powered by ChatGPT, will help you accomplish your tasks faster and more efficiently | | 484|TheRamU/Fay !2025-03-281060312 |Fay is a complete open source project that includes Fay controller and numeral models, which can be used in different applications such as virtual hosts, live promotion, numeral human interaction and so on| | 485|zyronon/douyin !2025-03-28105566|Vue3 + Pinia + Vite5 仿抖音,Vue 在移动端的最佳实践 . Imitate TikTok ,Vue Best practices on Mobile| | 486|THU-MIG/yolov10 !2025-03-28105485|YOLOv10: Real-Time End-to-End Object Detection| | 487|idootop/mi-gpt !2025-03-281052522|? Transform XiaoAi speaker into a personal voice assistant with ChatGPT and DouBao integration.| | 488|SakanaAI/AI-Scientist !2025-03-281051310|The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery 🧑‍🔬| | 489|szimek/sharedrop !2025-03-28105101|Easy P2P file transfer powered by WebRTC - inspired by Apple AirDrop| | 490|salesforce/LAVIS !2025-03-28103942 |LAVIS - A One-stop Library for Language-Vision Intelligence| | 491|aws/amazon-sagemaker-examples !2025-03-28103654|Example 📓 Jupyter notebooks that demonstrate how to build, train, and deploy machine learning models using 🧠 Amazon SageMaker.| | 492|artidoro/qlora !2025-03-28103402 |QLoRA: Efficient Finetuning of Quantized LLMs| | 493|lllyasviel/stable-diffusion-webui-forge !2025-03-281029314| a platform on top of Stable Diffusion WebUI (based on Gradio) to make development easier, optimize resource management, and speed up inference| | 494|NielsRogge/Transformers-Tutorials !2025-03-28102487|This repository contains demos I made with the Transformers library by HuggingFace.| | 495|kedro-org/kedro !2025-03-28102371|Kedro is a toolbox for production-ready data science. It uses software engineering best practices to help you create data engineering and data science pipelines that are reproducible, maintainable, and modular.| | 496| chathub-dev/chathub !2025-03-28102301 | All-in-one chatbot client | | 497|microsoft/promptflow !2025-03-28101612|Build high-quality LLM apps - from prototyping, testing to production deployment and monitoring.| | 498|mistralai/mistral-src !2025-03-28101372|Reference implementation of Mistral AI 7B v0.1 model.| | 499|burn-rs/burn !2025-03-28101183|Burn - A Flexible and Comprehensive Deep Learning Framework in Rust| | 500|AIGC-Audio/AudioGPT !2025-03-28101150 |AudioGPT: Understanding and Generating Speech, Music, Sound, and Talking Head| | 501|facebookresearch/dinov2 !2025-03-281011210 |PyTorch code and models for the DINOv2 self-supervised learning method.| | 502|RockChinQ/LangBot !2025-03-281008455|😎丰富生态、🧩支持扩展、🦄多模态 - 大模型原生即时通信机器人平台 🤖 | | 503|78/xiaozhi-esp32 !2025-03-281008180|Build your own AI friend| | 504|cumulo-autumn/StreamDiffusion !2025-03-28100761|StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation| | 505|DataTalksClub/machine-learning-zoomcamp !2025-03-28100664|The code from the Machine Learning Bookcamp book and a free course based on the book| | 506|nerfstudio-project/nerfstudio !2025-03-28100343|A collaboration friendly studio for NeRFs| | 507|cupy/cupy !2025-03-28100344|NumPy & SciPy for GPU| | 508|NVIDIA/TensorRT-LLM !2025-03-281000823|TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.| | 509|wasp-lang/open-saas !2025-03-2899665|A free, open-source SaaS app starter for React & Node.js with superpowers. Production-ready. Community-driven.| | 510|huggingface/text-generation-inference !2025-03-2899383|Large Language Model Text Generation Inference| | 511|jxnl/instructor !2025-03-2899224|structured outputs for llms| | 512|GoogleCloudPlatform/generative-ai !2025-03-2899086|Sample code and notebooks for Generative AI on Google Cloud| | 513|manticoresoftware/manticoresearch !2025-03-2898799|Easy to use open source fast database for search | | 514|langfuse/langfuse !2025-03-28985134|🪢 Open source LLM engineering platform. Observability, metrics, evals, prompt management, testing, prompt playground, datasets, LLM evaluations -- 🍊YC W23 🤖 integrate via Typescript, Python / Decorators, OpenAI, Langchain, LlamaIndex, Litellm, Instructor, Mistral, Perplexity, Claude, Gemini, Vertex| | 515|keephq/keep !2025-03-2897949|The open-source alert management and AIOps platform| | 516|sashabaranov/go-openai !2025-03-2897843|OpenAI ChatGPT, GPT-3, GPT-4, DALL·E, Whisper API wrapper for Go| | 517|autowarefoundation/autoware !2025-03-2897766|Autoware - the world's leading open-source software project for autonomous driving| | 518|anthropics/courses !2025-03-2897269|Anthropic's educational courses| | 519|popcorn-official/popcorn-desktop !2025-03-2896853|Popcorn Time is a multi-platform, free software BitTorrent client that includes an integrated media player ( Windows / Mac / Linux ) A Butter-Project Fork| | 520|getmaxun/maxun !2025-03-28968515|🔥 Open-source no-code web data extraction platform. Turn websites to APIs and spreadsheets with no-code robots in minutes! [In Beta]| | 521|wandb/wandb !2025-03-2896763|🔥 A tool for visualizing and tracking your machine learning experiments. This repo contains the CLI and Python API.| | 522|karpathy/minbpe !2025-03-2895353|Minimal, clean, code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.| | 523|bigscience-workshop/petals !2025-03-2895142|🌸 Run large language models at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading| | 524|OthersideAI/self-operating-computer !2025-03-2894931|A framework to enable multimodal models to operate a computer.| | 525|mshumer/gpt-prompt-engineer !2025-03-2894911|| | 526| BloopAI/bloop !2025-03-2894710 | A fast code search engine written in Rust| | 527|BlinkDL/ChatRWKV !2025-03-289467-1 |ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.| | 528|timlrx/tailwind-nextjs-starter-blog !2025-03-2894677|This is a Next.js, Tailwind CSS blogging starter template. Comes out of the box configured with the latest technologies to make technical writing a breeze. Easily configurable and customizable. Perfect as a replacement to existing Jekyll and Hugo individual blogs.| | 529|google/benchmark !2025-03-2893634|A microbenchmark support library| | 530|facebookresearch/nougat !2025-03-2893603|Implementation of Nougat Neural Optical Understanding for Academic Documents| | 531|modelscope/facechain !2025-03-2893536|FaceChain is a deep-learning toolchain for generating your Digital-Twin.| | 532|DrewThomasson/ebook2audiobook !2025-03-2893388|Convert ebooks to audiobooks with chapters and metadata using dynamic AI models and voice cloning. Supports 1,107+ languages!| | 533|RayTracing/raytracing.github.io !2025-03-2893035|Main Web Site (Online Books)| | 534|QwenLM/Qwen2.5-VL !2025-03-28930249|Qwen2.5-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.| | 535|WongKinYiu/yolov9 !2025-03-2892201|Implementation of paper - YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information| | 536|alibaba-damo-academy/FunASR !2025-03-28920222|A Fundamental End-to-End Speech Recognition Toolkit and Open Source SOTA Pretrained Models.| | 537|Visualize-ML/Book4Power-of-Matrix !2025-03-2891931|Book4 'Power of Matrix' | | 538|dice2o/BingGPT !2025-03-289185-1 |Desktop application of new Bing's AI-powered chat (Windows, macOS and Linux)| | 539|browserbase/stagehand !2025-03-28917621|An AI web browsing framework focused on simplicity and extensibility.| | 540|FlagOpen/FlagEmbedding !2025-03-28914111|Dense Retrieval and Retrieval-augmented LLMs| | 541|Const-me/Whisper !2025-03-2890979|High-performance GPGPU inference of OpenAI's Whisper automatic speech recognition (ASR) model| | 542|lucidrains/denoising-diffusion-pytorch !2025-03-2890942|Implementation of Denoising Diffusion Probabilistic Model in Pytorch| | 543|Chainlit/chainlit !2025-03-28904422|Build Conversational AI in minutes ⚡️| | 544|togethercomputer/OpenChatKit !2025-03-2890160 |OpenChatKit provides a powerful, open-source base to create both specialized and general purpose chatbots for various applications| | 545|Stability-AI/StableStudio !2025-03-2889631 |Community interface for generative AI| | 546|voicepaw/so-vits-svc-fork !2025-03-2889482 |so-vits-svc fork with realtime support, improved interface and more features.| | 547|pymc-devs/pymc !2025-03-2889413|Bayesian Modeling and Probabilistic Programming in Python| | 548|espnet/espnet !2025-03-2889302|End-to-End Speech Processing Toolkit| | 549|kedacore/keda !2025-03-2888991|KEDA is a Kubernetes-based Event Driven Autoscaling component. It provides event driven scale for any container running in Kubernetes| | 550|open-mmlab/Amphion !2025-03-28886911|Amphion (/æmˈfaɪən/) is a toolkit for Audio, Music, and Speech Generation. Its purpose is to support reproducible research and help junior researchers and engineers get started in the field of audio, music, and speech generation research and development.| | 551|gorse-io/gorse !2025-03-2888451|Gorse open source recommender system engine| | 552|adams549659584/go-proxy-bingai !2025-03-288768-1 |A Microsoft New Bing demo site built with Vue3 and Go, providing a consistent UI experience, supporting ChatGPT prompts, and accessible within China| | 553|open-mmlab/mmsegmentation !2025-03-2887513|OpenMMLab Semantic Segmentation Toolbox and Benchmark.| | 554|bytedance/monolith !2025-03-2887223|ByteDance's Recommendation System| | 555|LouisShark/chatgptsystemprompt !2025-03-2887216|store all agent's system prompt| | 556|brexhq/prompt-engineering !2025-03-2887080 |Tips and tricks for working with Large Language Models like OpenAI's GPT-4.| | 557|erincatto/box2d !2025-03-2886841|Box2D is a 2D physics engine for games| | 558|🔥microsoft/ai-agents-for-beginners !2025-03-288669323|10 Lessons to Get Started Building AI Agents| | 559|nashsu/FreeAskInternet !2025-03-2886102|FreeAskInternet is a completely free, private and locally running search aggregator & answer generate using LLM, without GPU needed. The user can ask a question and the system will make a multi engine search and combine the search result to the ChatGPT3.5 LLM and generate the answer based on search results.| | 560|goldmansachs/gs-quant !2025-03-2885981|Python toolkit for quantitative finance| | 561|srbhr/Resume-Matcher !2025-03-2885800|Open Source Free ATS Tool to compare Resumes with Job Descriptions and create a score to rank them.| | 562|facebookresearch/ImageBind !2025-03-2885681 |ImageBind One Embedding Space to Bind Them All| | 563|ashawkey/stable-dreamfusion !2025-03-2885481 |A pytorch implementation of text-to-3D dreamfusion, powered by stable diffusion.| | 564|meetecho/janus-gateway !2025-03-2885232|Janus WebRTC Server| | 565|google/magika !2025-03-2885003|Detect file content types with deep learning| | 566|huggingface/chat-ui !2025-03-2884871 |Open source codebase powering the HuggingChat app| | 567|EleutherAI/lm-evaluation-harness !2025-03-28843012|A framework for few-shot evaluation of autoregressive language models.| | 568|jina-ai/reader !2025-03-2884089|Convert any URL to an LLM-friendly input with a simple prefix https://r.jina.ai/| | 569|microsoft/TypeChat !2025-03-288406-1|TypeChat is a library that makes it easy to build natural language interfaces using types.| | 570|thuml/Time-Series-Library !2025-03-28839715|A Library for Advanced Deep Time Series Models.| | 571|OptimalScale/LMFlow !2025-03-2883882|An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Model for All.| | 572|baptisteArno/typebot.io !2025-03-2883845|💬 Typebot is a powerful chatbot builder that you can self-host.| | 573|jzhang38/TinyLlama !2025-03-2883504|The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.| | 574|fishaudio/Bert-VITS2 !2025-03-2883472|vits2 backbone with multilingual-bert| | 575|OpenBMB/XAgent !2025-03-2882683|An Autonomous LLM Agent for Complex Task Solving| | 576|Acly/krita-ai-diffusion !2025-03-2882387|Streamlined interface for generating images with AI in Krita. Inpaint and outpaint with optional text prompt, no tweaking required.| | 577|jasonppy/VoiceCraft !2025-03-2882151|Zero-Shot Speech Editing and Text-to-Speech in the Wild| | 578|SJTU-IPADS/PowerInfer !2025-03-2881693|High-speed Large Language Model Serving on PCs with Consumer-grade GPUs| | 579|modelscope/DiffSynth-Studio !2025-03-28814713|Enjoy the magic of Diffusion models!| | 580|o3de/o3de !2025-03-2881443|Open 3D Engine (O3DE) is an Apache 2.0-licensed multi-platform 3D engine that enables developers and content creators to build AAA games, cinema-quality 3D worlds, and high-fidelity simulations without any fees or commercial obligations.| | 581|zmh-program/chatnio !2025-03-2881325|🚀 Next Generation AI One-Stop Internationalization Solution. 🚀 下一代 AI 一站式 B/C 端解决方案,支持 OpenAI,Midjourney,Claude,讯飞星火,Stable Diffusion,DALL·E,ChatGLM,通义千问,腾讯混元,360 智脑,百川 AI,火山方舟,新必应,Gemini,Moonshot 等模型,支持对话分享,自定义预设,云端同步,模型市场,支持弹性计费和订阅计划模式,支持图片解析,支持联网搜索,支持模型缓存,丰富美观的后台管理与仪表盘数据统计。| | 582|leptonai/searchwithlepton !2025-03-2880632|Building a quick conversation-based search demo with Lepton AI.| | 583|sebastianstarke/AI4Animation !2025-03-2880620|Bringing Characters to Life with Computer Brains in Unity| | 584|wangrongding/wechat-bot !2025-03-2880528|🤖一个基于 WeChaty 结合 DeepSeek / ChatGPT / Kimi / 讯飞等Ai服务实现的微信机器人 ,可以用来帮助你自动回复微信消息,或者管理微信群/好友,检测僵尸粉等...| | 585|openvinotoolkit/openvino !2025-03-2880528|OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference| | 586|steven2358/awesome-generative-ai !2025-03-28802610|A curated list of modern Generative Artificial Intelligence projects and services| | 587|adam-maj/tiny-gpu !2025-03-2880234|A minimal GPU design in Verilog to learn how GPUs work from the ground up| | 588| anse-app/chatgpt-demo !2025-03-2880180 | A demo repo based on OpenAI API (gpt-3.5-turbo) | | 589| acheong08/EdgeGPT !2025-03-288015-1 |Reverse engineered API of Microsoft's Bing Chat | | 590|ai-collection/ai-collection !2025-03-2879994 |The Generative AI Landscape - A Collection of Awesome Generative AI Applications| | 591|GreyDGL/PentestGPT !2025-03-2879953 |A GPT-empowered penetration testing tool| | 592|delta-io/delta !2025-03-2879112|An open-source storage framework that enables building a Lakehouse architecture with compute engines including Spark, PrestoDB, Flink, Trino, and Hive and APIs| | 593|dataelement/bisheng !2025-03-2879085|Bisheng is an open LLM devops platform for next generation AI applications.| | 594|e2b-dev/e2b !2025-03-2878447 |Vercel for AI agents. We help developers to build, deploy, and monitor AI agents. Focusing on specialized AI agents that build software for you - your personal software developers.| | 595|01-ai/Yi !2025-03-2878311|A series of large language models trained from scratch by developers @01-ai| | 596|Plachtaa/VALL-E-X !2025-03-287830-1|An open source implementation of Microsoft's VALL-E X zero-shot TTS model. The demo is available at https://plachtaa.github.io| | 597|abhishekkrthakur/approachingalmost !2025-03-2878204|Approaching (Almost) Any Machine Learning Problem| | 598|pydantic/pydantic-ai !2025-03-28781041|Agent Framework / shim to use Pydantic with LLMs| | 599|rany2/edge-tts !2025-03-2877901|Use Microsoft Edge's online text-to-speech service from Python WITHOUT needing Microsoft Edge or Windows or an API key| | 600|CASIA-IVA-Lab/FastSAM !2025-03-2877881|Fast Segment Anything| | 601|netease-youdao/EmotiVoice !2025-03-2877817|EmotiVoice 😊: a Multi-Voice and Prompt-Controlled TTS Engine| | 602|lllyasviel/IC-Light !2025-03-2877804|More relighting!| | 603|kroma-network/tachyon !2025-03-287774-1|Modular ZK(Zero Knowledge) backend accelerated by GPU| | 604|deep-floyd/IF !2025-03-2877731 |A novel state-of-the-art open-source text-to-image model with a high degree of photorealism and language understanding| | 605|oumi-ai/oumi !2025-03-2877705|Everything you need to build state-of-the-art foundation models, end-to-end.| | 606|reorproject/reor !2025-03-2877681|AI note-taking app that runs models locally.| | 607|lightpanda-io/browser !2025-03-28775813|Lightpanda: the headless browser designed for AI and automation| | 608|xiangsx/gpt4free-ts !2025-03-287755-1|Providing a free OpenAI GPT-4 API ! This is a replication project for the typescript version of xtekky/gpt4free| | 609|IDEA-Research/GroundingDINO !2025-03-28773311|Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"| | 610|bunkerity/bunkerweb !2025-03-2877326|🛡️ Make your web services secure by default !| | 611|vikhyat/moondream !2025-03-2877057|tiny vision language model| | 612|firmai/financial-machine-learning !2025-03-287703-1|A curated list of practical financial machine learning tools and applications.| | 613|n8n-io/self-hosted-ai-starter-kit !2025-03-28765121|The Self-hosted AI Starter Kit is an open-source template that quickly sets up a local AI environment. Curated by n8n, it provides essential tools for creating secure, self-hosted AI workflows.| | 614|intel-analytics/ipex-llm !2025-03-2876507|Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with llama.cpp, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, ModelScope, etc.| | 615|jrouwe/JoltPhysics !2025-03-28764510|A multi core friendly rigid body physics and collision detection library. Written in C++. Suitable for games and VR applications. Used by Horizon Forbidden West.| | 616|THUDM/CodeGeeX2 !2025-03-2876270|CodeGeeX2: A More Powerful Multilingual Code Generation Model| | 617|meta-llama/llama-stack !2025-03-2875866|Composable building blocks to build Llama Apps| | 618|sweepai/sweep !2025-03-287530-1|Sweep is an AI junior developer| | 619|lllyasviel/Omost !2025-03-2875301|Your image is almost there!| | 620|ahmedbahaaeldin/From-0-to-Research-Scientist-resources-guide !2025-03-2875050|Detailed and tailored guide for undergraduate students or anybody want to dig deep into the field of AI with solid foundation.| | 621|dair-ai/ML-Papers-Explained !2025-03-2875050|Explanation to key concepts in ML| | 622|zaidmukaddam/scira !2025-03-28750110|Scira (Formerly MiniPerplx) is a minimalistic AI-powered search engine that helps you find information on the internet. Powered by Vercel AI SDK! Search with models like Grok 2.0.| | 623|Portkey-AI/gateway !2025-03-28749416|A Blazing Fast AI Gateway. Route to 100+ LLMs with 1 fast & friendly API.| | 624|web-infra-dev/midscene !2025-03-28748729|An AI-powered automation SDK can control the page, perform assertions, and extract data in JSON format using natural language.| | 625|zilliztech/GPTCache !2025-03-2874801 |GPTCache is a library for creating semantic cache to store responses from LLM queries.| | 626|niedev/RTranslator !2025-03-2874742|RTranslator is the world's first open source real-time translation app.| |!green-up-arrow.svg 627|roboflow/notebooks !2025-03-2874666|Examples and tutorials on using SOTA computer vision models and techniques. Learn everything from old-school ResNet, through YOLO and object-detection transformers like DETR, to the latest models like Grounding DINO and SAM.| |!red-down-arrow 628|openlm-research/openllama !2025-03-2874652|OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset| | 629|LiheYoung/Depth-Anything !2025-03-2874155|Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data| | 630|enso-org/enso !2025-03-2874040|Hybrid visual and textual functional programming.| | 631|bigcode-project/starcoder !2025-03-287401-1 |Home of StarCoder: fine-tuning & inference!| | 632|git-ecosystem/git-credential-manager !2025-03-2873975|Secure, cross-platform Git credential storage with authentication to GitHub, Azure Repos, and other popular Git hosting services.| | 633|OpenGVLab/InternVL !2025-03-2873634|[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4V. 接近GPT-4V表现的可商用开源模型| | 634|WooooDyy/LLM-Agent-Paper-List !2025-03-2873551|The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.| | 635|lencx/Noi !2025-03-2873157|🦄 AI + Tools + Plugins + Community| | 636|udlbook/udlbook !2025-03-2873075|Understanding Deep Learning - Simon J.D. Prince| | 637|OpenBMB/MiniCPM !2025-03-2872841|MiniCPM-2B: An end-side LLM outperforms Llama2-13B.| | 638|jaywalnut310/vits !2025-03-2872815 |VITS: Conditional Variational Autoencoder with Adversarial Learning for End-to-End Text-to-Speech| | 639|xorbitsai/inference !2025-03-28727528|Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.| | 640|PWhiddy/PokemonRedExperiments !2025-03-2872492|Playing Pokemon Red with Reinforcement Learning| | 641|Canner/WrenAI !2025-03-28723213|🤖 Open-source AI Agent that empowers data-driven teams to chat with their data to generate Text-to-SQL, charts, spreadsheets, reports, and BI. 📈📊📋🧑‍💻| | 642|miurla/morphic !2025-03-2872258|An AI-powered answer engine with a generative UI| | 643|ml-explore/mlx-examples !2025-03-2872168|Examples in the MLX framework| | 644|PKU-YuanGroup/ChatLaw !2025-03-2872010|Chinese Legal Large Model| | 645|NVIDIA/cutlass !2025-03-2871883|CUDA Templates for Linear Algebra Subroutines| | 646|FoundationVision/VAR !2025-03-28717444|[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction"| | 647|ymcui/Chinese-LLaMA-Alpaca-2 !2025-03-2871561|Chinese LLaMA-2 & Alpaca-2 LLMs| | 648|nadermx/backgroundremover !2025-03-2871514 |Background Remover lets you Remove Background from images and video using AI with a simple command line interface that is free and open source.| | 649|onuratakan/gpt-computer-assistant !2025-03-28714514|gpt-4o for windows, macos and ubuntu| | 650|graviraja/MLOps-Basics !2025-03-2871326|| | 651|Future-House/paper-qa !2025-03-287118-1|High accuracy RAG for answering questions from scientific documents with citations| | 652|open-mmlab/mmagic !2025-03-2871102 |OpenMMLab Multimodal Advanced, Generative, and Intelligent Creation Toolbox| | 653|bhaskatripathi/pdfGPT !2025-03-2870941 |PDF GPT allows you to chat with the contents of your PDF file by using GPT capabilities. The only open source solution to turn your pdf files in a chatbot!| | 654|ollama/ollama-python !2025-03-28709117|Ollama Python library| | 655|facebookresearch/DiT !2025-03-2870376|Official PyTorch Implementation of "Scalable Diffusion Models with Transformers"| | 656|geekyutao/Inpaint-Anything !2025-03-2870262 |Inpaint anything using Segment Anything and inpainting models.| | 657|AbdullahAlfaraj/Auto-Photoshop-StableDiffusion-Plugin !2025-03-2870160 |A user-friendly plug-in that makes it easy to generate stable diffusion images inside Photoshop using Automatic1111-sd-webui as a backend.| | 658|apple/corenet !2025-03-2869990|CoreNet: A library for training deep neural networks| | 659|openstatusHQ/openstatus !2025-03-2869926|🏓 The open-source synthetic monitoring platform 🏓| | 660|weaviate/Verba !2025-03-2869772|Retrieval Augmented Generation (RAG) chatbot powered by Weaviate| | 661|meshery/meshery !2025-03-2869630|Meshery, the cloud native manager| | 662|OpenTalker/video-retalking !2025-03-2869530|[SIGGRAPH Asia 2022] VideoReTalking: Audio-based Lip Synchronization for Talking Head Video Editing In the Wild| | 663|digitalinnovationone/dio-lab-open-source !2025-03-28689013|Repositório do lab "Contribuindo em um Projeto Open Source no GitHub" da Digital Innovation One.| | 664|jianchang512/ChatTTS-ui !2025-03-2868842|一个简单的本地网页界面,直接使用ChatTTS将文字合成为语音,同时支持对外提供API接口。| | 665|patchy631/ai-engineering-hub !2025-03-28686434|In-depth tutorials on LLMs, RAGs and real-world AI agent applications.| | 666|gunnarmorling/1brc !2025-03-2868512|1️⃣🐝🏎️ The One Billion Row Challenge -- A fun exploration of how quickly 1B rows from a text file can be aggregated with Java| | 667|Azure-Samples/azure-search-openai-demo !2025-03-2868482 |A sample app for the Retrieval-Augmented Generation pattern running in Azure, using Azure Cognitive Search for retrieval and Azure OpenAI large language models to power ChatGPT-style and Q&A experiences.| | 668|mit-han-lab/streaming-llm !2025-03-2868382|Efficient Streaming Language Models with Attention Sinks| | 669|InternLM/InternLM !2025-03-2868352|InternLM has open-sourced a 7 billion parameter base model, a chat model tailored for practical scenarios and the training system.| | 670|dependency-check/DependencyCheck !2025-03-2868191|OWASP dependency-check is a software composition analysis utility that detects publicly disclosed vulnerabilities in application dependencies.| | 671|Soulter/AstrBot !2025-03-28678643|✨易上手的多平台 LLM 聊天机器人及开发框架✨。支持 QQ、QQ频道、Telegram、微信平台(Gewechat, 企业微信)、内置 Web Chat,OpenAI GPT、DeepSeek、Ollama、Llama、GLM、Gemini、OneAPI、LLMTuner,支持 LLM Agent 插件开发,可视化面板。一键部署。支持 Dify 工作流、代码执行器、Whisper 语音转文字。| | 672|react-native-webview/react-native-webview !2025-03-2867792|React Native Cross-Platform WebView| | 673|modelscope/agentscope !2025-03-28676916|Start building LLM-empowered multi-agent applications in an easier way.| | 674|mylxsw/aidea !2025-03-2867381|AIdea is a versatile app that supports GPT and domestic large language models,also supports "Stable Diffusion" text-to-image generation, image-to-image generation, SDXL 1.0, super-resolution, and image colorization| | 675|langchain-ai/ollama-deep-researcher !2025-03-28668635|Fully local web research and report writing assistant| | 676|threestudio-project/threestudio !2025-03-2866653|A unified framework for 3D content generation.| | 677|gaomingqi/Track-Anything !2025-03-2866631 |A flexible and interactive tool for video object tracking and segmentation, based on Segment Anything, XMem, and E2FGVI.| | 678|spdustin/ChatGPT-AutoExpert !2025-03-2866570|🚀🧠💬 Supercharged Custom Instructions for ChatGPT (non-coding) and ChatGPT Advanced Data Analysis (coding).| | 679|HariSekhon/DevOps-Bash-tools !2025-03-2866463|1000+ DevOps Bash Scripts - AWS, GCP, Kubernetes, Docker, CI/CD, APIs, SQL, PostgreSQL, MySQL, Hive, Impala, Kafka, Hadoop, Jenkins, GitHub, GitLab, BitBucket, Azure DevOps, TeamCity, Spotify, MP3, LDAP, Code/Build Linting, pkg mgmt for Linux, Mac, Python, Perl, Ruby, NodeJS, Golang, Advanced dotfiles: .bashrc, .vimrc, .gitconfig, .screenrc, tmux..| | 680|modelscope/swift !2025-03-28661530|ms-swift: Use PEFT or Full-parameter to finetune 200+ LLMs or 15+ MLLMs| | 681|langchain-ai/opengpts !2025-03-2866080|This is an open source effort to create a similar experience to OpenAI's GPTs and Assistants API| | 682| yihong0618/xiaogpt !2025-03-2865131 | Play ChatGPT with xiaomi ai speaker | | 683| civitai/civitai !2025-03-2865111 | Build a platform where people can share their stable diffusion models | | 684|KoljaB/RealtimeSTT !2025-03-28649513|A robust, efficient, low-latency speech-to-text library with advanced voice activity detection, wake word activation and instant transcription.| | 685|qunash/chatgpt-advanced !2025-03-2864910 | A browser extension that augments your ChatGPT prompts with web results.| | 686|Licoy/ChatGPT-Midjourney !2025-03-2864850|🎨 Own your own ChatGPT+Midjourney web service with one click| | 687|friuns2/BlackFriday-GPTs-Prompts !2025-03-2864744|List of free GPTs that doesn't require plus subscription| | 688|PixarAnimationStudios/OpenUSD !2025-03-2864700|Universal Scene Description| | 689|linyiLYi/street-fighter-ai !2025-03-2864630 |This is an AI agent for Street Fighter II Champion Edition.| | 690|run-llama/rags !2025-03-2864380|Build ChatGPT over your data, all with natural language| | 691|frdel/agent-zero !2025-03-2864154|Agent Zero AI framework| | 692|microsoft/DeepSpeedExamples !2025-03-2863911 |Example models using DeepSpeed| | 693|k8sgpt-ai/k8sgpt !2025-03-2863882|Giving Kubernetes Superpowers to everyone| | 694|open-metadata/OpenMetadata !2025-03-2863514|OpenMetadata is a unified platform for discovery, observability, and governance powered by a central metadata repository, in-depth lineage, and seamless team collaboration.| | 695|google/gemma.cpp !2025-03-2863163|lightweight, standalone C++ inference engine for Google's Gemma models.| | 696|RayVentura/ShortGPT !2025-03-286314-1|🚀🎬 ShortGPT - An experimental AI framework for automated short/video content creation. Enables creators to rapidly produce, manage, and deliver content using AI and automation.| | 697|openai/consistencymodels !2025-03-2862940 |Official repo for consistency models.| | 698|yangjianxin1/Firefly !2025-03-2862924|Firefly: Chinese conversational large language model (full-scale fine-tuning + QLoRA), supporting fine-tuning of Llma2, Llama, Baichuan, InternLM, Ziya, Bloom, and other large models| | 699|enricoros/big-AGI !2025-03-2862665|Generative AI suite powered by state-of-the-art models and providing advanced AI/AGI functions. It features AI personas, AGI functions, multi-model chats, text-to-image, voice, response streaming, code highlighting and execution, PDF import, presets for developers, much more. Deploy on-prem or in the cloud.| | 700|aptos-labs/aptos-core !2025-03-2862633|Aptos is a layer 1 blockchain built to support the widespread use of blockchain through better technology and user experience.| | 701|wenda-LLM/wenda !2025-03-286262-1 |Wenda: An LLM invocation platform. Its objective is to achieve efficient content generation tailored to specific environments while considering the limited computing resources of individuals and small businesses, as well as knowledge security and privacy concerns| | 702|Project-MONAI/MONAI !2025-03-2862603|AI Toolkit for Healthcare Imaging| | 703|HVision-NKU/StoryDiffusion !2025-03-2862470|Create Magic Story!| | 704|deepseek-ai/DeepSeek-LLM !2025-03-2862463|DeepSeek LLM: Let there be answers| | 705|Tohrusky/Final2x !2025-03-2862393|2^x Image Super-Resolution| | 706|OpenSPG/KAG !2025-03-28619611|KAG is a logical form-guided reasoning and retrieval framework based on OpenSPG engine and LLMs. It is used to build logical reasoning and factual Q&A solutions for professional domain knowledge bases. It can effectively overcome the shortcomings of the traditional RAG vector similarity calculation model.| | 707|Moonvy/OpenPromptStudio !2025-03-2861861 |AIGC Hint Word Visualization Editor| | 708|levihsu/OOTDiffusion !2025-03-2861761|Official implementation of OOTDiffusion| | 709|tmc/langchaingo !2025-03-2861729|LangChain for Go, the easiest way to write LLM-based programs in Go| | 710|vladmandic/automatic !2025-03-2861374|SD.Next: Advanced Implementation of Stable Diffusion and other Diffusion-based generative image models| | 711|clovaai/donut !2025-03-2861231 |Official Implementation of OCR-free Document Understanding Transformer (Donut) and Synthetic Document Generator (SynthDoG), ECCV 2022| | 712|Shaunwei/RealChar !2025-03-286121-1|🎙️🤖Create, Customize and Talk to your AI Character/Companion in Realtime(All in One Codebase!). Have a natural seamless conversation with AI everywhere(mobile, web and terminal) using LLM OpenAI GPT3.5/4, Anthropic Claude2, Chroma Vector DB, Whisper Speech2Text, ElevenLabs Text2Speech🎙️🤖| | 713|microsoft/TinyTroupe !2025-03-2861142|LLM-powered multiagent persona simulation for imagination enhancement and business insights.| | 714| rustformers/llm !2025-03-2861010 | Run inference for Large Language Models on CPU, with Rust| | 715|firebase/firebase-ios-sdk !2025-03-2860950|Firebase SDK for Apple App Development| | 716|vespa-engine/vespa !2025-03-2860824|The open big data serving engine. https://vespa.ai| | 717|n4ze3m/page-assist !2025-03-28607610|Use your locally running AI models to assist you in your web browsing| | 718|Dooy/chatgpt-web-midjourney-proxy !2025-03-2860646|chatgpt web, midjourney, gpts,tts, whisper 一套ui全搞定| | 719|ethereum-optimism/optimism !2025-03-2860213|Optimism is Ethereum, scaled.| | 720|sczhou/ProPainter !2025-03-2859971|[ICCV 2023] ProPainter: Improving Propagation and Transformer for Video Inpainting| | 721|MineDojo/Voyager !2025-03-2859951 |An Open-Ended Embodied Agent with Large Language Models| | 722|lavague-ai/LaVague !2025-03-2859800|Automate automation with Large Action Model framework| | 723|SevaSk/ecoute !2025-03-2859770 |Ecoute is a live transcription tool that provides real-time transcripts for both the user's microphone input (You) and the user's speakers output (Speaker) in a textbox. It also generates a suggested response using OpenAI's GPT-3.5 for the user to say based on the live transcription of the conversation.| | 724|google/mesop !2025-03-2859661|| | 725|pengxiao-song/LaWGPT !2025-03-2859542 |Repo for LaWGPT, Chinese-Llama tuned with Chinese Legal knowledge| | 726|fr0gger/Awesome-GPT-Agents !2025-03-2859434|A curated list of GPT agents for cybersecurity| | 727|google-deepmind/graphcast !2025-03-2859412|| | 728|comet-ml/opik !2025-03-28594126|Open-source end-to-end LLM Development Platform| | 729|SciPhi-AI/R2R !2025-03-28594033|A framework for rapid development and deployment of production-ready RAG systems| | 730|SkalskiP/courses !2025-03-2859272 |This repository is a curated collection of links to various courses and resources about Artificial Intelligence (AI)| | 731|QuivrHQ/MegaParse !2025-03-2859122|File Parser optimised for LLM Ingestion with no loss 🧠 Parse PDFs, Docx, PPTx in a format that is ideal for LLMs.| | 732|pytorch-labs/gpt-fast !2025-03-2858971|Simple and efficient pytorch-native transformer text generation in !2025-03-2858886|Curated list of chatgpt prompts from the top-rated GPTs in the GPTs Store. Prompt Engineering, prompt attack & prompt protect. Advanced Prompt Engineering papers.| | 734|nilsherzig/LLocalSearch !2025-03-2858852|LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a chain of LLMs to find the answer. The user can see the progress of the agents and the final answer. No OpenAI or Google API keys are needed.| | 735|kuafuai/DevOpsGPT !2025-03-285874-2|Multi agent system for AI-driven software development. Convert natural language requirements into working software. Supports any development language and extends the existing base code.| | 736|myshell-ai/MeloTTS !2025-03-2858486|High-quality multi-lingual text-to-speech library by MyShell.ai. Support English, Spanish, French, Chinese, Japanese and Korean.| | 737|OpenGVLab/LLaMA-Adapter !2025-03-2858421 |Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters| | 738|volcengine/verl !2025-03-28582563|veRL: Volcano Engine Reinforcement Learning for LLM| | 739|a16z-infra/companion-app !2025-03-2858171|AI companions with memory: a lightweight stack to create and host your own AI companions| | 740|HumanAIGC/OutfitAnyone !2025-03-285816-1|Outfit Anyone: Ultra-high quality virtual try-on for Any Clothing and Any Person| | 741|josStorer/RWKV-Runner !2025-03-2857472|A RWKV management and startup tool, full automation, only 8MB. And provides an interface compatible with the OpenAI API. RWKV is a large language model that is fully open source and available for commercial use.| | 742|648540858/wvp-GB28181-pro !2025-03-2857414|WEB VIDEO PLATFORM是一个基于GB28181-2016标准实现的网络视频平台,支持NAT穿透,支持海康、大华、宇视等品牌的IPC、NVR、DVR接入。支持国标级联,支持rtsp/rtmp等视频流转发到国标平台,支持rtsp/rtmp等推流转发到国标平台。| | 743|ToonCrafter/ToonCrafter !2025-03-2857345|a research paper for generative cartoon interpolation| | 744|PawanOsman/ChatGPT !2025-03-2857191|OpenAI API Free Reverse Proxy| | 745|apache/hudi !2025-03-2857091|Upserts, Deletes And Incremental Processing on Big Data.| | 746| nsarrazin/serge !2025-03-2857081 | A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API| | 747|homanp/superagent !2025-03-2857021|🥷 Superagent - Build, deploy, and manage LLM-powered agents| | 748|ramonvc/freegpt-webui !2025-03-2856910|GPT 3.5/4 with a Chat Web UI. No API key is required.| | 749|baichuan-inc/baichuan-7B !2025-03-2856901|A large-scale 7B pretraining language model developed by BaiChuan-Inc.| | 750|Azure/azure-sdk-for-net !2025-03-2856792|This repository is for active development of the Azure SDK for .NET. For consumers of the SDK we recommend visiting our public developer docs at https://learn.microsoft.com/dotnet/azure/ or our versioned developer docs at https://azure.github.io/azure-sdk-for-net.| | 751|mnotgod96/AppAgent !2025-03-2856643|AppAgent: Multimodal Agents as Smartphone Users, an LLM-based multimodal agent framework designed to operate smartphone apps.| | 752|microsoft/TaskWeaver !2025-03-2856243|A code-first agent framework for seamlessly planning and executing data analytics tasks.| | 753| yetone/bob-plugin-openai-translator !2025-03-285600-1 | A Bob Plugin base ChatGPT API | | 754|PrefectHQ/marvin !2025-03-2855840 |A batteries-included library for building AI-powered software| | 755|microsoft/promptbase !2025-03-2855832|All things prompt engineering| | 756|fullstackhero/dotnet-starter-kit !2025-03-2855560|Production Grade Cloud-Ready .NET 8 Starter Kit (Web API + Blazor Client) with Multitenancy Support, and Clean/Modular Architecture that saves roughly 200+ Development Hours! All Batteries Included.| | 757|deepseek-ai/DeepSeek-Coder-V2 !2025-03-2855435|DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence| | 758|aiwaves-cn/agents !2025-03-2855391|An Open-source Framework for Autonomous Language Agents| | 759|microsoft/Mastering-GitHub-Copilot-for-Paired-Programming !2025-03-2855158|A 6 Lesson course teaching everything you need to know about harnessing GitHub Copilot and an AI Paired Programing resource.| | 760|allenai/OLMo !2025-03-2854506|Modeling, training, eval, and inference code for OLMo| | 761|apify/crawlee-python !2025-03-2854493|Crawlee—A web scraping and browser automation library for Python to build reliable crawlers. Extract data for AI, LLMs, RAG, or GPTs. Download HTML, PDF, JPG, PNG, and other files from websites. Works with BeautifulSoup, Playwright, and raw HTTP. Both headful and headless mode. With proxy rotation.| | 762|k2-fsa/sherpa-onnx !2025-03-28541520|Speech-to-text, text-to-speech, and speaker recongition using next-gen Kaldi with onnxruntime without Internet connection. Support embedded systems, Android, iOS, Raspberry Pi, RISC-V, x86_64 servers, websocket server/client, C/C++, Python, Kotlin, C#, Go, NodeJS, Java, Swift| | 763|TEN-framework/TEN-Agent !2025-03-28541411|TEN Agent is a realtime conversational AI agent powered by TEN. It seamlessly integrates the OpenAI Realtime API, RTC capabilities, and advanced features like weather updates, web search, computer vision, and Retrieval-Augmented Generation (RAG).| | 764|google/gemmapytorch !2025-03-2854010|The official PyTorch implementation of Google's Gemma models| | 765|snakers4/silero-vad !2025-03-2853858|Silero VAD: pre-trained enterprise-grade Voice Activity Detector| | 766|livekit/agents !2025-03-2853836|Build real-time multimodal AI applications 🤖🎙️📹| | 767|pipecat-ai/pipecat !2025-03-28537811|Open Source framework for voice and multimodal conversational AI| | 768|EricLBuehler/mistral.rs !2025-03-28536324|Blazingly fast LLM inference.| | 769|asg017/sqlite-vec !2025-03-28535810|Work-in-progress vector search SQLite extension that runs anywhere.| | 770|albertan017/LLM4Decompile !2025-03-2853563|Reverse Engineering: Decompiling Binary Code with Large Language Models| | 771|Permify/permify !2025-03-2853235|An open-source authorization as a service inspired by Google Zanzibar, designed to build and manage fine-grained and scalable authorization systems for any application.| | 772|imoneoi/openchat !2025-03-2853171|OpenChat: Advancing Open-source Language Models with Imperfect Data| | 773|mosaicml/composer !2025-03-2853140|Train neural networks up to 7x faster| | 774|dsdanielpark/Bard-API !2025-03-285277-1 |The python package that returns a response of Google Bard through API.| | 775|lxfater/inpaint-web !2025-03-2852552|A free and open-source inpainting & image-upscaling tool powered by webgpu and wasm on the browser。| | 776|leanprover/lean4 !2025-03-2852441|Lean 4 programming language and theorem prover| | 777|AILab-CVC/YOLO-World !2025-03-2852415|Real-Time Open-Vocabulary Object Detection| | 778|openchatai/OpenChat !2025-03-2852260 |Run and create custom ChatGPT-like bots with OpenChat, embed and share these bots anywhere, the open-source chatbot console.| | 779|mufeedvh/code2prompt !2025-03-28519414|A CLI tool to convert your codebase into a single LLM prompt with source tree, prompt templating, and token counting.| | 780|biobootloader/wolverine !2025-03-2851700 |Automatically repair python scripts through GPT-4 to give them regenerative abilities.| | 781|huggingface/parler-tts !2025-03-2851671|Inference and training library for high-quality TTS models.| | 782|Akegarasu/lora-scripts !2025-03-2851308 |LoRA training scripts use kohya-ss's trainer, for diffusion model.| | 783|openchatai/OpenCopilot !2025-03-285128-3|🤖 🔥 Let your users chat with your product features and execute things by text - open source Shopify sidekick| | 784|e2b-dev/fragments !2025-03-2851228|Open-source Next.js template for building apps that are fully generated by AI. By E2B.| | 785|microsoft/SynapseML !2025-03-2851132|Simple and Distributed Machine Learning| | 786|aigc-apps/sd-webui-EasyPhoto !2025-03-285108-1|📷 EasyPhoto | | 787|ChaoningZhang/MobileSAM !2025-03-2850944|This is the official code for Faster Segment Anything (MobileSAM) project that makes SAM lightweight| | 788|huggingface/alignment-handbook !2025-03-2850932|Robust recipes for to align language models with human and AI preferences| | 789|alpkeskin/mosint !2025-03-2850920|An automated e-mail OSINT tool| | 790|TaskingAI/TaskingAI !2025-03-2850891|The open source platform for AI-native application development.| | 791|lipku/metahuman-stream !2025-03-28507615|Real time interactive streaming digital human| | 792|OpenInterpreter/01 !2025-03-2850530|The open-source language model computer| | 793|open-compass/opencompass !2025-03-28505111|OpenCompass is an LLM evaluation platform, supporting a wide range of models (InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.| | 794|xxlong0/Wonder3D !2025-03-2850491|A cross-domain diffusion model for 3D reconstruction from a single image| | 795|pytorch/torchtune !2025-03-2850342|A Native-PyTorch Library for LLM Fine-tuning| | 796|SuperDuperDB/superduperdb !2025-03-2850192|🔮 SuperDuperDB: Bring AI to your database: Integrate, train and manage any AI models and APIs directly with your database and your data.| | 797|WhiskeySockets/Baileys !2025-03-2850057|Lightweight full-featured typescript/javascript WhatsApp Web API| | 798| mpociot/chatgpt-vscode !2025-03-2849890 | A VSCode extension that allows you to use ChatGPT | | 799|OpenGVLab/DragGAN !2025-03-2849880|Unofficial Implementation of DragGAN - "Drag Your GAN: Interactive Point-based Manipulation on the Generative Image Manifold" (DragGAN 全功能实现,在线Demo,本地部署试用,代码、模型已全部开源,支持Windows, macOS, Linux)| | 800|microsoft/LLMLingua !2025-03-2849824|To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.| | 801|Zipstack/unstract !2025-03-2849745|No-code LLM Platform to launch APIs and ETL Pipelines to structure unstructured documents| | 802|OpenBMB/ToolBench !2025-03-2849621|An open platform for training, serving, and evaluating large language model for tool learning.| | 803|Fanghua-Yu/SUPIR !2025-03-2849593|SUPIR aims at developing Practical Algorithms for Photo-Realistic Image Restoration In the Wild| | 804|GaiaNet-AI/gaianet-node !2025-03-2849360|Install and run your own AI agent service| | 805|qodo-ai/qodo-cover !2025-03-284922-1|Qodo-Cover: An AI-Powered Tool for Automated Test Generation and Code Coverage Enhancement! 💻🤖🧪🐞| | 806|Zejun-Yang/AniPortrait !2025-03-2849042|AniPortrait: Audio-Driven Synthesis of Photorealistic Portrait Animation| | 807|lvwzhen/law-cn-ai !2025-03-2848901 |⚖️ AI Legal Assistant| | 808|developersdigest/llm-answer-engine !2025-03-2848740|Build a Perplexity-Inspired Answer Engine Using Next.js, Groq, Mixtral, Langchain, OpenAI, Brave & Serper| | 809|Plachtaa/VITS-fast-fine-tuning !2025-03-2848640|This repo is a pipeline of VITS finetuning for fast speaker adaptation TTS, and many-to-many voice conversion| | 810|espeak-ng/espeak-ng !2025-03-2848601|eSpeak NG is an open source speech synthesizer that supports more than hundred languages and accents.| | 811|ant-research/CoDeF !2025-03-2848581|[CVPR'24 Highlight] Official PyTorch implementation of CoDeF: Content Deformation Fields for Temporally Consistent Video Processing| | 812|deepseek-ai/DeepSeek-V2 !2025-03-2848512|| | 813|XRPLF/rippled !2025-03-2848210|Decentralized cryptocurrency blockchain daemon implementing the XRP Ledger protocol in C++| | 814|AutoMQ/automq !2025-03-28478721|AutoMQ is a cloud-first alternative to Kafka by decoupling durability to S3 and EBS. 10x cost-effective. Autoscale in seconds. Single-digit ms latency.| | 815|AILab-CVC/VideoCrafter !2025-03-2847800|VideoCrafter1: Open Diffusion Models for High-Quality Video Generation| | 816|nautechsystems/nautilustrader !2025-03-2847702|A high-performance algorithmic trading platform and event-driven backtester| | 817|kyegomez/swarms !2025-03-2847563|The Enterprise-Grade Production-Ready Multi-Agent Orchestration Framework Join our Community: https://discord.com/servers/agora-999382051935506503| | 818|Deci-AI/super-gradients !2025-03-2847310 |Easily train or fine-tune SOTA computer vision models with one open source training library. The home of Yolo-NAS.| | 819|QwenLM/Qwen2.5-Coder !2025-03-2847236|Qwen2.5-Coder is the code version of Qwen2.5, the large language model series developed by Qwen team, Alibaba Cloud.| | 820|SCIR-HI/Huatuo-Llama-Med-Chinese !2025-03-2847191 |Repo for HuaTuo (华驼), Llama-7B tuned with Chinese medical knowledge| | 821|togethercomputer/RedPajama-Data !2025-03-2846841 |code for preparing large datasets for training large language models| | 822|mishushakov/llm-scraper !2025-03-2846704|Turn any webpage into structured data using LLMs| | 823|1rgs/jsonformer !2025-03-2846663 |A Bulletproof Way to Generate Structured JSON from Language Models| | 824|anti-work/shortest !2025-03-2846565|QA via natural language AI tests| | 825|dnhkng/GlaDOS !2025-03-2846510|This is the Personality Core for GLaDOS, the first steps towards a real-life implementation of the AI from the Portal series by Valve.| | 826|Nukem9/dlssg-to-fsr3 !2025-03-2846380|Adds AMD FSR3 Frame Generation to games by replacing Nvidia DLSS-G Frame Generation (nvngx_dlssg).| | 827|BuilderIO/ai-shell !2025-03-2846373 |A CLI that converts natural language to shell commands.| | 828|facebookincubator/AITemplate !2025-03-2846220 |AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.| | 829|terraform-aws-modules/terraform-aws-eks !2025-03-2846030|Terraform module to create AWS Elastic Kubernetes (EKS) resources 🇺🇦| | 830|timescale/pgai !2025-03-2845915|A suite of tools to develop RAG, semantic search, and other AI applications more easily with PostgreSQL| | 831|awslabs/multi-agent-orchestrator !2025-03-2845788|Flexible and powerful framework for managing multiple AI agents and handling complex conversations| | 832|sanchit-gandhi/whisper-jax !2025-03-2845771 |Optimised JAX code for OpenAI's Whisper Model, largely built on the Hugging Face Transformers Whisper implementation| | 833|NVIDIA/NeMo-Guardrails !2025-03-2845755|NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.| | 834|PathOfBuildingCommunity/PathOfBuilding !2025-03-2845480|Offline build planner for Path of Exile.| | 835|UX-Decoder/Segment-Everything-Everywhere-All-At-Once !2025-03-2845412 |Official implementation of the paper "Segment Everything Everywhere All at Once"| | 836|build-trust/ockam !2025-03-2845171|Orchestrate end-to-end encryption, cryptographic identities, mutual authentication, and authorization policies between distributed applications – at massive scale.| | 837|google-research/timesfm !2025-03-2845135|TimesFM (Time Series Foundation Model) is a pretrained time-series foundation model developed by Google Research for time-series forecasting.| | 838|luosiallen/latent-consistency-model !2025-03-2844842|Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference| | 839|NVlabs/neuralangelo !2025-03-2844740|Official implementation of "Neuralangelo: High-Fidelity Neural Surface Reconstruction" (CVPR 2023)| | 840|kyegomez/tree-of-thoughts !2025-03-2844720 |Plug in and Play Implementation of Tree of Thoughts: Deliberate Problem Solving with Large Language Models that Elevates Model Reasoning by atleast 70%| | 841|sjvasquez/handwriting-synthesis !2025-03-2844720 |Handwriting Synthesis with RNNs ✏️| | 842| madawei2699/myGPTReader !2025-03-2844420 | A slack bot that can read any webpage, ebook or document and summarize it with chatGPT | | 843|OpenBMB/AgentVerse !2025-03-2844413|🤖 AgentVerse 🪐 provides a flexible framework that simplifies the process of building custom multi-agent environments for large language models (LLMs).| | 844|argmaxinc/WhisperKit !2025-03-2844395|Swift native speech recognition on-device for iOS and macOS applications.| | 845|landing-ai/vision-agent !2025-03-2844346|Vision agent| | 846|InternLM/xtuner !2025-03-2844273|An efficient, flexible and full-featured toolkit for fine-tuning large models (InternLM, Llama, Baichuan, Qwen, ChatGLM)| | 847|google-deepmind/alphageometry !2025-03-284421-1|Solving Olympiad Geometry without Human Demonstrations| | 848|ostris/ai-toolkit !2025-03-2844093|Various AI scripts. Mostly Stable Diffusion stuff.| | 849|LLM-Red-Team/kimi-free-api !2025-03-2844004|🚀 KIMI AI 长文本大模型白嫖服务,支持高速流式输出、联网搜索、长文档解读、图像解析、多轮对话,零配置部署,多路token支持,自动清理会话痕迹。| | 850|argilla-io/argilla !2025-03-2843991|Argilla is a collaboration platform for AI engineers and domain experts that require high-quality outputs, full data ownership, and overall efficiency.| | 851|spring-projects/spring-ai !2025-03-28438419|An Application Framework for AI Engineering| | 852|alibaba-damo-academy/FunClip !2025-03-2843555|Open-source, accurate and easy-to-use video clipping tool, LLM based AI clipping intergrated | | 853|yisol/IDM-VTON !2025-03-2843541|IDM-VTON : Improving Diffusion Models for Authentic Virtual Try-on in the Wild| | 854|fchollet/ARC-AGI !2025-03-2843368|The Abstraction and Reasoning Corpus| | 855|MahmoudAshraf97/whisper-diarization !2025-03-2843064|Automatic Speech Recognition with Speaker Diarization based on OpenAI Whisper| | 856|Speykious/cve-rs !2025-03-2843047|Blazingly 🔥 fast 🚀 memory vulnerabilities, written in 100% safe Rust. 🦀| | 857|Blealtan/efficient-kan !2025-03-2842770|An efficient pure-PyTorch implementation of Kolmogorov-Arnold Network (KAN).| | 858|smol-ai/GodMode !2025-03-284249-1|AI Chat Browser: Fast, Full webapp access to ChatGPT / Claude / Bard / Bing / Llama2! I use this 20 times a day.| | 859|openai/plugins-quickstart !2025-03-284235-4 |Get a ChatGPT plugin up and running in under 5 minutes!| | 860|Doriandarko/maestro !2025-03-2842260|A framework for Claude Opus to intelligently orchestrate subagents.| | 861|philz1337x/clarity-upscaler !2025-03-2842204|Clarity-Upscaler: Reimagined image upscaling for everyone| | 862|facebookresearch/co-tracker !2025-03-2842142|CoTracker is a model for tracking any point (pixel) on a video.| | 863|xlang-ai/OpenAgents !2025-03-2842031|OpenAgents: An Open Platform for Language Agents in the Wild| | 864|alibaba/higress !2025-03-28419514|🤖 AI Gateway | | 865|ray-project/llm-numbers !2025-03-2841920 |Numbers every LLM developer should know| | 866|fudan-generative-vision/champ !2025-03-2841820|Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance| | 867|NVIDIA/garak !2025-03-2841795|the LLM vulnerability scanner| | 868|leetcode-mafia/cheetah !2025-03-2841740 |Whisper & GPT-based app for passing remote SWE interviews| | 869|ragapp/ragapp !2025-03-2841710|The easiest way to use Agentic RAG in any enterprise| | 870|collabora/WhisperSpeech !2025-03-2841692|An Open Source text-to-speech system built by inverting Whisper.| | 871|Facico/Chinese-Vicuna !2025-03-2841520 |Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model| | 872|openai/grok !2025-03-2841381|| | 873|CrazyBoyM/llama3-Chinese-chat !2025-03-2841361|Llama3 Chinese Repository with modified versions, and training and deployment resources| | 874|luban-agi/Awesome-AIGC-Tutorials !2025-03-2841301|Curated tutorials and resources for Large Language Models, AI Painting, and more.| | 875|damo-vilab/AnyDoor !2025-03-2841192|Official implementations for paper: Anydoor: zero-shot object-level image customization| | 876|raspberrypi/pico-sdk !2025-03-2841072|| | 877|mshumer/gpt-llm-trainer !2025-03-284097-1|| | 878|metavoiceio/metavoice-src !2025-03-284076-1|AI for human-level speech intelligence| | 879|intelowlproject/IntelOwl !2025-03-2840763|IntelOwl: manage your Threat Intelligence at scale| | 880|a16z-infra/ai-getting-started !2025-03-2840682|A Javascript AI getting started stack for weekend projects, including image/text models, vector stores, auth, and deployment configs| | 881|MarkFzp/mobile-aloha !2025-03-2840641|Mobile ALOHA: Learning Bimanual Mobile Manipulation with Low-Cost Whole-Body Teleoperation| | 882| keijiro/AICommand !2025-03-2840380 | ChatGPT integration with Unity Editor | | 883|Tencent/HunyuanDiT !2025-03-2840214|Hunyuan-DiT : A Powerful Multi-Resolution Diffusion Transformer with Fine-Grained Chinese Understanding| | 884|hengyoush/kyanos !2025-03-2840061|Visualize the time packets spend in the kernel, watch & analyze in command line.| | 885|agiresearch/AIOS !2025-03-2840045|AIOS: LLM Agent Operating System| | 886|truefoundry/cognita !2025-03-2839773|RAG (Retrieval Augmented Generation) Framework for building modular, open source applications for production by TrueFoundry| | 887|X-PLUG/MobileAgent !2025-03-2839557|Mobile-Agent: Autonomous Multi-Modal Mobile Device Agent with Visual Perception| | 888|jackMort/ChatGPT.nvim !2025-03-2839231|ChatGPT Neovim Plugin: Effortless Natural Language Generation with OpenAI's ChatGPT API| | 889|microsoft/RD-Agent !2025-03-28388422|Research and development (R&D) is crucial for the enhancement of industrial productivity, especially in the AI era, where the core aspects of R&D are mainly focused on data and models. We are committed to automate these high-value generic R&D processes through our open source R&D automation tool RD-Agent, which let AI drive data-driven AI.| | 890|Significant-Gravitas/Auto-GPT-Plugins !2025-03-283882-1 |Plugins for Auto-GPT| | 891|apple/ml-mgie !2025-03-2838770|| | 892|OpenDriveLab/UniAD !2025-03-2838727|[CVPR 2023 Best Paper] Planning-oriented Autonomous Driving| | 893|llSourcell/DoctorGPT !2025-03-2838640|DoctorGPT is an LLM that can pass the US Medical Licensing Exam. It works offline, it's cross-platform, & your health data stays private.| | 894|FlagAI-Open/FlagAI !2025-03-2838601|FlagAI (Fast LArge-scale General AI models) is a fast, easy-to-use and extensible toolkit for large-scale model.| | 895|krishnaik06/Roadmap-To-Learn-Generative-AI-In-2024 !2025-03-2838513|Roadmap To Learn Generative AI In 2024| | 896|SysCV/sam-hq !2025-03-2838491|Segment Anything in High Quality| | 897|google/security-research !2025-03-2838420|This project hosts security advisories and their accompanying proof-of-concepts related to research conducted at Google which impact non-Google owned code.| | 898|shroominic/codeinterpreter-api !2025-03-2838330|Open source implementation of the ChatGPT Code Interpreter 👾| | 899|Yonom/assistant-ui !2025-03-2838308|React Components for AI Chat 💬 🚀| | 900|nucleuscloud/neosync !2025-03-2838262|Open source data anonymization and synthetic data orchestration for developers. Create high fidelity synthetic data and sync it across your environments.| | 901|ravenscroftj/turbopilot !2025-03-2838230 |Turbopilot is an open source large-language-model based code completion engine that runs locally on CPU| | 902|NVlabs/Sana !2025-03-28380810|SANA: Efficient High-Resolution Image Synthesis with Linear Diffusion Transformer| | 903|huggingface/distil-whisper !2025-03-2838061|Distilled variant of Whisper for speech recognition. 6x faster, 50% smaller, within 1% word error rate.| | 904|Codium-ai/AlphaCodium !2025-03-2837971|code generation tool that surpasses most human competitors in CodeContests| | 905|fixie-ai/ultravox !2025-03-2837710|A fast multimodal LLM for real-time voice| | 906|unit-mesh/auto-dev !2025-03-28375715|🧙‍AutoDev: The AI-powered coding wizard with multilingual support 🌐, auto code generation 🏗️, and a helpful bug-slaying assistant 🐞! Customizable prompts 🎨 and a magic Auto Dev/Testing/Document/Agent feature 🧪 included! 🚀| | 907|Marker-Inc-Korea/AutoRAG !2025-03-2837432|AutoML tool for RAG| | 908|deepseek-ai/DeepSeek-VL !2025-03-283734-1|DeepSeek-VL: Towards Real-World Vision-Language Understanding| | 909|hiyouga/ChatGLM-Efficient-Tuning !2025-03-283692-1|Fine-tuning ChatGLM-6B with PEFT | | 910| Yue-Yang/ChatGPT-Siri !2025-03-2836921 | Shortcuts for Siri using ChatGPT API gpt-3.5-turbo model | | 911|0hq/WebGPT !2025-03-2836901 |Run GPT model on the browser with WebGPU. An implementation of GPT inference in less than ~2000 lines of vanilla Javascript.| | 912|cvg/LightGlue !2025-03-2836903|LightGlue: Local Feature Matching at Light Speed (ICCV 2023)| | 913|deanxv/coze-discord-proxy !2025-03-2836791|代理Discord-Bot对话Coze-Bot,实现API形式请求GPT4对话模型/微调模型| | 914|MervinPraison/PraisonAI !2025-03-2836764|PraisonAI application combines AutoGen and CrewAI or similar frameworks into a low-code solution for building and managing multi-agent LLM systems, focusing on simplicity, customisation, and efficient human-agent collaboration.| | 915|Ironclad/rivet !2025-03-2836345 |The open-source visual AI programming environment and TypeScript library| | 916|BasedHardware/OpenGlass !2025-03-2835851|Turn any glasses into AI-powered smart glasses| | 917|ricklamers/gpt-code-ui !2025-03-2835840 |An open source implementation of OpenAI's ChatGPT Code interpreter| | 918|whoiskatrin/chart-gpt !2025-03-2835830 |AI tool to build charts based on text input| | 919|github/CopilotForXcode !2025-03-2835788|Xcode extension for GitHub Copilot| | 920|hemansnation/God-Level-Data-Science-ML-Full-Stack !2025-03-2835570 |A collection of scientific methods, processes, algorithms, and systems to build stories & models. This roadmap contains 16 Chapters, whether you are a fresher in the field or an experienced professional who wants to transition into Data Science & AI| | 921|pytorch/torchchat !2025-03-2835461|Run PyTorch LLMs locally on servers, desktop and mobile| | 922| Kent0n-Li/ChatDoctor !2025-03-2835451 | A Medical Chat Model Fine-tuned on LLaMA Model using Medical Domain Knowledge | | 923|xtekky/chatgpt-clone !2025-03-283519-1 |ChatGPT interface with better UI| | 924|jupyterlab/jupyter-ai !2025-03-2835120|A generative AI extension for JupyterLab| | 925|pytorch/torchtitan !2025-03-2835064|A native PyTorch Library for large model training| | 926|minimaxir/simpleaichat !2025-03-2835031|Python package for easily interfacing with chat apps, with robust features and minimal code complexity.| | 927|srush/Tensor-Puzzles !2025-03-2834930|Solve puzzles. Improve your pytorch.| | 928|Helicone/helicone !2025-03-2834918|🧊 Open source LLM-Observability Platform for Developers. One-line integration for monitoring, metrics, evals, agent tracing, prompt management, playground, etc. Supports OpenAI SDK, Vercel AI SDK, Anthropic SDK, LiteLLM, LLamaIndex, LangChain, and more. 🍓 YC W23| | 929|run-llama/llama-hub !2025-03-2834740|A library of data loaders for LLMs made by the community -- to be used with LlamaIndex and/or LangChain| | 930|NExT-GPT/NExT-GPT !2025-03-2834700|Code and models for NExT-GPT: Any-to-Any Multimodal Large Language Model| | 931|souzatharsis/podcastfy !2025-03-2834661|An Open Source Python alternative to NotebookLM's podcast feature: Transforming Multimodal Content into Captivating Multilingual Audio Conversations with GenAI| | 932|Dataherald/dataherald !2025-03-2834450|Interact with your SQL database, Natural Language to SQL using LLMs| | 933|iryna-kondr/scikit-llm !2025-03-2834350 |Seamlessly integrate powerful language models like ChatGPT into scikit-learn for enhanced text analysis tasks.| | 934|Netflix/maestro !2025-03-2834230|Maestro: Netflix’s Workflow Orchestrator| | 935|CanadaHonk/porffor !2025-03-2833560|A from-scratch experimental AOT JS engine, written in JS| | 936|hustvl/Vim !2025-03-2833323|Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model| | 937|pashpashpash/vault-ai !2025-03-2833250 |OP Vault ChatGPT: Give ChatGPT long-term memory using the OP Stack (OpenAI + Pinecone Vector Database). Upload your own custom knowledge base files (PDF, txt, etc) using a simple React frontend.| | 938|tencentmusic/supersonic !2025-03-28330611|SuperSonic is the next-generation BI platform that integrates Chat BI (powered by LLM) and Headless BI (powered by semantic layer) paradigms.| | 939|billmei/every-chatgpt-gui !2025-03-2832981|Every front-end GUI client for ChatGPT| | 940|microsoft/torchgeo !2025-03-2832772|TorchGeo: datasets, samplers, transforms, and pre-trained models for geospatial data| | 941|LLMBook-zh/LLMBook-zh.github.io !2025-03-28326110|《大语言模型》作者:赵鑫,李军毅,周昆,唐天一,文继荣| | 942|dvlab-research/MiniGemini !2025-03-2832601|Official implementation for Mini-Gemini| | 943|rashadphz/farfalle !2025-03-2832460|🔍 AI search engine - self-host with local or cloud LLMs| | 944|Luodian/Otter !2025-03-2832450|🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.| | 945|AprilNEA/ChatGPT-Admin-Web !2025-03-2832370 | ChatGPT WebUI with user management and admin dashboard system| | 946|MarkFzp/act-plus-plus !2025-03-2832365|Imitation Learning algorithms with Co-traing for Mobile ALOHA: ACT, Diffusion Policy, VINN| | 947|ethen8181/machine-learning !2025-03-2832310|🌎 machine learning tutorials (mainly in Python3)| | 948|opengeos/segment-geospatial !2025-03-2832312 |A Python package for segmenting geospatial data with the Segment Anything Model (SAM)| | 949|iusztinpaul/hands-on-llms !2025-03-283225-2|🦖 𝗟𝗲𝗮𝗿𝗻 about 𝗟𝗟𝗠𝘀, 𝗟𝗟𝗠𝗢𝗽𝘀, and 𝘃𝗲𝗰𝘁𝗼𝗿 𝗗𝗕𝘀 for free by designing, training, and deploying a real-time financial advisor LLM system ~ 𝘴𝘰𝘶𝘳𝘤𝘦 𝘤𝘰𝘥𝘦 + 𝘷𝘪𝘥𝘦𝘰 & 𝘳𝘦𝘢𝘥𝘪𝘯𝘨 𝘮𝘢𝘵𝘦𝘳𝘪𝘢𝘭𝘴| | 950|ToTheBeginning/PuLID !2025-03-2832221|Official code for PuLID: Pure and Lightning ID Customization via Contrastive Alignment| | 951|neo4j-labs/llm-graph-builder !2025-03-2832164|Neo4j graph construction from unstructured data using LLMs| | 952|OpenGVLab/InternGPT !2025-03-2832150 |InternGPT (iGPT) is an open source demo platform where you can easily showcase your AI models. Now it supports DragGAN, ChatGPT, ImageBind, multimodal chat like GPT-4, SAM, interactive image editing, etc. Try it at igpt.opengvlab.com (支持DragGAN、ChatGPT、ImageBind、SAM的在线Demo系统)| | 953|PKU-YuanGroup/Video-LLaVA !2025-03-2832060 |Video-LLaVA: Learning United Visual Representation by Alignment Before Projection| | 954|DataTalksClub/llm-zoomcamp !2025-03-2832030|LLM Zoomcamp - a free online course about building an AI bot that can answer questions about your knowledge base| | 955|gptscript-ai/gptscript !2025-03-2832010|Natural Language Programming| |!green-up-arrow.svg 956|isaac-sim/IsaacLab !2025-03-28320113|Unified framework for robot learning built on NVIDIA Isaac Sim| |!red-down-arrow 957|ai-boost/Awesome-GPTs !2025-03-2832003|Curated list of awesome GPTs 👍.| | 958|huggingface/safetensors !2025-03-2831901|Simple, safe way to store and distribute tensors| | 959|linyiLYi/bilibot !2025-03-2831771|A local chatbot fine-tuned by bilibili user comments.| | 960| project-baize/baize-chatbot !2025-03-283168-1 | Let ChatGPT teach your own chatbot in hours with a single GPU! | | 961|Azure-Samples/cognitive-services-speech-sdk !2025-03-2831280|Sample code for the Microsoft Cognitive Services Speech SDK| | 962|microsoft/Phi-3CookBook !2025-03-2831231|This is a Phi-3 book for getting started with Phi-3. Phi-3, a family of open AI models developed by Microsoft. Phi-3 models are the most capable and cost-effective small language models (SLMs) available, outperforming models of the same size and next size up across a variety of language, reasoning, coding, and math benchmarks.| | 963|neuralmagic/deepsparse !2025-03-2831180|Sparsity-aware deep learning inference runtime for CPUs| | 964|sugarforever/chat-ollama !2025-03-2831000|ChatOllama is an open source chatbot based on LLMs. It supports a wide range of language models, and knowledge base management.| | 965|amazon-science/chronos-forecasting !2025-03-2830974|Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting| | 966|damo-vilab/i2vgen-xl !2025-03-2830902|Official repo for VGen: a holistic video generation ecosystem for video generation building on diffusion models| | 967|google-deepmind/gemma !2025-03-2830733|Open weights LLM from Google DeepMind.| | 968|iree-org/iree !2025-03-2830733|A retargetable MLIR-based machine learning compiler and runtime toolkit.| | 969|NVlabs/VILA !2025-03-2830724|VILA - a multi-image visual language model with training, inference and evaluation recipe, deployable from cloud to edge (Jetson Orin and laptops)| | 970|microsoft/torchscale !2025-03-2830661|Foundation Architecture for (M)LLMs| | 971|openai/openai-realtime-console !2025-03-2830656|React app for inspecting, building and debugging with the Realtime API| | 972|daveshap/OpenAIAgentSwarm !2025-03-2830610|HAAS = Hierarchical Autonomous Agent Swarm - "Resistance is futile!"| | 973|microsoft/PromptWizard !2025-03-2830555|Task-Aware Agent-driven Prompt Optimization Framework| | 974|CVI-SZU/Linly !2025-03-2830490 |Chinese-LLaMA basic model; ChatFlow Chinese conversation model; NLP pre-training/command fine-tuning dataset| | 975|cohere-ai/cohere-toolkit !2025-03-2830130|Toolkit is a collection of prebuilt components enabling users to quickly build and deploy RAG applications.| | 976|adamcohenhillel/ADeus !2025-03-2830131|An open source AI wearable device that captures what you say and hear in the real world and then transcribes and stores it on your own server. You can then chat with Adeus using the app, and it will have all the right context about what you want to talk about - a truly personalized, personal AI.| | 977|Lightning-AI/LitServe !2025-03-2830132|Lightning-fast serving engine for AI models. Flexible. Easy. Enterprise-scale.| | 978|potpie-ai/potpie !2025-03-2829973|Prompt-To-Agent : Create custom engineering agents for your codebase| | 979|ant-design/x !2025-03-28299529|Craft AI-driven interfaces effortlessly 🤖| | 980|meta-llama/PurpleLlama !2025-03-2829832|Set of tools to assess and improve LLM security.| | 981|williamyang1991/RerenderAVideo !2025-03-2829800|[SIGGRAPH Asia 2023] Rerender A Video: Zero-Shot Text-Guided Video-to-Video Translation| | 982|baichuan-inc/Baichuan-13B !2025-03-2829790|A 13B large language model developed by Baichuan Intelligent Technology| | 983|Stability-AI/stable-audio-tools !2025-03-2829761|Generative models for conditional audio generation| | 984|li-plus/chatglm.cpp !2025-03-2829720|C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & more LLMs| | 985|NVIDIA/GenerativeAIExamples !2025-03-2829546|Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.| | 986|Josh-XT/AGiXT !2025-03-2829521 |AGiXT is a dynamic AI Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.| | 987|MrForExample/ComfyUI-3D-Pack !2025-03-2829515|An extensive node suite that enables ComfyUI to process 3D inputs (Mesh & UV Texture, etc) using cutting edge algorithms (3DGS, NeRF, etc.)| | 988|olimorris/codecompanion.nvim !2025-03-28295111|✨ AI-powered coding, seamlessly in Neovim. Supports Anthropic, Copilot, Gemini, Ollama, OpenAI and xAI LLMs| | 989|salesforce/CodeT5 !2025-03-282940-1 |Home of CodeT5: Open Code LLMs for Code Understanding and Generation| | 990|facebookresearch/ijepa !2025-03-2829391|Official codebase for I-JEPA, the Image-based Joint-Embedding Predictive Architecture. First outlined in the CVPR paper, "Self-supervised learning from images with a joint-embedding predictive architecture."| | 991|eureka-research/Eureka !2025-03-2829351|Official Repository for "Eureka: Human-Level Reward Design via Coding Large Language Models"| | 992|NVIDIA/trt-llm-rag-windows !2025-03-282934-1|A developer reference project for creating Retrieval Augmented Generation (RAG) chatbots on Windows using TensorRT-LLM| | 993|gmpetrov/databerry !2025-03-282930-1|The no-code platform for building custom LLM Agents| | 994|AI4Finance-Foundation/FinRobot !2025-03-28291946|FinRobot: An Open-Source AI Agent Platform for Financial Applications using LLMs 🚀 🚀 🚀| | 995|nus-apr/auto-code-rover !2025-03-2829013|A project structure aware autonomous software engineer aiming for autonomous program improvement| | 996|deepseek-ai/DreamCraft3D !2025-03-2828921|[ICLR 2024] Official implementation of DreamCraft3D: Hierarchical 3D Generation with Bootstrapped Diffusion Prior| | 997|mlabonne/llm-datasets !2025-03-2828848|High-quality datasets, tools, and concepts for LLM fine-tuning.| | 998|facebookresearch/jepa !2025-03-2828712|PyTorch code and models for V-JEPA self-supervised learning from video.| | 999|facebookresearch/habitat-sim !2025-03-2828604|A flexible, high-performance 3D simulator for Embodied AI research.| | 1000|xenova/whisper-web !2025-03-2828581|ML-powered speech recognition directly in your browser| | 1001|cvlab-columbia/zero123 !2025-03-2828530|Zero-1-to-3: Zero-shot One Image to 3D Object: https://zero123.cs.columbia.edu/| | 1002|yuruotong1/autoMate !2025-03-28285121|Like Manus, Computer Use Agent(CUA) and Omniparser, we are computer-using agents.AI-driven local automation assistant that uses natural language to make computers work by themselves| | 1003|muellerberndt/mini-agi !2025-03-282845-1 |A minimal generic autonomous agent based on GPT3.5/4. Can analyze stock prices, perform network security tests, create art, and order pizza.| | 1004|allenai/open-instruct !2025-03-2828432|| | 1005|CodingChallengesFYI/SharedSolutions !2025-03-2828360|Publicly shared solutions to Coding Challenges| | 1006|hegelai/prompttools !2025-03-2828220|Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate).| | 1007|mazzzystar/Queryable !2025-03-2828222|Run CLIP on iPhone to Search Photos.| | 1008|Doubiiu/DynamiCrafter !2025-03-2828173|DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors| | 1009|SamurAIGPT/privateGPT !2025-03-282805-1 |An app to interact privately with your documents using the power of GPT, 100% privately, no data leaks| | 1010|facebookresearch/Pearl !2025-03-2827951|A Production-ready Reinforcement Learning AI Agent Library brought by the Applied Reinforcement Learning team at Meta.| | 1011|intuitem/ciso-assistant-community !2025-03-2827954|CISO Assistant is a one-stop-shop for GRC, covering Risk, AppSec and Audit Management and supporting +70 frameworks worldwide with auto-mapping: NIST CSF, ISO 27001, SOC2, CIS, PCI DSS, NIS2, CMMC, PSPF, GDPR, HIPAA, Essential Eight, NYDFS-500, DORA, NIST AI RMF, 800-53, 800-171, CyFun, CJIS, AirCyber, NCSC, ECC, SCF and so much more| | 1012|facebookresearch/audio2photoreal !2025-03-2827840|Code and dataset for photorealistic Codec Avatars driven from audio| | 1013|Azure/azure-rest-api-specs !2025-03-2827770|The source for REST API specifications for Microsoft Azure.| | 1014|SCUTlihaoyu/open-chat-video-editor !2025-03-2827690 |Open source short video automatic generation tool| | 1015|Alpha-VLLM/LLaMA2-Accessory !2025-03-2827642|An Open-source Toolkit for LLM Development| | 1016|johnma2006/mamba-minimal !2025-03-2827601|Simple, minimal implementation of the Mamba SSM in one file of PyTorch.| | 1017|nerfstudio-project/gsplat !2025-03-2827576|CUDA accelerated rasterization of gaussian splatting| | 1018|Physical-Intelligence/openpi !2025-03-28274617|| | 1019|leptonai/leptonai !2025-03-2827246|A Pythonic framework to simplify AI service building| |!green-up-arrow.svg 1020|joanrod/star-vector !2025-03-28271149|StarVector is a foundation model for SVG generation that transforms vectorization into a code generation task. Using a vision-language modeling architecture, StarVector processes both visual and textual inputs to produce high-quality SVG code with remarkable precision.| |!red-down-arrow 1021|jqnatividad/qsv !2025-03-2827092|CSVs sliced, diced & analyzed.| | 1022|FranxYao/chain-of-thought-hub !2025-03-2826991|Benchmarking large language models' complex reasoning ability with chain-of-thought prompting| | 1023|princeton-nlp/SWE-bench !2025-03-2826965|[ICLR 2024] SWE-Bench: Can Language Models Resolve Real-world Github Issues?| | 1024|elastic/otel-profiling-agent !2025-03-2826930|The production-scale datacenter profiler| | 1025|src-d/hercules !2025-03-2826900|Gaining advanced insights from Git repository history.| | 1026|lanqian528/chat2api !2025-03-2826695|A service that can convert ChatGPT on the web to OpenAI API format.| | 1027|ishan0102/vimGPT !2025-03-2826681|Browse the web with GPT-4V and Vimium| | 1028|TMElyralab/MuseV !2025-03-2826650|MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising| | 1029|georgia-tech-db/eva !2025-03-2826600 |AI-Relational Database System | | 1030|kubernetes-sigs/controller-runtime !2025-03-2826590|Repo for the controller-runtime subproject of kubebuilder (sig-apimachinery)| | 1031|gptlink/gptlink !2025-03-2826550 |Build your own free commercial ChatGPT environment in 10 minutes. The setup is simple and includes features such as user management, orders, tasks, and payments| | 1032|pytorch/executorch !2025-03-2826534|On-device AI across mobile, embedded and edge for PyTorch| | 1033|NVIDIA/nv-ingest !2025-03-2826290|NVIDIA Ingest is an early access set of microservices for parsing hundreds of thousands of complex, messy unstructured PDFs and other enterprise documents into metadata and text to embed into retrieval systems.| | 1034|SuperTux/supertux !2025-03-2826081|SuperTux source code| | 1035|abi/secret-llama !2025-03-2826050|Fully private LLM chatbot that runs entirely with a browser with no server needed. Supports Mistral and LLama 3.| | 1036|liou666/polyglot !2025-03-2825841 |Desktop AI Language Practice Application| | 1037|janhq/nitro !2025-03-2825821|A fast, lightweight, embeddable inference engine to supercharge your apps with local AI. OpenAI-compatible API| | 1038|deepseek-ai/DeepSeek-Math !2025-03-2825825|DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models| | 1039|anthropics/prompt-eng-interactive-tutorial !2025-03-2825781|Anthropic's Interactive Prompt Engineering Tutorial| | 1040|microsoft/promptbench !2025-03-2825741|A unified evaluation framework for large language models| | 1041|baaivision/Painter !2025-03-2825580 |Painter & SegGPT Series: Vision Foundation Models from BAAI| | 1042|OpenPipe/OpenPipe !2025-03-2825581|Turn expensive prompts into cheap fine-tuned models| | 1043|TracecatHQ/tracecat !2025-03-2825531|😼 The AI-native, open source alternative to Tines / Splunk SOAR.| | 1044|JoshuaC215/agent-service-toolkit !2025-03-2825528|Full toolkit for running an AI agent service built with LangGraph, FastAPI and Streamlit| | 1045|databricks/dbrx !2025-03-2825460|Code examples and resources for DBRX, a large language model developed by Databricks| | 1046|lamini-ai/lamini !2025-03-2825271 |Official repo for Lamini's data generator for generating instructions to train instruction-following LLMs| | 1047|mshumer/gpt-author !2025-03-282510-1|| | 1048|TMElyralab/MusePose !2025-03-2824971|MusePose: a Pose-Driven Image-to-Video Framework for Virtual Human Generation| | 1049|Kludex/fastapi-tips !2025-03-2824974|FastAPI Tips by The FastAPI Expert!| | 1050|openai/simple-evals !2025-03-2824813|| | 1051|iterative/datachain !2025-03-2824732|AI-data warehouse to enrich, transform and analyze data from cloud storages| | 1052|girafe-ai/ml-course !2025-03-2824703|Open Machine Learning course| | 1053|kevmo314/magic-copy !2025-03-2824620 |Magic Copy is a Chrome extension that uses Meta's Segment Anything Model to extract a foreground object from an image and copy it to the clipboard.| | 1054|Eladlev/AutoPrompt !2025-03-2824432|A framework for prompt tuning using Intent-based Prompt Calibration| | 1055|OpenBMB/CPM-Bee !2025-03-282434-1 |A bilingual large-scale model with trillions of parameters| | 1056|IDEA-Research/T-Rex !2025-03-2824310|T-Rex2: Towards Generic Object Detection via Text-Visual Prompt Synergy| | 1057|microsoft/genaiscript !2025-03-2824202|Automatable GenAI Scripting| | 1058|paulpierre/RasaGPT !2025-03-2824090 |💬 RasaGPT is the first headless LLM chatbot platform built on top of Rasa and Langchain. Built w/ Rasa, FastAPI, Langchain, LlamaIndex, SQLModel, pgvector, ngrok, telegram| | 1059|ashishpatel26/LLM-Finetuning !2025-03-2823911|LLM Finetuning with peft| | 1060|SoraWebui/SoraWebui !2025-03-2823570|SoraWebui is an open-source Sora web client, enabling users to easily create videos from text with OpenAI's Sora model.| | 1061|6drf21e/ChatTTScolab !2025-03-2823491|🚀 一键部署(含离线整合包)!基于 ChatTTS ,支持音色抽卡、长音频生成和分角色朗读。简单易用,无需复杂安装。| | 1062|Azure/PyRIT !2025-03-2823343|The Python Risk Identification Tool for generative AI (PyRIT) is an open access automation framework to empower security professionals and machine learning engineers to proactively find risks in their generative AI systems.| | 1063|tencent-ailab/V-Express !2025-03-2823201|V-Express aims to generate a talking head video under the control of a reference image, an audio, and a sequence of V-Kps images.| | 1064|THUDM/CogVLM2 !2025-03-2823170|GPT4V-level open-source multi-modal model based on Llama3-8B| | 1065|dvmazur/mixtral-offloading !2025-03-2823001|Run Mixtral-8x7B models in Colab or consumer desktops| | 1066|semanser/codel !2025-03-2822950|✨ Fully autonomous AI Agent that can perform complicated tasks and projects using terminal, browser, and editor.| | 1067|mshumer/gpt-investor !2025-03-2822590|| | 1068|aixcoder-plugin/aiXcoder-7B !2025-03-2822550|official repository of aiXcoder-7B Code Large Language Model| | 1069|Azure-Samples/graphrag-accelerator !2025-03-2822503|One-click deploy of a Knowledge Graph powered RAG (GraphRAG) in Azure| | 1070|emcf/engshell !2025-03-2821830 |An English-language shell for any OS, powered by LLMs| | 1071|hncboy/chatgpt-web-java !2025-03-2821771|ChatGPT project developed in Java, based on Spring Boot 3 and JDK 17, supports both AccessToken and ApiKey modes| | 1072|openai/consistencydecoder !2025-03-2821692|Consistency Distilled Diff VAE| | 1073|Alpha-VLLM/Lumina-T2X !2025-03-2821681|Lumina-T2X is a unified framework for Text to Any Modality Generation| | 1074|bghira/SimpleTuner !2025-03-2821612|A general fine-tuning kit geared toward Stable Diffusion 2.1, Stable Diffusion 3, DeepFloyd, and SDXL.| | 1075|JiauZhang/DragGAN !2025-03-2821530 |Implementation of DragGAN: Interactive Point-based Manipulation on the Generative Image Manifold| | 1076|cgpotts/cs224u !2025-03-2821390|Code for Stanford CS224u| | 1077|PKU-YuanGroup/MoE-LLaVA !2025-03-2821300|Mixture-of-Experts for Large Vision-Language Models| | 1078|darrenburns/elia !2025-03-2820831|A snappy, keyboard-centric terminal user interface for interacting with large language models. Chat with ChatGPT, Claude, Llama 3, Phi 3, Mistral, Gemma and more.| | 1079|ageerle/ruoyi-ai !2025-03-28207898|RuoYi AI 是一个全栈式 AI 开发平台,旨在帮助开发者快速构建和部署个性化的 AI 应用。| | 1080|NVIDIA/gpu-operator !2025-03-2820510|NVIDIA GPU Operator creates/configures/manages GPUs atop Kubernetes| | 1081|BAAI-Agents/Cradle !2025-03-2820481|The Cradle framework is a first attempt at General Computer Control (GCC). Cradle supports agents to ace any computer task by enabling strong reasoning abilities, self-improvment, and skill curation, in a standardized general environment with minimal requirements.| | 1082|microsoft/aici !2025-03-2820080|AICI: Prompts as (Wasm) Programs| | 1083|PRIS-CV/DemoFusion !2025-03-2820040|Let us democratise high-resolution generation! (arXiv 2023)| | 1084|apple/axlearn !2025-03-2820012|An Extensible Deep Learning Library| | 1085|naver/mast3r !2025-03-2819685|Grounding Image Matching in 3D with MASt3R| | 1086|liltom-eth/llama2-webui !2025-03-281958-1|Run Llama 2 locally with gradio UI on GPU or CPU from anywhere (Linux/Windows/Mac). Supporting Llama-2-7B/13B/70B with 8-bit, 4-bit. Supporting GPU inference (6 GB VRAM) and CPU inference.| | 1087|GaParmar/img2img-turbo !2025-03-2819582|One-step image-to-image with Stable Diffusion turbo: sketch2image, day2night, and more| | 1088|Niek/chatgpt-web !2025-03-2819560|ChatGPT web interface using the OpenAI API| | 1089|huggingface/cookbook !2025-03-2819421|Open-source AI cookbook| | 1090|pytorch/ao !2025-03-2819241|PyTorch native quantization and sparsity for training and inference| | 1091|emcie-co/parlant !2025-03-2819053|The behavior guidance framework for customer-facing LLM agents| | 1092|ymcui/Chinese-LLaMA-Alpaca-3 !2025-03-2818980|中文羊驼大模型三期项目 (Chinese Llama-3 LLMs) developed from Meta Llama 3| | 1093|Nutlope/notesGPT !2025-03-2818811|Record voice notes & transcribe, summarize, and get tasks| | 1094|InstantStyle/InstantStyle !2025-03-2818791|InstantStyle: Free Lunch towards Style-Preserving in Text-to-Image Generation 🔥| | 1095|idaholab/moose !2025-03-2818771|Multiphysics Object Oriented Simulation Environment| | 1096|The-OpenROAD-Project/OpenROAD !2025-03-2818351|OpenROAD's unified application implementing an RTL-to-GDS Flow. Documentation at https://openroad.readthedocs.io/en/latest/| | 1097|alibaba/spring-ai-alibaba !2025-03-281831121|Agentic AI Framework for Java Developers| | 1098|ytongbai/LVM !2025-03-2817990|Sequential Modeling Enables Scalable Learning for Large Vision Models| | 1099|microsoft/sample-app-aoai-chatGPT !2025-03-2817981|[PREVIEW] Sample code for a simple web chat experience targeting chatGPT through AOAI.| | 1100|AI-Citizen/SolidGPT !2025-03-2817830|Chat everything with your code repository, ask repository level code questions, and discuss your requirements. AI Scan and learning your code repository, provide you code repository level answer🧱 🧱| | 1101|YangLing0818/RPG-DiffusionMaster !2025-03-2817784|Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs (PRG)| | 1102|kyegomez/BitNet !2025-03-2817710|Implementation of "BitNet: Scaling 1-bit Transformers for Large Language Models" in pytorch| | 1103|eloialonso/diamond !2025-03-2817671|DIAMOND (DIffusion As a Model Of eNvironment Dreams) is a reinforcement learning agent trained in a diffusion world model.| | 1104|flowdriveai/flowpilot !2025-03-2817250|flow-pilot is an openpilot based driver assistance system that runs on linux, windows and android powered machines.| | 1105|xlang-ai/OSWorld !2025-03-2817200|OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments| | 1106|linyiLYi/snake-ai !2025-03-2817031|An AI agent that beats the classic game "Snake".| | 1107|baaivision/Emu !2025-03-2816991|Emu Series: Generative Multimodal Models from BAAI| | 1108|kevmo314/scuda !2025-03-2816870|SCUDA is a GPU over IP bridge allowing GPUs on remote machines to be attached to CPU-only machines.| | 1109|SharifiZarchi/IntroductiontoMachineLearning !2025-03-2816701|دوره‌ی مقدمه‌ای بر یادگیری ماشین، برای دانشجویان| | 1110|google/maxtext !2025-03-2816670|A simple, performant and scalable Jax LLM!| | 1111|ml-explore/mlx-swift-examples !2025-03-2816471|Examples using MLX Swift| | 1112|unitreerobotics/unitreerlgym !2025-03-2816256|| | 1113|collabora/WhisperFusion !2025-03-2815901|WhisperFusion builds upon the capabilities of WhisperLive and WhisperSpeech to provide a seamless conversations with an AI.| | 1114|lichao-sun/Mora !2025-03-2815520|Mora: More like Sora for Generalist Video Generation| | 1115|GoogleCloudPlatform/localllm !2025-03-2815370|Run LLMs locally on Cloud Workstations| | 1116|TencentARC/BrushNet !2025-03-2815330|The official implementation of paper "BrushNet: A Plug-and-Play Image Inpainting Model with Decomposed Dual-Branch Diffusion"| | 1117|ai-christianson/RA.Aid !2025-03-2815288|Develop software autonomously.| | 1118|stephansturges/WALDO !2025-03-2815170|Whereabouts Ascertainment for Low-lying Detectable Objects. The SOTA in FOSS AI for drones!| | 1119|skills/copilot-codespaces-vscode !2025-03-2815112|Develop with AI-powered code suggestions using GitHub Copilot and VS Code| | 1120|andrewnguonly/Lumos !2025-03-2814920|A RAG LLM co-pilot for browsing the web, powered by local LLMs| | 1121|TeamNewPipe/NewPipeExtractor !2025-03-2814811|NewPipe's core library for extracting data from streaming sites| | 1122|mhamilton723/FeatUp !2025-03-2814770|Official code for "FeatUp: A Model-Agnostic Frameworkfor Features at Any Resolution" ICLR 2024| | 1123|AnswerDotAI/fsdpqlora !2025-03-2814671|Training LLMs with QLoRA + FSDP| | 1124|jgravelle/AutoGroq !2025-03-2814330|| | 1125|OpenGenerativeAI/llm-colosseum !2025-03-2814130|Benchmark LLMs by fighting in Street Fighter 3! The new way to evaluate the quality of an LLM| | 1126|microsoft/vscode-ai-toolkit !2025-03-2814000|| | 1127|McGill-NLP/webllama !2025-03-2813930|Llama-3 agents that can browse the web by following instructions and talking to you| | 1128|lucidrains/self-rewarding-lm-pytorch !2025-03-2813760|Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI| | 1129|ishaan1013/sandbox !2025-03-2813650|A cloud-based code editing environment with an AI copilot and real-time collaboration.| | 1130|goatcorp/Dalamud !2025-03-2813275|FFXIV plugin framework and API| | 1131|Lightning-AI/lightning-thunder !2025-03-2813151|Make PyTorch models Lightning fast! Thunder is a source to source compiler for PyTorch. It enables using different hardware executors at once.| | 1132|PKU-YuanGroup/MagicTime !2025-03-2813052|MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators| | 1133|SakanaAI/evolutionary-model-merge !2025-03-2813000|Official repository of Evolutionary Optimization of Model Merging Recipes| | 1134|a-real-ai/pywinassistant !2025-03-2812950|The first open source Large Action Model generalist Artificial Narrow Intelligence that controls completely human user interfaces by only using natural language. PyWinAssistant utilizes Visualization-of-Thought Elicits Spatial Reasoning in Large Language Models.| | 1135|TraceMachina/nativelink !2025-03-2812630|NativeLink is an open source high-performance build cache and remote execution server, compatible with Bazel, Buck2, Reclient, and other RBE-compatible build systems. It offers drastically faster builds, reduced test flakiness, and significant infrastructure cost savings.| | 1136|MLSysOps/MLE-agent !2025-03-2812500|🤖 MLE-Agent: Your intelligent companion for seamless AI engineering and research. 🔍 Integrate with arxiv and paper with code to provide better code/research plans 🧰 OpenAI, Ollama, etc supported. 🎆 Code RAG| | 1137|wpilibsuite/allwpilib !2025-03-2811610|Official Repository of WPILibJ and WPILibC| | 1138|elfvingralf/macOSpilot-ai-assistant !2025-03-2811470|Voice + Vision powered AI assistant that answers questions about any application, in context and in audio.| | 1139|langchain-ai/langchain-extract !2025-03-2811210|🦜⛏️ Did you say you like data?| | 1140|FoundationVision/GLEE !2025-03-2811120|【CVPR2024】GLEE: General Object Foundation Model for Images and Videos at Scale| | 1141|Profluent-AI/OpenCRISPR !2025-03-2810990|AI-generated gene editing systems| | 1142|zju3dv/EasyVolcap !2025-03-2810821|[SIGGRAPH Asia 2023 (Technical Communications)] EasyVolcap: Accelerating Neural Volumetric Video Research| | 1143|PaddlePaddle/PaddleHelix !2025-03-2810560|Bio-Computing Platform Featuring Large-Scale Representation Learning and Multi-Task Deep Learning “螺旋桨”生物计算工具集| | 1144|myshell-ai/JetMoE !2025-03-289800|Reaching LLaMA2 Performance with 0.1M Dollars| | 1145|likejazz/llama3.np !2025-03-289770|llama3.np is pure NumPy implementation for Llama 3 model.| | 1146|mustafaaljadery/gemma-2B-10M !2025-03-289500|Gemma 2B with 10M context length using Infini-attention.| | 1147|HITsz-TMG/FilmAgent !2025-03-289382|Resources of our paper "FilmAgent: A Multi-Agent Framework for End-to-End Film Automation in Virtual 3D Spaces". New versions in the making!| | 1148|aws-samples/amazon-bedrock-samples !2025-03-289362|This repository contains examples for customers to get started using the Amazon Bedrock Service. This contains examples for all available foundational models| | 1149|Akkudoktor-EOS/EOS !2025-03-2893154|This repository features an Energy Optimization System (EOS) that optimizes energy distribution, usage for batteries, heat pumps& household devices. It includes predictive models for electricity prices (planned), load forecasting& dynamic optimization to maximize energy efficiency & minimize costs. Founder Dr. Andreas Schmitz (YouTube @akkudoktor)| Tip: | symbol| rule | | :----| :---- | |🔥 | 256 1k| |!green-up-arrow.svg !red-down-arrow | ranking up / down| |⭐ | on trending page today| [Back to Top] Tools | No. | Tool | Description | | ----:|:----------------------------------------------- |:------------------------------------------------------------------------------------------- | | 1 | ChatGPT | A sibling model to InstructGPT, which is trained to follow instructions in a prompt and provide a detailed response | | 2 | DALL·E 2 | Create original, realistic images and art from a text description | | 3 | Murf AI | AI enabled, real people's voices| | 4 | Midjourney | An independent research lab that produces an artificial intelligence program under the same name that creates images from textual descriptions, used in Discord | 5 | Make-A-Video | Make-A-Video is a state-of-the-art AI system that generates videos from text | | 6 | Creative Reality™ Studio by D-ID| Use generative AI to create future-facing videos| | 7 | chat.D-ID| The First App Enabling Face-to-Face Conversations with ChatGPT| | 8 | Notion AI| Access the limitless power of AI, right inside Notion. Work faster. Write better. Think bigger. | | 9 | Runway| Text to Video with Gen-2 | | 10 | Resemble AI| Resemble’s AI voice generator lets you create human–like voice overs in seconds | | 11 | Cursor| Write, edit, and chat about your code with a powerful AI | | 12 | Hugging Face| Build, train and deploy state of the art models powered by the reference open source in machine learning | | 13 | Claude | A next-generation AI assistant for your tasks, no matter the scale | | 14 | Poe| Poe lets you ask questions, get instant answers, and have back-and-forth conversations with AI. Gives access to GPT-4, gpt-3.5-turbo, Claude from Anthropic, and a variety of other bots| [Back to Top] Websites | No. | WebSite |Description | | ----:|:------------------------------------------ |:---------------------------------------------------------------------------------------- | | 1 | OpenAI | An artificial intelligence research lab | | 2 | Bard | Base Google's LaMDA chatbots and pull from internet | | 3 | ERNIE Bot | Baidu’s new generation knowledge-enhanced large language model is a new member of the Wenxin large model family | | 4 | DALL·E 2 | An AI system that can create realistic images and art from a description in natural language | | 5 | Whisper | A general-purpose speech recognition model | | 6| CivitAI| A platform that makes it easy for people to share and discover resources for creating AI art| | 7|D-ID| D-ID’s Generative AI enables users to transform any picture or video into extraordinary experiences| | 8| Nvidia eDiff-I| Text-to-Image Diffusion Models with Ensemble of Expert Denoisers | | 9| Stability AI| The world's leading open source generative AI company which opened source Stable Diffusion | | 10| Meta AI| Whether it be research, product or infrastructure development, we’re driven to innovate responsibly with AI to benefit the world | | 11| ANTHROPIC| AI research and products that put safety at the frontier | [Back to Top] Reports&Papers | No. | Report&Paper | Description | |:---- |:-------------------------------------------------------------------------------------------------------------- |:---------------------------------------------------- | | 1 | GPT-4 Technical Report | GPT-4 Technical Report | | 2 | mli/paper-reading | Deep learning classics and new papers are read carefully paragraph by paragraph. | | 3 | labmlai/annotateddeeplearningpaperimplementations| A collection of simple PyTorch implementations of neural networks and related algorithms, which are documented with explanations | | 4 | Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models | Talking, Drawing and Editing with Visual Foundation Models | | 5 | OpenAI Research | The latest research report and papers from OpenAI | | 6 | Make-A-Video: Text-to-Video Generation without Text-Video Data|Meta's Text-to-Video Generation| | 7 | eDiff-I: Text-to-Image Diffusion Models with Ensemble of Expert Denoisers| Nvidia eDiff-I - New generation of generative AI content creation tool | | 8 | Training an Assistant-style Chatbot with Large Scale Data Distillation from GPT-3.5-Turbo | 2023 GPT4All Technical Report | | 9 | Segment Anything| Meta Segment Anything | | 10 | LLaMA: Open and Efficient Foundation Language Models| LLaMA: a collection of foundation language models ranging from 7B to 65B parameters| | 11 | papers-we-love/papers-we-love |Papers from the computer science community to read and discuss| | 12 | CVPR 2023 papers |The most exciting and influential CVPR 2023 papers| [Back to Top] Tutorials | No. | Tutorial | Description| |:---- |:---------------------------------------------------------------- | --- | | 1 | Coursera - Machine Learning | The Machine Learning Specialization Course taught by Dr. Andrew Ng| | 2 | microsoft/ML-For-Beginners | 12 weeks, 26 lessons, 52 quizzes, classic Machine Learning for all| | 3 | ChatGPT Prompt Engineering for Developers | This short course taught by Isa Fulford (OpenAI) and Andrew Ng (DeepLearning.AI) will teach how to use a large language model (LLM) to quickly build new and powerful applications | | 4 | Dive into Deep Learning |Targeting Chinese readers, functional and open for discussion. The Chinese and English versions are used for teaching in over 400 universities across more than 60 countries | | 5 | AI Expert Roadmap | Roadmap to becoming an Artificial Intelligence Expert in 2022 | | 6 | Computer Science courses |List of Computer Science courses with video lectures| | 7 | Machine Learning with Python | Machine Learning with Python Certification on freeCodeCamp| | 8 | Building Systems with the ChatGPT API | This short course taught by Isa Fulford (OpenAI) and Andrew Ng (DeepLearning.AI), you will learn how to automate complex workflows using chain calls to a large language model| | 9 | LangChain for LLM Application Development | This short course taught by Harrison Chase (Co-Founder and CEO at LangChain) and Andrew Ng. you will gain essential skills in expanding the use cases and capabilities of language models in application development using the LangChain framework| | 10 | How Diffusion Models Work | This short course taught by Sharon Zhou (CEO, Co-founder, Lamini). you will gain a deep familiarity with the diffusion process and the models which carry it out. More than simply pulling in a pre-built model or using an API, this course will teach you to build a diffusion model from scratch| | 11 | Free Programming Books For AI |📚 Freely available programming books for AI | | 12 | microsoft/AI-For-Beginners |12 Weeks, 24 Lessons, AI for All!| | 13 | hemansnation/God-Level-Data-Science-ML-Full-Stack |A collection of scientific methods, processes, algorithms, and systems to build stories & models. This roadmap contains 16 Chapters, whether you are a fresher in the field or an experienced professional who wants to transition into Data Science & AI| | 14 | datawhalechina/prompt-engineering-for-developers |Chinese version of Andrew Ng's Big Model Series Courses, including "Prompt Engineering", "Building System", and "LangChain"| | 15 | ossu/computer-science |🎓 Path to a free self-taught education in Computer Science!| | 16 | microsoft/Data-Science-For-Beginners | 10 Weeks, 20 Lessons, Data Science for All! | |17 |jwasham/coding-interview-university !2023-09-29268215336 |A complete computer science study plan to become a software engineer.| [Back to Top] Thanks If this project has been helpful to you in any way, please give it a ⭐️ by clicking on the star.

h2o-llmstudio
github
LLM Vibe Score0.499
Human Vibe Score0.04822694170894296
h2oaiMar 28, 2025

h2o-llmstudio

Welcome to H2O LLM Studio, a framework and no-code GUI designed for fine-tuning state-of-the-art large language models (LLMs). Jump to With H2O LLM Studio, you can Quickstart What's New Setup Recommended Install Virtual Environments Run H2O LLM Studio GUI Run H2O LLM Studio GUI using Docker Run H2O LLM Studio with command line interface (CLI) Troubleshooting Data format and example data Training your model Example: Run on OASST data via CLI Model checkpoints Documentation Contributing License With H2O LLM Studio, you can easily and effectively fine-tune LLMs without the need for any coding experience. use a graphic user interface (GUI) specially designed for large language models. finetune any LLM using a large variety of hyperparameters. use recent finetuning techniques such as Low-Rank Adaptation (LoRA) and 8-bit model training with a low memory footprint. use Reinforcement Learning (RL) to finetune your model (experimental) use advanced evaluation metrics to judge generated answers by the model. track and compare your model performance visually. In addition, Neptune and W&B integration can be used. chat with your model and get instant feedback on your model performance. easily export your model to the Hugging Face Hub and share it with the community. Quickstart For questions, discussing, or just hanging out, come and join our Discord! Use cloud-based runpod.io instance to run the H2O LLM Studio GUI. Using CLI for fine-tuning LLMs: What's New PR 788 New problem type for Causal Regression Modeling allows to train single target regression data using LLMs. PR 747 Fully removed RLHF in favor of DPO/IPO/KTO optimization. PR 741 Removing separate max length settings for prompt and answer in favor of a single maxlength settings better resembling chattemplate functionality from transformers. PR 592 Added KTOPairLoss for DPO modeling allowing to train models with simple preference data. Data currently needs to be manually prepared by randomly matching positive and negative examples as pairs. PR 592 Starting to deprecate RLHF in favor of DPO/IPO optimization. Training is disabled, but old experiments are still viewable. RLHF will be fully removed in a future release. PR 530 Introduced a new problem type for DPO/IPO optimization. This optimization technique can be used as an alternative to RLHF. PR 288 Introduced Deepspeed for sharded training allowing to train larger models on machines with multiple GPUs. Requires NVLink. This feature replaces FSDP and offers more flexibility. Deepspeed requires a system installation of cudatoolkit and we recommend using version 12.1. See Recommended Install. PR 449 New problem type for Causal Classification Modeling allows to train binary and multiclass models using LLMs. PR 364 User secrets are now handled more securely and flexible. Support for handling secrets using the 'keyring' library was added. User settings are tried to be migrated automatically. Please note that due to current rapid development we cannot guarantee full backwards compatibility of new functionality. We thus recommend to pin the version of the framework to the one you used for your experiments. For resetting, please delete/backup your data and output folders. Setup H2O LLM Studio requires a machine with Ubuntu 16.04+ and at least one recent Nvidia GPU with Nvidia drivers version >= 470.57.02. For larger models, we recommend at least 24GB of GPU memory. For more information about installation prerequisites, see the Set up H2O LLM Studio guide in the documentation. For a performance comparison of different GPUs, see the H2O LLM Studio performance guide in the documentation. Recommended Install The recommended way to install H2O LLM Studio is using pipenv with Python 3.10. To install Python 3.10 on Ubuntu 16.04+, execute the following commands: System installs (Python 3.10) Installing NVIDIA Drivers (if required) If deploying on a 'bare metal' machine running Ubuntu, one may need to install the required Nvidia drivers and CUDA. The following commands show how to retrieve the latest drivers for a machine running Ubuntu 20.04 as an example. One can update the following based on their OS. alternatively, one can install cudatoolkits in a conda environment: Virtual environments We offer various ways of setting up the necessary python environment. Pipenv virtual environment The following command will create a virtual environment using pipenv and will install the dependencies using pipenv: If you are having troubles installing the flash_attn package, consider running instead. This will install the dependencies without the flash_attn package. Note that this will disable the use of Flash Attention 2 and model training will be slower and consume more memory. Nightly Conda virtual environment You can also setup a conda virtual environment that can also deviate from the recommended setup. The contains a command that installs a fresh conda environment with CUDA 12.4 and current nightly PyTorch. Using requirements.txt If you wish to use another virtual environment, you can also install the dependencies using the requirements.txt file: Run H2O LLM Studio GUI You can start H2O LLM Studio using the following command: This command will start the H2O wave server and app. Navigate to (we recommend using Chrome) to access H2O LLM Studio and start fine-tuning your models! If you are running H2O LLM Studio with a custom environment other than Pipenv, you need to start the app as follows: If you are using the nightly conda environment, you can run . Run H2O LLM Studio GUI using Docker Install Docker first by following instructions from NVIDIA Containers. Make sure to have nvidia-container-toolkit installed on your machine as outlined in the instructions. H2O LLM Studio images are stored in the h2oai dockerhub container repository. Navigate to (we recommend using Chrome) to access H2O LLM Studio and start fine-tuning your models! (Note other helpful docker commands are docker ps and docker kill.) If you prefer to build your own Docker image from source, follow the instructions below. Run H2O LLM Studio with command line interface (CLI) You can also use H2O LLM Studio with the command line interface (CLI) and specify the configuration .yaml file that contains all the experiment parameters. To finetune using H2O LLM Studio with CLI, activate the pipenv environment by running make shell, and then use the following command: To run on multiple GPUs in DDP mode, run the following command: By default, the framework will run on the first k GPUs. If you want to specify specific GPUs to run on, use the CUDAVISIBLEDEVICES environment variable before the command. To start an interactive chat with your trained model, use the following command: where experiment_name is the output folder of the experiment you want to chat with (see configuration). The interactive chat will also work with model that were finetuned using the UI. To publish the model to Hugging Face, use the following command: pathtoexperiment is the output folder of the experiment. device is the target device for running the model, either 'cpu' or 'cuda:0'. Default is 'cuda:0'. api_key is the Hugging Face API Key. If user logged in, it can be omitted. user_id is the Hugging Face user ID. If user logged in, it can be omitted. model_name is the name of the model to be published on Hugging Face. It can be omitted. safe_serialization is a flag indicating whether safe serialization should be used. Default is True. Troubleshooting If running on cloud based machines such as runpod, you may need to set the following environment variable to allow the H2O Wave server to accept connections from the proxy: If you are experiencing timeouts when running the H2O Wave server remotely, you can increase the timeout by setting the following environment variables: All default to 5 (seconds). Increase them if you are experiencing timeouts. Use -1 to disable the timeout. Data format and example data For details on the data format required when importing your data or example data that you can use to try out H2O LLM Studio, see Data format in the H2O LLM Studio documentation. Training your model With H2O LLM Studio, training your large language model is easy and intuitive. First, upload your dataset and then start training your model. Start by creating an experiment. You can then monitor and manage your experiment, compare experiments, or push the model to Hugging Face to share it with the community. Example: Run on OASST data via CLI As an example, you can run an experiment on the OASST data via CLI. For instructions, see Run an experiment on the OASST data guide in the H2O LLM Studio documentation. Model checkpoints All open-source datasets and models are posted on H2O.ai's Hugging Face page and our H2OGPT repository. Documentation Detailed documentation and frequently asked questions (FAQs) for H2O LLM Studio can be found at . If you wish to contribute to the docs, navigate to the /documentation folder of this repo and refer to the README.md for more information. Contributing We are happy to accept contributions to the H2O LLM Studio project. Please refer to the CONTRIBUTING.md file for more information. License H2O LLM Studio is licensed under the Apache 2.0 license. Please see the LICENSE file for more information.

rpaframework
github
LLM Vibe Score0.527
Human Vibe Score0.11594284776995417
robocorpMar 28, 2025

rpaframework

RPA Framework ============= REQUEST for user input! We are looking at improving our keyword usage to cover situations where developer might be struggling to smoothly write task for a Robot. Describe the situation where your implementation speed slows due to the lack of easier syntax. Comment HERE _ .. contents:: Table of Contents :local: :depth: 1 .. include-docs-readme Introduction RPA Framework is a collection of open-source libraries and tools for Robotic Process Automation (RPA), and it is designed to be used with both Robot Framework and Python. The goal is to offer well-documented and actively maintained core libraries for Software Robot Developers. Learn more about RPA at Robocorp Documentation_. The project is: 100% Open Source Sponsored by Robocorp_ Optimized for Robocorp Control Room and Developer Tools Accepting external contributions .. _Robot Framework: https://robotframework.org .. _Robot Framework Foundation: https://robotframework.org/foundation/ .. _Python: https://www.python.org/ .. _Robocorp: https://robocorp.com .. _Robocorp Documentation: https://robocorp.com/docs-robot-framework .. _Control Room: https://robocorp.com/docs/control-room .. _Developer Tools: https://robocorp.com/downloads .. _Installing Python Packages: https://robocorp.com/docs/setup/installing-python-package-dependencies Links ^^^^^ Homepage: `_ Documentation: _ PyPI: _ Release notes: _ RSS feed: _ .. image:: https://img.shields.io/github/actions/workflow/status/robocorp/rpaframework/main.yaml?style=for-the-badge :target: https://github.com/robocorp/rpaframework/actions/workflows/main.yaml :alt: Status .. image:: https://img.shields.io/pypi/dw/rpaframework?style=for-the-badge :target: https://pypi.python.org/pypi/rpaframework :alt: rpaframework .. image:: https://img.shields.io/pypi/l/rpaframework.svg?style=for-the-badge&color=brightgreen :target: http://www.apache.org/licenses/LICENSE-2.0.html :alt: License Packages .. image:: https://img.shields.io/pypi/v/rpaframework.svg?label=rpaframework&style=for-the-badge :target: https://pypi.python.org/pypi/rpaframework :alt: rpaframework latest version .. image:: https://img.shields.io/pypi/v/rpaframework-assistant.svg?label=rpaframework-assistant&style=for-the-badge :target: https://pypi.python.org/pypi/rpaframework-assistant :alt: rpaframework-assistant latest version .. image:: https://img.shields.io/pypi/v/rpaframework-aws.svg?label=rpaframework-aws&style=for-the-badge :target: https://pypi.python.org/pypi/rpaframework-aws :alt: rpaframework-aws latest version .. image:: https://img.shields.io/pypi/v/rpaframework-core.svg?label=rpaframework-core&style=for-the-badge :target: https://pypi.python.org/pypi/rpaframework-core :alt: rpaframework-core latest version .. image:: https://img.shields.io/pypi/v/rpaframework-google.svg?label=rpaframework-google&style=for-the-badge&color=blue :target: https://pypi.python.org/pypi/rpaframework-google :alt: rpaframework-google latest version .. image:: https://img.shields.io/pypi/v/rpaframework-hubspot.svg?label=rpaframework-hubspot&style=for-the-badge&color=blue :target: https://pypi.python.org/pypi/rpaframework-hubspot :alt: rpaframework-hubspot latest version .. image:: https://img.shields.io/pypi/v/rpaframework-openai.svg?label=rpaframework-openai&style=for-the-badge&color=blue :target: https://pypi.python.org/pypi/rpaframework-openai :alt: rpaframework-openai latest version .. image:: https://img.shields.io/pypi/v/rpaframework-pdf.svg?label=rpaframework-pdf&style=for-the-badge&color=blue :target: https://pypi.python.org/pypi/rpaframework-pdf :alt: rpaframework-pdf latest version .. image:: https://img.shields.io/pypi/v/rpaframework-recognition.svg?label=rpaframework-recognition&style=for-the-badge&color=blue :target: https://pypi.python.org/pypi/rpaframework-recognition :alt: rpaframework-recognition latest version .. image:: https://img.shields.io/pypi/v/rpaframework-windows.svg?label=rpaframework-windows&style=for-the-badge&color=blue :target: https://pypi.python.org/pypi/rpaframework-windows :alt: rpaframework-windows latest version From the above packages, rpaframework-core and rpaframework-recognition are support packages, which alone do not contain any libraries. Libraries The RPA Framework project currently includes the following libraries: The x in the PACKAGE column means that library is included in the rpaframework package and for example. x,pdf means that RPA.PDF library is provided in both the rpaframework and rpaframework-pdf packages. +----------------------------+-------------------------------------------------------+------------------------+ | LIBRARY NAME | DESCRIPTION | PACKAGE | +----------------------------+-------------------------------------------------------+------------------------+ | Archive_ | Archiving TAR and ZIP files | x | +----------------------------+-------------------------------------------------------+------------------------+ | Assistant_ | Display information to a user and request input. | assistant | +----------------------------+-------------------------------------------------------+------------------------+ | Browser.Selenium_ | Control browsers and automate the web | x | +----------------------------+-------------------------------------------------------+------------------------+ | Browser.Playwright_ | Newer way to control browsers | special (more below) | +----------------------------+-------------------------------------------------------+------------------------+ | Calendar_ | For date and time manipulations | x | +----------------------------+-------------------------------------------------------+------------------------+ | Cloud.AWS_ | Use Amazon AWS services | x,aws | +----------------------------+-------------------------------------------------------+------------------------+ | Cloud.Azure_ | Use Microsoft Azure services | x | +----------------------------+-------------------------------------------------------+------------------------+ | Cloud.Google_ | Use Google Cloud services | google | +----------------------------+-------------------------------------------------------+------------------------+ | Crypto_ | Common hashing and encryption operations | x | +----------------------------+-------------------------------------------------------+------------------------+ | Database_ | Interact with databases | x | +----------------------------+-------------------------------------------------------+------------------------+ | Desktop_ | Cross-platform desktop automation | x | +----------------------------+-------------------------------------------------------+------------------------+ | Desktop.Clipboard_ | Interact with the system clipboard | x | +----------------------------+-------------------------------------------------------+------------------------+ | Desktop.OperatingSystem_ | Read OS information and manipulate processes | x | +----------------------------+-------------------------------------------------------+------------------------+ | DocumentAI_ | Intelligent Document Processing wrapper | x | +----------------------------+-------------------------------------------------------+------------------------+ | DocumentAI.Base64AI_ | Intelligent Document Processing service | x | +----------------------------+-------------------------------------------------------+------------------------+ | DocumentAI.Nanonets_ | Intelligent Document Processing service | x | +----------------------------+-------------------------------------------------------+------------------------+ | Email.Exchange_ | E-Mail operations (Exchange protocol) | x | +----------------------------+-------------------------------------------------------+------------------------+ | Email.ImapSmtp_ | E-Mail operations (IMAP & SMTP) | x | +----------------------------+-------------------------------------------------------+------------------------+ | Excel.Application_ | Control the Excel desktop application | x | +----------------------------+-------------------------------------------------------+------------------------+ | Excel.Files_ | Manipulate Excel files directly | x | +----------------------------+-------------------------------------------------------+------------------------+ | FileSystem_ | Read and manipulate files and paths | x | +----------------------------+-------------------------------------------------------+------------------------+ | FTP_ | Interact with FTP servers | x | +----------------------------+-------------------------------------------------------+------------------------+ | HTTP_ | Interact directly with web APIs | x | +----------------------------+-------------------------------------------------------+------------------------+ | Hubspot_ | Access HubSpot CRM data objects | hubspot | +----------------------------+-------------------------------------------------------+------------------------+ | Images_ | Manipulate images | x | +----------------------------+-------------------------------------------------------+------------------------+ | JavaAccessBridge_ | Control Java applications | x | +----------------------------+-------------------------------------------------------+------------------------+ | JSON_ | Manipulate JSON objects | x | +----------------------------+-------------------------------------------------------+------------------------+ | MFA_ | Authenticate using one-time passwords (OTP) & OAuth2 | x | +----------------------------+-------------------------------------------------------+------------------------+ | Notifier_ | Notify messages using different services | x | +----------------------------+-------------------------------------------------------+------------------------+ | OpenAI_ | Artificial Intelligence service | openai | +----------------------------+-------------------------------------------------------+------------------------+ | Outlook.Application_ | Control the Outlook desktop application | x | +----------------------------+-------------------------------------------------------+------------------------+ | PDF_ | Read and create PDF documents | x,pdf | +----------------------------+-------------------------------------------------------+------------------------+ | Robocorp.Process_ | Use the Robocorp Process API | x | +----------------------------+-------------------------------------------------------+------------------------+ | Robocorp.WorkItems_ | Use the Robocorp Work Items API | x | +----------------------------+-------------------------------------------------------+------------------------+ | Robocorp.Vault_ | Use the Robocorp Secrets API | x | +----------------------------+-------------------------------------------------------+------------------------+ | Robocorp.Storage_ | Use the Robocorp Asset Storage API | x | +----------------------------+-------------------------------------------------------+------------------------+ | Salesforce_ | Salesforce operations | x | +----------------------------+-------------------------------------------------------+------------------------+ | SAP_ | Control SAP GUI desktop client | x | +----------------------------+-------------------------------------------------------+------------------------+ | Smartsheet_ | Access Smartsheet sheets | x | +----------------------------+-------------------------------------------------------+------------------------+ | Tables_ | Manipulate, sort, and filter tabular data | x | +----------------------------+-------------------------------------------------------+------------------------+ | Tasks_ | Control task execution | x | +----------------------------+-------------------------------------------------------+------------------------+ | Twitter_ | Twitter API interface | x | +----------------------------+-------------------------------------------------------+------------------------+ | Windows_ | Alternative library for Windows automation | x,windows | +----------------------------+-------------------------------------------------------+------------------------+ | Word.Application_ | Control the Word desktop application | x | +----------------------------+-------------------------------------------------------+------------------------+ .. _Archive: https://rpaframework.org/libraries/archive/ .. _Assistant: https://rpaframework.org/libraries/assistant/ .. Browser.Playwright: https://rpaframework.org/libraries/browserplaywright/ .. Browser.Selenium: https://rpaframework.org/libraries/browserselenium/ .. _Calendar: https://rpaframework.org/libraries/calendar/ .. Cloud.AWS: https://rpaframework.org/libraries/cloudaws/ .. Cloud.Azure: https://rpaframework.org/libraries/cloudazure/ .. Cloud.Google: https://rpaframework.org/libraries/cloudgoogle/ .. _Crypto: https://rpaframework.org/libraries/crypto/ .. _Database: https://rpaframework.org/libraries/database/ .. _Desktop: https://rpaframework.org/libraries/desktop/ .. Desktop.Clipboard: https://rpaframework.org/libraries/desktopclipboard/ .. Desktop.Operatingsystem: https://rpaframework.org/libraries/desktopoperatingsystem/ .. _DocumentAI: https://rpaframework.org/libraries/documentai .. DocumentAI.Base64AI: https://rpaframework.org/libraries/documentaibase64ai/ .. DocumentAI.Nanonets: https://rpaframework.org/libraries/documentainanonets/ .. Email.Exchange: https://rpaframework.org/libraries/emailexchange/ .. Email.ImapSmtp: https://rpaframework.org/libraries/emailimapsmtp/ .. Excel.Application: https://rpaframework.org/libraries/excelapplication/ .. Excel.Files: https://rpaframework.org/libraries/excelfiles/ .. _FileSystem: https://rpaframework.org/libraries/filesystem/ .. _FTP: https://rpaframework.org/libraries/ftp/ .. _HTTP: https://rpaframework.org/libraries/http/ .. _Hubspot: https://rpaframework.org/libraries/hubspot/ .. _Images: https://rpaframework.org/libraries/images/ .. _JavaAccessBridge: https://rpaframework.org/libraries/javaaccessbridge/ .. _JSON: https://rpaframework.org/libraries/json/ .. _MFA: https://rpaframework.org/libraries/mfa/ .. _Notifier: https://rpaframework.org/libraries/notifier/ .. _OpenAI: https://rpaframework.org/libraries/openai/ .. Outlook.Application: https://rpaframework.org/libraries/outlookapplication/ .. _PDF: https://rpaframework.org/libraries/pdf/ .. Robocorp.Process: https://rpaframework.org/libraries/robocorpprocess/ .. Robocorp.WorkItems: https://rpaframework.org/libraries/robocorpworkitems/ .. Robocorp.Vault: https://rpaframework.org/libraries/robocorpvault/ .. Robocorp.Storage: https://rpaframework.org/libraries/robocorpstorage/ .. _Salesforce: https://rpaframework.org/libraries/salesforce/ .. _SAP: https://rpaframework.org/libraries/sap/ .. _Smartsheet: https://rpaframework.org/libraries/smartsheet/ .. _Tables: https://rpaframework.org/libraries/tables/ .. _Tasks: https://rpaframework.org/libraries/tasks/ .. _Twitter: https://rpaframework.org/libraries/twitter/ .. _Windows: https://rpaframework.org/libraries/windows/ .. Word.Application: https://rpaframework.org/libraries/wordapplication/ Installation of RPA.Browser.Playwright The RPA.Browser.Playwright at the moment requires special installation, because of the package size and the post install step it needs to be fully installed. Minimum required conda.yaml to install Playwright: .. code-block:: yaml channels: conda-forge dependencies: python=3.10.14 nodejs=22.9.0 pip=24.0 pip: robotframework-browser==18.8.1 rpaframework==28.6.3 rccPostInstall: rfbrowser init Installation Learn about installing Python packages at Installing Python Packages_. Default installation method with Robocorp Developer Tools_ using conda.yaml: .. code-block:: yaml channels: conda-forge dependencies: python=3.10.14 pip=24.0 pip: rpaframework==28.6.3 To install all extra packages (including Playwright dependencies), you can use: .. code-block:: yaml channels: conda-forge dependencies: python=3.10.14 tesseract=5.4.1 nodejs=22.9.0 pip=24.0 pip: robotframework-browser==18.8.1 rpaframework==28.6.3 rpaframework-aws==5.3.3 rpaframework-google==9.0.2 rpaframework-recognition==5.2.5 rccPostInstall: rfbrowser init Separate installation of AWS, PDF and Windows libraries without the main rpaframework: .. code-block:: yaml channels: conda-forge dependencies: python=3.10.14 pip=24.0 pip: rpaframework-aws==5.3.3 included in the rpaframework as an extra rpaframework-pdf==7.3.3 included in the rpaframework by default rpaframework-windows==7.5.2 included in the rpaframework by default Installation method with pip using Python venv_: .. code-block:: shell python -m venv .venv source .venv/bin/activate pip install rpaframework .. note:: Python 3.8 or higher is required Example After installation the libraries can be directly imported inside Robot Framework_: .. code:: robotframework Settings Library RPA.Browser.Selenium Tasks Login as user Open available browser https://example.com Input text id:user-name ${USERNAME} Input text id:password ${PASSWORD} The libraries are also available inside Python_: .. code:: python from RPA.Browser.Selenium import Selenium lib = Selenium() lib.openavailablebrowser("https://example.com") lib.input_text("id:user-name", username) lib.input_text("id:password", password) Support and contact rpaframework.org _ for library documentation Robocorp Documentation_ for guides and tutorials #rpaframework channel in Robot Framework Slack_ if you have open questions or want to contribute Communicate with your fellow Software Robot Developers and Robocorp experts at Robocorp Developers Slack_ .. _Robot Framework Slack: https://robotframework-slack-invite.herokuapp.com/ .. _Robocorp Developers Slack: https://robocorp-developers.slack.com Contributing Found a bug? Missing a critical feature? Interested in contributing? Head over to the Contribution guide _ to see where to get started. Development Repository development is Python_ based and requires at minimum Python version 3.8+ installed on the development machine. The default Python version used in the Robocorp Robot template is 3.10.14 so it is a good choice for the version to install. Not recommended versions are 3.7.6 and 3.8.1, because they have issues with some of the dependencies related to rpaframework. At the time the newer Python versions starting from 3.12 are also not recommended, because some of the dependencies might cause issues. Repository development tooling is based on poetry and invoke. Poetry is the underlying tool used for compiling, building and running the package. Invoke is used for scripting purposes, for example for linting, testing and publishing tasks. Before writing any code, please read and acknowledge our extensive Dev Guide_. .. _Dev Guide: https://github.com/robocorp/rpaframework/blob/master/docs/source/contributing/development.md First steps to start developing: initial poetry configuration .. code:: shell poetry config virtualenvs.path null poetry config virtualenvs.in-project true poetry config repositories.devpi "https://devpi.robocorp.cloud/ci/test" git clone the repository #. create a new Git branch or switch to correct branch or stay in master branch some branch naming conventions feature/name-of-feature, hotfix/name-of-the-issue, release/number-of-release #. poetry install which install package with its dependencies into the .venv directory of the package, for example packages/main/.venv #. if testing against Robocorp Robot which is using devdata/env.json set environment variables or poetry build and use resulting .whl file (in the dist/ directory) in the Robot conda.yaml or poetry build and push resulting .whl file (in the dist/ directory) into a repository and use raw url to include it in the Robot conda.yaml another possibility for Robocorp internal development is to use Robocorp devpi instance, by poetry publish --ci and point conda.yaml to use rpaframework version in devpi #. poetry run python -m robot common ROBOT_ARGS from Robocorp Robot template: --report NONE --outputdir output --logtitle "Task log" #. poetry run python #. invoke lint to make sure that code formatting is according to rpaframework repository guidelines. It is possible and likely that Github action will fail the if developer has not linted the code changes. Code formatting is based on black and flake8 and those are run with the invoke lint. #. the library documentation can be created in the repository root (so called "meta" package level). The documentation is built by the docgen tools using the locally installed version of the project, local changes for the main package will be reflected each time you generate the docs, but if you want to see local changes for optional packages, you must utilize invoke install-local --package using the appropriate package name (e.g., rpaframework-aws). This will reinstall that package as a local editable version instead of from PyPI. Multiple such packages can be added by repeating the use of the --package option. In order to reset this, use invoke install --reset. poetry update and/or invoke install-local --package make docs open docs/build/html/index.html with the browser to view the changes or execute make local and navigate to localhost:8000 to view docs as a live local webpage. .. code-block:: toml Before [tool.poetry.dependencies] python = "^3.8" rpaframework = { path = "packages/main", extras = ["cv", "playwright", "aws"] } rpaframework-google = "^4.0.0" rpaframework-windows = "^4.0.0" After [tool.poetry.dependencies] python = "^3.8" rpaframework = { path = "packages/main", extras = ["cv", "playwright"] } rpaframework-aws = { path = "packages/aws" } rpaframework-google = "^4.0.0" rpaframework-windows = "^4.0.0" #. invoke test (this will run both Python unittests and robotframework tests defined in the packages tests/ directory) to run specific Python test: poetry run pytest path/to/test.py::test_function to run specific Robotframework test: inv testrobot -r -t #. git commit changes #. git push changes to remote #. create pull request from the branch describing changes included in the description #. update docs/source/releasenotes.rst with changes (commit and push) Packaging and publishing are done after changes have been merged into master branch. All the following steps should be done within master branch. #. git pull latest changes into master branch #. in the package directory containing changes execute invoke lint and invoke test #. update pyproject.toml with new version according to semantic versioning #. update docs/source/releasenotes.rst with changes #. in the repository root (so called "meta" package level) run command poetry update #. git commit changed poetry.lock files (on meta and target package level), releasenotes.rst and pyproject.toml with message "PACKAGE. version x.y.z" #. git push #. invoke publish after Github action on master branch is all green Some recommended tools for development Visual Studio Code_ as a code editor with following extensions: Sema4.ai_ Robot Framework Language Server_ GitLens_ Python extension_ GitHub Desktop_ will make version management less prone to errors .. _poetry: https://python-poetry.org .. _invoke: https://www.pyinvoke.org .. _Visual Studio Code: https://code.visualstudio.com .. _GitHub Desktop: https://desktop.github.com .. _Sema4.ai: https://marketplace.visualstudio.com/items?itemName=sema4ai.sema4ai .. _Robot Framework Language Server: https://marketplace.visualstudio.com/items?itemName=robocorp.robotframework-lsp .. _GitLens: https://marketplace.visualstudio.com/items?itemName=eamodio.gitlens .. _Python extension: https://marketplace.visualstudio.com/items?itemName=ms-python.python .. _black: https://pypi.org/project/black/ .. _flake8: https://pypi.org/project/flake8/ .. _venv: https://docs.python.org/3/library/venv.html License This project is open-source and licensed under the terms of the Apache License 2.0 `_.

awesome-quantum-machine-learning
github
LLM Vibe Score0.64
Human Vibe Score1
krishnakumarsekarMar 27, 2025

awesome-quantum-machine-learning

Awesome Quantum Machine Learning A curated list of awesome quantum machine learning algorithms,study materials,libraries and software (by language). Table of Contents INTRODUCTION Why Quantum Machine Learning? BASICS What is Quantum Mechanics? What is Quantum Computing? What is Topological Quantum Computing? Quantum Computing vs Classical Computing QUANTUM COMPUTING Atom Structure Photon wave Electron Fluctuation or spin States SuperPosition SuperPosition specific for machine learning(Quantum Walks) Classical Bit Quantum Bit or Qubit or Qbit Basic Gates in Quantum Computing Quantum Diode Quantum Transistor Quantum Processor Quantum Registery QRAM Quantum Entanglement QUANTUM COMPUTING MACHINE LEARNING BRIDGE Complex Numbers Tensors Tensors Network Oracle Hadamard transform Hilbert Space eigenvalues and eigenvectors Schr¨odinger Operators Quantum lambda calculus Quantum Amplitute Phase Qubits Encode and Decode convert classical bit to qubit Quantum Dirac and Kets Quantum Complexity Arbitrary State Generation QUANTUM ALGORITHMS Quantum Fourier Transform Variational-Quantum-Eigensolver Grovers Algorithm Shor's algorithm Hamiltonian Oracle Model Bernstein-Vazirani Algorithm Simon’s Algorithm Deutsch-Jozsa Algorithm Gradient Descent Phase Estimation Haar Tansform Quantum Ridgelet Transform Quantum NP Problem QUANTUM MACHINE LEARNING ALGORITHMS Quantum K-Nearest Neighbour Quantum K-Means Quantum Fuzzy C-Means Quantum Support Vector Machine Quantum Genetic Algorithm Quantum Hidden Morkov Models Quantum state classification with Bayesian methods Quantum Ant Colony Optimization Quantum Cellular Automata Quantum Classification using Principle Component Analysis Quantum Inspired Evolutionary Algorithm Quantum Approximate Optimization Algorithm Quantum Elephant Herding Optimization Quantum-behaved Particle Swarm Optimization Quantum Annealing Expectation-Maximization QAUNTUM NEURAL NETWORK Quantum perceptrons Qurons Quantum Auto Encoder Quantum Annealing Photonic Implementation of Quantum Neural Network Quantum Feed Forward Neural Network Quantum Boltzman Neural Network Quantum Neural Net Weight Storage Quantum Upside Down Neural Net Quantum Hamiltonian Neural Net QANN QPN SAL Quantum Hamiltonian Learning Compressed Quantum Hamiltonian Learning QAUNTUM STATISTICAL DATA ANALYSIS Quantum Probability Theory Kolmogorovian Theory Quantum Measurement Problem Intuitionistic Logic Heyting Algebra Quantum Filtering Paradoxes Quantum Stochastic Process Double Negation Quantum Stochastic Calculus Hamiltonian Calculus Quantum Ito's Formula Quantum Stochastic Differential Equations(QSDE) Quantum Stochastic Integration Itō Integral Quasiprobability Distributions Quantum Wiener Processes Quantum Statistical Ensemble Quantum Density Operator or Density Matrix Gibbs Canonical Ensemble Quantum Mean Quantum Variance Envariance Polynomial Optimization Quadratic Unconstrained Binary Optimization Quantum Gradient Descent Quantum Based Newton's Method for Constrained Optimization Quantum Based Newton's Method for UnConstrained Optimization Quantum Ensemble Quantum Topology Quantum Topological Data Analysis Quantum Bayesian Hypothesis Quantum Statistical Decision Theory Quantum Minimax Theorem Quantum Hunt-Stein Theorem Quantum Locally Asymptotic Normality Quantum Ising Model Quantum Metropolis Sampling Quantum Monte Carlo Approximation Quantum Bootstrapping Quantum Bootstrap Aggregation Quantum Decision Tree Classifier Quantum Outlier Detection Cholesky-Decomposition for Quantum Chemistry Quantum Statistical Inference Asymptotic Quantum Statistical Inference Quantum Gaussian Mixture Modal Quantum t-design Quantum Central Limit Theorem Quantum Hypothesis Testing Quantum Chi-squared and Goodness of Fit Testing Quantum Estimation Theory Quantum Way of Linear Regression Asymptotic Properties of Quantum Outlier Detection in Quantum Concepts QAUNTUM ARTIFICIAL INTELLIGENCE Heuristic Quantum Mechanics Consistent Quantum Reasoning Quantum Reinforcement Learning QAUNTUM COMPUTER VISION QUANTUM PROGRAMMING LANGUAGES , TOOLs and SOFTWARES ALL QUANTUM ALGORITHMS SOURCE CODES , GITHUBS QUANTUM HOT TOPICS Quantum Cognition Quantum Camera Quantum Mathematics Quantum Information Processing Quantum Image Processing Quantum Cryptography Quantum Elastic Search Quantum DNA Computing Adiabetic Quantum Computing Topological Big Data Anlytics using Quantum Hamiltonian Time Based Quantum Computing Deep Quantum Learning Quantum Tunneling Quantum Entanglment Quantum Eigen Spectrum Quantum Dots Quantum elctro dynamics Quantum teleportation Quantum Supremacy Quantum Zeno Effect Quantum Cohomology Quantum Chromodynamics Quantum Darwinism Quantum Coherence Quantum Decoherence Topological Quantum Computing Topological Quantum Field Theory Quantum Knots Topological Entanglment Boson Sampling Quantum Convolutional Code Stabilizer Code Quantum Chaos Quantum Game Theory Quantum Channel Tensor Space Theory Quantum Leap Quantum Mechanics for Time Travel Quantum Secured Block Chain Quantum Internet Quantum Optical Network Quantum Interference Quantum Optical Network Quantum Operating System Electron Fractionalization Flip-Flop Quantum Computer Quantum Information with Gaussian States Quantum Anomaly Detection Distributed Secure Quantum Machine Learning Decentralized Quantum Machine Learning Artificial Agents for Quantum Designs Light Based Quantum Chips for AI Training QUANTUM STATE PREPARATION ALGORITHM FOR MACHINE LEARNING Pure Quantum State Product State Matrix Product State Greenberger–Horne–Zeilinger State W state AKLT model Majumdar–Ghosh Model Multistate Landau–Zener Models Projected entangled-pair States Infinite Projected entangled-pair States Corner Transfer Matrix Method Tensor-entanglement Renormalization Tree Tensor Network for Supervised Learning QUANTUM MACHINE LEARNING VS DEEP LEARNING QUANTUM MEETUPS QUANTUM GOOGLE GROUPS QUANTUM BASED COMPANIES QUANTUM LINKEDLIN QUANTUM BASED DEGREES CONSOLIDATED QUANTUM ML BOOKS CONSOLIDATED QUANTUM ML VIDEOS CONSOLIDATED QUANTUM ML Reserach Papers CONSOLIDATED QUANTUM ML Reserach Scientist RECENT QUANTUM UPDATES FORUM ,PAGES AND NEWSLETTER INTRODUCTION Why Quantum Machine Learning? Machine Learning(ML) is just a term in recent days but the work effort start from 18th century. What is Machine Learning ? , In Simple word the answer is making the computer or application to learn themselves . So its totally related with computing fields like computer science and IT ? ,The answer is not true . ML is a common platform which is mingled in all the aspects of the life from agriculture to mechanics . Computing is a key component to use ML easily and effectively . To be more clear ,Who is the mother of ML ?, As no option Mathematics is the mother of ML . The world tremendous invention complex numbers given birth to this field . Applying mathematics to the real life problem always gives a solution . From Neural Network to the complex DNA is running under some specific mathematical formulas and theorems. As computing technology growing faster and faster mathematics entered into this field and makes the solution via computing to the real world . In the computing technology timeline once a certain achievements reached peoples interested to use advanced mathematical ideas such as complex numbers ,eigen etc and its the kick start for the ML field such as Artificial Neural Network ,DNA Computing etc. Now the main question, why this field is getting boomed now a days ? , From the business perspective , 8-10 Years before during the kick start time for ML ,the big barrier is to merge mathematics into computing field . people knows well in computing has no idea on mathematics and research mathematician has no idea on what is computing . The education as well as the Job Opportunities is like that in that time . Even if a person tried to study both then the business value for making a product be not good. Then the top product companies like Google ,IBM ,Microsoft decided to form a team with mathematician ,a physician and a computer science person to come up with various ideas in this field . Success of this team made some wonderful products and they started by providing cloud services using this product . Now we are in this stage. So what's next ? , As mathematics reached the level of time travel concepts but the computing is still running under classical mechanics . the companies understood, the computing field must have a change from classical to quantum, and they started working on the big Quantum computing field, and the market named this field as Quantum Information Science .The kick start is from Google and IBM with the Quantum Computing processor (D-Wave) for making Quantum Neural Network .The field of Quantum Computer Science and Quantum Information Science will do a big change in AI in the next 10 years. Waiting to see that........... .(google, ibm). References D-Wave - Owner of a quantum processor Google - Quantum AI Lab IBM - Quantum Computer Lab Quora - Question Regarding future of quantum AI NASA - NASA Quantum Works Youtube - Google Video of a Quantum Processor external-link - MIT Review microsoft new product - Newly Launched Microsoft Quantum Language and Development Kit microsoft - Microsoft Quantum Related Works Google2 - Google Quantum Machine Learning Blog BBC - About Google Quantum Supremacy,IBM Quantum Computer and Microsoft Q Google Quantum Supremacy - Latest 2019 Google Quantum Supremacy Achievement IBM Quantum Supremacy - IBM Talk on Quantum Supremacy as a Primer VICE on the fight - IBM Message on Google Quantum Supremacy IBM Zurich Quantum Safe Cryptography - An interesting startup to replace all our Certificate Authority Via Cloud and IBM Q BASICS What is Quantum Mechanics? In a single line study of an electron moved out of the atom then its classical mechanic ,vibrates inside the atom its quantum mechanics WIKIPEDIA - Basic History and outline LIVESCIENCE. - A survey YOUTUBE - Simple Animation Video Explanining Great. What is Quantum Computing? A way of parallel execution of multiple processess in a same time using qubit ,It reduces the computation time and size of the processor probably in neuro size WIKIPEDIA - Basic History and outline WEBOPEDIA. - A survey YOUTUBE - Simple Animation Video Explanining Great. Quantum Computing vs Classical Computing LINK - Basic outline Quantum Computing Atom Structure one line : Electron Orbiting around the nucleous in an eliptical format YOUTUBE - A nice animation video about the basic atom structure Photon Wave one line : Light nornmally called as wave transmitted as photons as similar as atoms in solid particles YOUTUBE - A nice animation video about the basic photon 1 YOUTUBE - A nice animation video about the basic photon 2 Electron Fluctuation or spin one line : When a laser light collide with solid particles the electrons of the atom will get spin between the orbitary layers of the atom ) YOUTUBE - A nice animation video about the basic Electron Spin 1 YOUTUBE - A nice animation video about the basic Electron Spin 2 YOUTUBE - A nice animation video about the basic Electron Spin 3 States one line : Put a point on the spinning electron ,if the point is in the top then state 1 and its in bottom state 0 YOUTUBE - A nice animation video about the Quantum States SuperPosition two line : During the spin of the electron the point may be in the middle of upper and lower position, So an effective decision needs to take on the point location either 0 or 1 . Better option to analyse it along with other electrons using probability and is called superposition YOUTUBE - A nice animation video about the Quantum Superposition SuperPosition specific for machine learning(Quantum Walks) one line : As due to computational complexity ,quantum computing only consider superposition between limited electrons ,In case to merge more than one set quantum walk be the idea YOUTUBE - A nice video about the Quantum Walks Classical Bits one line : If electron moved from one one atom to other ,from ground state to excited state a bit value 1 is used else bit value 0 used Qubit one line : The superposition value of states of a set of electrons is Qubit YOUTUBE - A nice video about the Quantum Bits 1 YOUTUBE - A nice video about the Bits and Qubits 2 Basic Gates in Quantum Computing one line : As like NOT, OR and AND , Basic Gates like NOT, Hadamard gate , SWAP, Phase shift etc can be made with quantum gates YOUTUBE - A nice video about the Quantum Gates Quantum Diode one line : Quantum Diodes using a different idea from normal diode, A bunch of laser photons trigger the electron to spin and the quantum magnetic flux will capture the information YOUTUBE - A nice video about the Quantum Diode Quantum Transistors one line : A transistor default have Source ,drain and gate ,Here source is photon wave ,drain is flux and gate is classical to quantum bits QUORA -Discussion about the Quantum Transistor YOUTUBE - Well Explained Quantum Processor one line : A nano integration circuit performing the quantum gates operation sorrounded by cooling units to reduce the tremendous amount of heat YOUTUBE - Well Explained Quantum Registery QRAM one line : Comapring the normal ram ,its ultrafast and very small in size ,the address location can be access using qubits superposition value ,for a very large memory set coherent superposition(address of address) be used PDF - very Well Explained QUANTUM COMPUTING MACHINE LEARNING BRIDGE Complex Numbers one line : Normally Waves Interference is in n dimensional structure , to find a polynomial equation n order curves ,better option is complex number YOUTUBE - Wonderful Series very super Explained Tensors one line : Vectors have a direction in 2D vector space ,If on a n dimensional vector space ,vectors direction can be specify with the tensor ,The best solution to find the superposition of a n vector electrons spin space is representing vectors as tensors and doing tensor calculus YOUTUBE - Wonderful super Explained tensors basics YOUTUBE - Quantum tensors basics Tensors Network one line : As like connecting multiple vectors ,multple tensors form a network ,solving such a network reduce the complexity of processing qubits YOUTUBE - Tensors Network Some ideas specifically for quantum algorithms QUANTUM MACHINE LEARNING ALGORITHMS Quantum K-Nearest Neighbour info : Here the centroid(euclidean distance) can be detected using the swap gates test between two states of the qubit , As KNN is regerssive loss can be tally using the average PDF1 from Microsoft - Theory Explanation PDF2 - A Good Material to understand the basics Matlab - Yet to come soon Python - Yet to come soon Quantum K-Means info : Two Approaches possible ,1. FFT and iFFT to make an oracle and calculate the means of superposition 2. Adiobtic Hamiltonian generation and solve the hamiltonian to determine the cluster PDF1 - Applying Quantum Kmeans on Images in a nice way PDF2 - Theory PDF3 - Explaining well the K-means clustering using hamiltonian Matlab - Yet to come soon Python - Yet to come soon Quantum Fuzzy C-Means info : As similar to kmeans fcm also using the oracle dialect ,but instead of means,here oracle optimization followed by a rotation gate is giving a good result PDF1 - Theory Matlab - Yet to come soon Python - Yet to come soon Quantum Support Vector Machine info : A little different from above as here kernel preparation is via classical and the whole training be in oracles and oracle will do the classification, As SVM is linear ,An optimal Error(Optimum of the Least Squares Dual Formulation) Based regression is needed to improve the performance PDF1 - Nice Explanation but little hard to understand :) PDF2 - Nice Application of QSVM Matlab - Yet to come soon Python - Yet to come soon Quantum Genetic Algorithm info : One of the best algorithm suited for Quantum Field ,Here the chromosomes act as qubit vectors ,the crossover part carrying by an evaluation and the mutation part carrying by the rotation of gates ![Flow Chart]() PDF1 - Very Beautiful Article , well explained and superp PDF2 - A big theory :) PDF3 - Super Comparison Matlab - Simulation Python1 - Simulation Python2 - Yet to come Quantum Hidden Morkov Models info : As HMM is already state based ,Here the quantum states acts as normal for the markov chain and the shift between states is using quantum operation based on probability distribution ![Flow Chart]() PDF1 - Nice idea and explanation PDF2 - Nice but a different concept little Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come Quantum state classification with Bayesian methods info : Quantum Bayesian Network having the same states concept using quantum states,But here the states classification to make the training data as reusable is based on the density of the states(Interference) ![Bayesian Network Sample1]() ![Bayesian Network Sample2]() ![Bayesian Network Sample3]() PDF1 - Good Theory PDF2 - Good Explanation Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come Quantum Ant Colony Optimization info : A good algorithm to process multi dimensional equations, ACO is best suited for Sales man issue , QACO is best suited for Sales man in three or more dimension, Here the quantum rotation circuit is doing the peromene update and qubits based colony communicating all around the colony in complex space ![Ant Colony Optimization 1]() PDF1 - Good Concept PDF2 - Good Application Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come Quantum Cellular Automata info : One of the very complex algorithm with various types specifically used for polynomial equations and to design the optimistic gates for a problem, Here the lattice is formed using the quatum states and time calculation is based on the change of the state between two qubits ,Best suited for nano electronics ![Quantum Cellular Automata]() Wikipedia - Basic PDF1 - Just to get the keywords PDF2 - Nice Explanation and an easily understandable application Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come QAUNTUM NEURAL NETWORK one line : Its really one of the hardest topic , To understand easily ,Normal Neural Network is doing parallel procss ,QNN is doing parallel of parallel processess ,In theory combination of various activation functions is possible in QNN ,In Normal NN more than one activation function reduce the performance and increase the complexity Quantum perceptrons info : Perceptron(layer) is the basic unit in Neural Network ,The quantum version of perceptron must satisfy both linear and non linear problems , Quantum Concepts is combination of linear(calculus of superposition) and nonlinear(State approximation using probability) ,To make a perceptron in quantum world ,Transformation(activation function) of non linearity to certain limit is needed ,which is carrying by phase estimation algorithm ![Quantum Perceptron 3]() PDF1 - Good Theory PDF2 - Good Explanation Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come QAUNTUM STATISTICAL DATA ANALYSIS one line : An under research concept ,It can be seen in multiple ways, one best way if you want to apply n derivative for a problem in current classical theory its difficult to compute as its serialization problem instead if you do parallelization of differentiation you must estimate via probability the value in all flows ,Quantum Probability Helps to achieve this ,as the loss calculation is very less . the other way comparatively booming is Quantum Bayesianism, its a solution to solve most of the uncertainity problem in statistics to combine time and space in highly advanced physical research QUANTUM PROGRAMMING LANGUAGES , TOOLs and SOFTWARES All info : All Programming languages ,softwares and tools in alphabetical order Software - Nice content of all Python library - A python library Matlab based python library - Matlab Python Library Quantum Tensor Network Github - Tensor Network Bayesforge - A Beautiful Amazon Web Service Enabled Framework for Quantum Alogorithms and Data Analytics Rigetti - A best tools repository to use quantum computer in real time Rigetti Forest - An API to connect Quantum Computer quil/pyQuil - A quantum instruction language to use forest framework Grove - Grove is a repository to showcase quantum Fourier transform, phase estimation, the quantum approximate optimization algorithm, and others developed using Forest QISKit - A IBM Kit to access quantum computer and mainly for quantum circuits IBM Bluemix Simulator - A Bluemix Simulator for Quantum Circuits Microsoft Quantum Development Kit - Microsoft Visual Studio Enbaled Kit for Quantum Circuit Creation Microsoft "Q#" - Microsoft Q Sharp a new Programming Language for Quantum Circuit Creation qiskit api python - An API to connect IBM Quantum Computer ,With the generated token its easy to connect ,but very limited utils ,Lot of new utils will come soon Cyclops Tensor Framework - A framework to do tensor network simulations Python ToolKit for chemistry and physics Quantum Algorithm simulations - A New Started Project for simulating molecule and solids Bayesian Based Quatum Projects Repository - A nice repository and the kickstarter of bayesforge Google Fermion Products - A newly launched product specifivally for chemistry simulation Tree Tensor Networks - Interesting Tensor Network in Incubator Deep Tensor Neural Network - Some useful information about Tensor Neural Network in Incubator Generative Tensorial Networks - A startup to apply machine learning via tensor network for drug discovery Google Bristlecone - A new Quantum Processor from Google , Aimed for Future Hardwares with full fledged AI support XANADU - A Light based Quantum Hardware(chips supports) and Software Company Started in Preparation Stage. Soon will be in market fathom computing - A new concept to train the ai in a processor using light and quantum based concepts. soon products will be launch Alibaba Quantum Computing Cloud Service - Cloud Service to access 11 Bit Quantum Computing Processor Atomistic Machine Learning Project - Seems something Interesting with Deep Tensor Network for Quantum Chemistry Applications circQ and Google Works - Google Top Efforts on Tools IBM Safe Cryptography on Cloud - IBM Started and Developing a Quantm Safe Cryptography to replace all our Certificate Authority via Cloud Google Tensor Network Open Source - Google Started the Most Scientist Preferred Way To Use a Quantum Computer Circuit. Tensor Flow Which Makes Easy to Design the Network and Will Leave the Work Effect Of Gates, Processor Preparation and also going to tell the beauty of Maths Google Tensor Network Github - Github Project of Google Tensor Network Quantum Tensorflow - Yet to come soon Quantum Spark - Yet to come soon Quatum Map Reduce - Yet to come soon Quantum Database - Yet to come soon Quantum Server - Yet to come soon Quantum Data Analytics - Yet to come soon QUANTUM HOT TOPICS Deep Quantum Learning why and what is deep learning? In one line , If you know deep learning you can get a good job :) ,Even a different platform undergraduated and graduated person done a master specialization in deep learning can work in this big sector :), Practically speaking machine learning (vector mathematics) , deep learning (vector space(Graphics) mathematics) and big data are the terms created by big companies to make a trend in the market ,but in science and research there is no word such that , Now a days if you ask a junior person working in this big companies ,what is deep learning ,you will get some reply as "doing linear regression with stochastic gradient for a unsupervised data using Convolutional Neural Network :)" ,They knows the words clearly and knows how to do programming using that on a bunch of "relative data" , If you ask them about the FCM , SVM and HMM etc algorithms ,they will simply say these are olden days algorithms , deep learning replaced all :), But actually they dont know from the birth to the till level and the effectiveness of algorithms and mathematics ,How many mathematical theorems in vector, spaces , tensors etc solved to find this "hiding the complexity technology", They did not played with real non relative data like medical images, astro images , geology images etc , finding a relation and features is really complex and looping over n number of images to do pattern matching is a giant work , Now a days the items mentioned as deep learning (= multiple hidden artifical neural network) is not suitable for that why quantum deep learning or deep quantum learning? In the mid of Artificial Neural Network Research people realised at the maximum extreme only certain mathematical operations possible to do with ANN and the aim of this ANN is to achieve parallel execution of many mathematical operations , In artificial Intelligence ,the world intelligence stands for mathematics ,how effective if a probem can be solvable is based on the mathematics logic applying on the problem , more the logic will give more performance(more intelligent), This goal open the gate for quantum artificial neural network, On applying the ideas behind the deep learning to quantum mechanics environment, its possible to apply complex mathematical equations to n number of non relational data to find more features and can improve the performance Quantum Machine Learning vs Deep Learning Its fun to discuss about this , In recent days most of the employees from Product Based Companies Like google,microsoft etc using the word deep learning ,What actually Deep Learning ? and is it a new inventions ? how to learn this ? Is it replacing machine learning ? these question come to the mind of junior research scholars and mid level employees The one answer to all questions is deep learning = parallel "for" loops ,No more than that ,Its an effective way of executing multiple tasks repeatly and to reduce the computation cost, But it introduce a big cap between mathematics and computerscience , How ? All classical algorithms based on serial processing ,Its depends on the feedback of the first loop ,On applying a serial classical algorithm in multiple clusters wont give a good result ,but some light weight parallel classical algorithms(Deep learning) doing the job in multiple clusters and its not suitable for complex problems, What is the solution for then? As in the title Quantum Machine Learning ,The advantage behind is deep learning is doing the batch processing simply on the data ,but quantum machine learning designed to do batch processing as per the algorithm The product companies realised this one and they started migrating to quantum machine learning and executing the classical algorithms on quantum concept gives better result than deep learning algorithms on classical computer and the target to merge both to give very wonderful result References Quora - Good Discussion Quora - The Bridge Discussion Pdf - Nice Discussion Google - Google Research Discussion Microsoft - Microsoft plan to merge both IBM - IBM plan to merge both IBM Project - IBM Project idea MIT and Google - Solutions for all questions QUANTUM MEETUPS Meetup 1 - Quantum Physics Meetup 2 - Quantum Computing London Meetup 3 - Quantum Computing New York Meetup 4 - Quantum Computing Canada Meetup 5 - Quantum Artificial Intelligence Texas Meetup 6 - Genarl Quantum Mechanics , Mathematics New York Meetup 7 - Quantum Computing Mountain View California Meetup 8 - Statistical Analysis New York Meetup 9 - Quantum Mechanics London UK Meetup 10 - Quantum Physics Sydney Australia Meetup 11 - Quantum Physics Berkeley CA Meetup 12 - Quantum Computing London UK Meetup 13 - Quantum Mechanics Carmichael CA Meetup 14 - Maths and Science Group Portland Meetup 15 - Quantum Physics Santa Monica, CA Meetup 16 - Quantum Mechanics London Meetup 17 - Quantum Computing London Meetup 18 - Quantum Meta Physics ,Kansas City , Missouri ,US Meetup 19 - Quantum Mechanics and Physics ,Boston ,Massachusetts ,US Meetup 20 - Quantum Physics and Mechanics ,San Francisco ,California Meetup 21 - Quantum Mechanics ,Langhorne, Pennsylvania Meetup 22 - Quantum Mechanics ,Portland QUANTUM BASED DEGREES Plenty of courses around the world and many Universities Launching it day by day ,Instead of covering only Quantum ML , Covering all Quantum Related topics gives more idea in the order below Available Courses Quantum Mechanics for Science and Engineers Online Standford university - Nice Preparatory Course edx - Quantum Mechanics for Everyone NPTEL 1 - Nice Series of Courses to understand basics and backbone of quantum mechanics NPTEL 2 NPTEL 3 NPTEL 4 NPTEL 5 Class Based Course UK Bristol Australia Australian National University Europe Maxs Planks University Quantum Physics Online MIT - Super Explanation and well basics NPTEL - Nice Series of Courses to understand basics and backbone of quantum Physics Class Based Course Europe University of Copenhagen Quantum Chemistry Online NPTEL 1 - Nice Series of Courses to understand basics and backbone of quantum Chemistry NPTEL 2 - Class Based Course Europe UGent Belgium Quantum Computing Online MIT - Super Explanation and well basics edx - Nice Explanation NPTEL - Nice Series of Courses to understand basics and backbone of quantum Computing Class Based Course Canada uwaterloo Singapore National University Singapore USA Berkley China Baidu Quantum Technology Class Based Course Canada uwaterloo Singapore National University Singapore Europe Munich Russia Skoltech Quantum Information Science External Links quantwiki Online MIT - Super Explanation and well basics edx - Nice Explanation NPTEL - Nice Series of Courses to understand basics and backbone of quantum information and computing Class Based Course USA MIT Standford University Joint Center for Quantum Information and Computer Science - University of Maryland Canada Perimeter Institute Singapore National University Singapore Europe ULB Belgium IQOQI Quantum Electronics Online MIT - Wonderful Course NPTEL - Nice Series of Courses to understand basics and backbone of quantum Electronics Class Based Course USA Texas Europe Zurich ICFO Asia Tata Institute Quantum Field Theory Online Standford university - Nice Preparatory Course edx - Some QFT Concepts available Class Based Course UK Imperial Europe Vrije Quantum Computer Science Class Based Course USA Oxford Joint Center for Quantum Information and Computer Science - University of Maryland Quantum Artificial Intelligence and Machine Learning External Links Quora 1 Quora 1 Artificial Agents Research for Quantum Designs Quantum Mathematics Class Based Course USA University of Notre CONSOLIDATED Quantum Research Papers scirate - Plenty of Quantum Research Papers Available Peter Wittek - Famous Researcher for the Quantum Machine Leanrning , Published a book in this topic [Murphy Yuezhen Niu] (https://scholar.google.com/citations?user=0wJPxfkAAAAJ&hl=en) - A good researcher published some nice articles Recent Quantum Updates forum ,pages and newsletter Quantum-Tech - A Beautiful Newsletter Page Publishing Amazing Links facebook Quantum Machine Learning - Running By me . Not that much good :). You can get some ideas Linkedlin Quantum Machine Learning - A nice page running by experts. Can get plenty of ideas FOSDEM 2019 Quantum Talks - A one day talk in fosdem 2019 with more than 10 research topics,tools and ideas FOSDEM 2020 Quantum Talks - Live talk in fosdem 2020 with plenty new research topics,tools and ideas License Dedicated Opensources ![Dedicated Opensources]() Source code of plenty of Algortihms in Image Processing , Data Mining ,etc in Matlab, Python ,Java and VC++ Scripts Good Explanations of Plenty of algorithms with flow chart etc Comparison Matrix of plenty of algorithms Is Quantum Machine Learning Will Reveal the Secret Maths behind Astrology? Awesome Machine Learning and Deep Learning Mathematics is online Published Basic Presentation of the series Quantum Machine Learning Contribution If you think this page might helpful. Please help for World Education Charity or kids who wants to learn

aioquic
github
LLM Vibe Score0.518
Human Vibe Score0.04117299426077279
aiortcMar 27, 2025

aioquic

aioquic ======= .. image:: https://img.shields.io/pypi/l/aioquic.svg :target: https://pypi.python.org/pypi/aioquic :alt: License .. image:: https://img.shields.io/pypi/v/aioquic.svg :target: https://pypi.python.org/pypi/aioquic :alt: Version .. image:: https://img.shields.io/pypi/pyversions/aioquic.svg :target: https://pypi.python.org/pypi/aioquic :alt: Python versions .. image:: https://github.com/aiortc/aioquic/workflows/tests/badge.svg :target: https://github.com/aiortc/aioquic/actions :alt: Tests .. image:: https://img.shields.io/codecov/c/github/aiortc/aioquic.svg :target: https://codecov.io/gh/aiortc/aioquic :alt: Coverage .. image:: https://readthedocs.org/projects/aioquic/badge/?version=latest :target: https://aioquic.readthedocs.io/ :alt: Documentation What is `aioquic? aioquic is a library for the QUIC network protocol in Python. It features a minimal TLS 1.3 implementation, a QUIC stack and an HTTP/3 stack. aioquic is used by Python opensource projects such as dnspython_, hypercorn, mitmproxy and the Web Platform Tests_ cross-browser test suite. It has also been used extensively in research papers about QUIC. To learn more about aioquic please read the documentation_. Why should I use aioquic? aioquic has been designed to be embedded into Python client and server libraries wishing to support QUIC and / or HTTP/3. The goal is to provide a common codebase for Python libraries in the hope of avoiding duplicated effort. Both the QUIC and the HTTP/3 APIs follow the "bring your own I/O" pattern, leaving actual I/O operations to the API user. This approach has a number of advantages including making the code testable and allowing integration with different concurrency models. A lot of effort has gone into writing an extensive test suite for the aioquic code to ensure best-in-class code quality, and it is regularly tested for interoperability against other QUIC implementations. Features minimal TLS 1.3 implementation conforming with RFC 8446_ QUIC stack conforming with RFC 9000 (QUIC v1) and RFC 9369 (QUIC v2) IPv4 and IPv6 support connection migration and NAT rebinding logging TLS traffic secrets logging QUIC events in QLOG format version negotiation conforming with RFC 9368_ HTTP/3 stack conforming with RFC 9114_ server push support WebSocket bootstrapping conforming with RFC 9220_ datagram support conforming with RFC 9297_ Installing The easiest way to install aioquic is to run: .. code:: bash pip install aioquic Building from source If there are no wheels for your system or if you wish to build aioquic from source you will need the OpenSSL development headers. Linux ..... On Debian/Ubuntu run: .. code-block:: console sudo apt install libssl-dev python3-dev On Alpine Linux run: .. code-block:: console sudo apk add openssl-dev python3-dev bsd-compat-headers libffi-dev OS X .... On OS X run: .. code-block:: console brew install openssl You will need to set some environment variables to link against OpenSSL: .. code-block:: console export CFLAGS=-I$(brew --prefix openssl)/include export LDFLAGS=-L$(brew --prefix openssl)/lib Windows ....... On Windows the easiest way to install OpenSSL is to use Chocolatey_. .. code-block:: console choco install openssl You will need to set some environment variables to link against OpenSSL: .. code-block:: console $Env:INCLUDE = "C:\Progra~1\OpenSSL\include" $Env:LIB = "C:\Progra~1\OpenSSL\lib" Running the examples aioquic comes with a number of examples illustrating various QUIC usecases. You can browse these examples here: https://github.com/aiortc/aioquic/tree/main/examples License aioquic is released under the BSD license`_. .. _read the documentation: https://aioquic.readthedocs.io/en/latest/ .. _dnspython: https://github.com/rthalley/dnspython .. _hypercorn: https://github.com/pgjones/hypercorn .. _mitmproxy: https://github.com/mitmproxy/mitmproxy .. _Web Platform Tests: https://github.com/web-platform-tests/wpt .. _tested for interoperability: https://interop.seemann.io/ .. _QUIC implementations: https://github.com/quicwg/base-drafts/wiki/Implementations .. _cryptography: https://cryptography.io/ .. _Chocolatey: https://chocolatey.org/ .. _BSD license: https://aioquic.readthedocs.io/en/latest/license.html .. _RFC 8446: https://datatracker.ietf.org/doc/html/rfc8446 .. _RFC 9000: https://datatracker.ietf.org/doc/html/rfc9000 .. _RFC 9114: https://datatracker.ietf.org/doc/html/rfc9114 .. _RFC 9220: https://datatracker.ietf.org/doc/html/rfc9220 .. _RFC 9297: https://datatracker.ietf.org/doc/html/rfc9297 .. _RFC 9368: https://datatracker.ietf.org/doc/html/rfc9368 .. _RFC 9369: https://datatracker.ietf.org/doc/html/rfc9369

obsei
github
LLM Vibe Score0.545
Human Vibe Score0.10175553624190911
obseiMar 27, 2025

obsei

Note: Obsei is still in alpha stage hence carefully use it in Production. Also, as it is constantly undergoing development hence master branch may contain many breaking changes. Please use released version. Obsei (pronounced "Ob see" | /əb-'sē/) is an open-source, low-code, AI powered automation tool. Obsei consists of - Observer: Collect unstructured data from various sources like tweets from Twitter, Subreddit comments on Reddit, page post's comments from Facebook, App Stores reviews, Google reviews, Amazon reviews, News, Website, etc. Analyzer: Analyze unstructured data collected with various AI tasks like classification, sentiment analysis, translation, PII, etc. Informer: Send analyzed data to various destinations like ticketing platforms, data storage, dataframe, etc so that the user can take further actions and perform analysis on the data. All the Observers can store their state in databases (Sqlite, Postgres, MySQL, etc.), making Obsei suitable for scheduled jobs or serverless applications. !Obsei diagram Future direction - Text, Image, Audio, Documents and Video oriented workflows Collect data from every possible private and public channels Add every possible workflow to an AI downstream application to automate manual cognitive workflows Use cases Obsei use cases are following, but not limited to - Social listening: Listening about social media posts, comments, customer feedback, etc. Alerting/Notification: To get auto-alerts for events such as customer complaints, qualified sales leads, etc. Automatic customer issue creation based on customer complaints on Social Media, Email, etc. Automatic assignment of proper tags to tickets based content of customer complaint for example login issue, sign up issue, delivery issue, etc. Extraction of deeper insight from feedbacks on various platforms Market research Creation of dataset for various AI tasks Many more based on creativity 💡 Installation Prerequisite Install the following (if not present already) - Install Python 3.7+ Install PIP Install Obsei You can install Obsei either via PIP or Conda based on your preference. To install latest released version - Install from master branch (if you want to try the latest features) - Note: all option will install all the dependencies which might not be needed for your workflow, alternatively following options are available to install minimal dependencies as per need - pip install obsei[source]: To install dependencies related to all observers pip install obsei[sink]: To install dependencies related to all informers pip install obsei[analyzer]: To install dependencies related to all analyzers, it will install pytorch as well pip install obsei[twitter-api]: To install dependencies related to Twitter observer pip install obsei[google-play-scraper]: To install dependencies related to Play Store review scrapper observer pip install obsei[google-play-api]: To install dependencies related to Google official play store review API based observer pip install obsei[app-store-scraper]: To install dependencies related to Apple App Store review scrapper observer pip install obsei[reddit-scraper]: To install dependencies related to Reddit post and comment scrapper observer pip install obsei[reddit-api]: To install dependencies related to Reddit official api based observer pip install obsei[pandas]: To install dependencies related to TSV/CSV/Pandas based observer and informer pip install obsei[google-news-scraper]: To install dependencies related to Google news scrapper observer pip install obsei[facebook-api]: To install dependencies related to Facebook official page post and comments api based observer pip install obsei[atlassian-api]: To install dependencies related to Jira official api based informer pip install obsei[elasticsearch]: To install dependencies related to elasticsearch informer pip install obsei[slack-api]:To install dependencies related to Slack official api based informer You can also mix multiple dependencies together in single installation command. For example to install dependencies Twitter observer, all analyzer, and Slack informer use following command - How to use Expand the following steps and create a workflow - Step 1: Configure Source/Observer Twitter Youtube Scrapper Facebook Email Google Maps Reviews Scrapper AppStore Reviews Scrapper Play Store Reviews Scrapper Reddit Reddit Scrapper Note: Reddit heavily rate limit scrappers, hence use it to fetch small data during long period Google News Web Crawler Pandas DataFrame Step 2: Configure Analyzer Note: To run transformers in an offline mode, check transformers offline mode. Some analyzer support GPU and to utilize pass device parameter. List of possible values of device parameter (default value auto): auto: GPU (cuda:0) will be used if available otherwise CPU will be used cpu: CPU will be used cuda:{id} - GPU will be used with provided CUDA device id Text Classification Text classification: Classify text into user provided categories. Sentiment Analyzer Sentiment Analyzer: Detect the sentiment of the text. Text classification can also perform sentiment analysis but if you don't want to use heavy-duty NLP model then use less resource hungry dictionary based Vader Sentiment detector. NER Analyzer NER (Named-Entity Recognition) Analyzer: Extract information and classify named entities mentioned in text into pre-defined categories such as person names, organizations, locations, medical codes, time expressions, quantities, monetary values, percentages, etc Translator PII Anonymizer Dummy Analyzer Dummy Analyzer: Does nothing. Its simply used for transforming the input (TextPayload) to output (TextPayload) and adding the user supplied dummy data. Step 3: Configure Sink/Informer Slack Zendesk Jira ElasticSearch Http Pandas DataFrame Logger This is useful for testing and dry running the pipeline. Step 4: Join and create workflow source will fetch data from the selected source, then feed it to the analyzer for processing, whose output we feed into a sink to get notified at that sink. Step 5: Execute workflow Copy the code snippets from Steps 1 to 4 into a python file, for example example.py and execute the following command - Demo We have a minimal streamlit based UI that you can use to test Obsei. !Screenshot Watch UI demo video Check demo at (Note: Sometimes the Streamlit demo might not work due to rate limiting, use the docker image (locally) in such cases.) To test locally, just run To run Obsei workflow easily using GitHub Actions (no sign ups and cloud hosting required), refer to this repo. Companies/Projects using Obsei Here are some companies/projects (alphabetical order) using Obsei. To add your company/project to the list, please raise a PR or contact us via email. Oraika: Contextually understand customer feedback 1Page: Giving a better context in meetings and calls Spacepulse: The operating system for spaces Superblog: A blazing fast alternative to WordPress and Medium Zolve: Creating a financial world beyond borders Utilize: No-code app builder for businesses with a deskless workforce Articles Sr. No. Title Author 1 AI based Comparative Customer Feedback Analysis Using Obsei Reena Bapna 2 LinkedIn App - User Feedback Analysis Himanshu Sharma Tutorials Sr. No. Workflow Colab Binder 1 Observe app reviews from Google play store, Analyze them by performing text classification and then Inform them on console via logger PlayStore Reviews → Classification → Logger 2 Observe app reviews from Google play store, PreProcess text via various text cleaning functions, Analyze them by performing text classification, Inform them to Pandas DataFrame and store resultant CSV to Google Drive PlayStore Reviews → PreProcessing → Classification → Pandas DataFrame → CSV in Google Drive 3 Observe app reviews from Apple app store, PreProcess text via various text cleaning function, Analyze them by performing text classification, Inform them to Pandas DataFrame and store resultant CSV to Google Drive AppStore Reviews → PreProcessing → Classification → Pandas DataFrame → CSV in Google Drive 4 Observe news article from Google news, PreProcess text via various text cleaning function, Analyze them via performing text classification while splitting text in small chunks and later computing final inference using given formula Google News → Text Cleaner → Text Splitter → Classification → Inference Aggregator 💡Tips: Handle large text classification via Obsei Documentation For detailed installation instructions, usages and examples, refer to our documentation. Support and Release Matrix Linux Mac Windows Remark Tests ✅ ✅ ✅ Low Coverage as difficult to test 3rd party libs PIP ✅ ✅ ✅ Fully Supported Conda ❌ ❌ ❌ Not Supported Discussion forum Discussion about Obsei can be done at community forum Changelogs Refer releases for changelogs Security Issue For any security issue please contact us via email Stargazers over time Maintainers This project is being maintained by Oraika Technologies. Lalit Pagaria and Girish Patel are maintainers of this project. License Copyright holder: Oraika Technologies Overall Apache 2.0 and you can read License file. Multiple other secondary permissive or weak copyleft licenses (LGPL, MIT, BSD etc.) for third-party components refer Attribution. To make project more commercial friendly, we void third party components which have strong copyleft licenses (GPL, AGPL etc.) into the project. Attribution This could not have been possible without these open source softwares. Contribution First off, thank you for even considering contributing to this package, every contribution big or small is greatly appreciated. Please refer our Contribution Guideline and Code of Conduct. Thanks so much to all our contributors

He makes $750 a day 'Vibe Coding' Apps (using Replit, ChatGPT, Upwork)
youtube
LLM Vibe Score0.379
Human Vibe Score0.77
Greg IsenbergMar 21, 2025

He makes $750 a day 'Vibe Coding' Apps (using Replit, ChatGPT, Upwork)

Billy Howell shares his strategy for making money by building and selling custom web applications using AI tools like Replit. He demonstrates the process by finding projects on Upwork, creating a product requirements document with ChatGPT, and using Replit to automatically generate a functional web application. Billy explains that this approach is less risky than building SaaS products because it validates demand before significant development work. Timestamps: 00:00 - Intro 02:19 - Searching for App Ideas on Upwork 11:04 - Using ChatGPT for PRD Creation 12:22 - Why choose Replit for Development 15:15 - Building Prototype with Replit 19:53 - Areas of Concern when building with AI coders 23:30 - Earning Potential on Upwork 27:55 - The process for selling these Apps 32:03 - Comparing Different Business Models 35:40 - Huge opportunity: Unbundling SaaS 37:44 - Testing App 39:39 - How to standout on Upwork 40:35 - Integrating v0 UI to Replit Key Points • Billy Howell explains his method of "vibe coding" - using AI tools like Replit to quickly build and sell custom web applications • The process involves finding clients on Upwork who need solutions, creating a prototype, and selling it before building the complete app • Billy demonstrates how to use Repl.it with AI assistance to rapidly build a case management system for a nonprofit • The approach focuses on creating simple CRUD (Create, Read, Update, Delete) applications rather than complex systems 1) The "Sell First, Build Later" Framework Billy's #1 rule: Find someone to BUY your app BEFORE you build it. Most developers get this backward - they build something cool then struggle to find users. The secret? Don't market. SELL. How? Look for people ALREADY trying to pay for solutions 2) Upwork Gold Mining Strategy Billy's exact process: • Search Upwork for jobs mentioning expensive SaaS tools (Airtable, HubSpot, etc) • Look for simple CRUD apps (data entry, visualization) • Build a quick prototype in Repl.it • Send a Loom video demo to potential clients His first sale? $750 replacing an Airtable solution! 3) The Vibe Coding Tech Stack Billy's weapons of choice: • Replit for rapid prototyping (zero setup friction!) • ChatGPT to format requirements into PRDs • V0 for beautiful UI mockups • ShadCN components for clean interfaces The magic combo: Feed requirements to Replit + "build me this app" = working prototype in MINUTES. 4) What to Avoid When Vibe Coding Not all projects are created equal! Watch out for: • Payment processing (risky) • DocuSign integrations (complex) • Calendar functionality (AI struggles with time zones) • Anything changing data in other apps Start with simple CRUD apps that store and display information. 5) The Real Money-Making Model Billy's approach isn't just about one-off projects: • Initial build: $750-2,500 • Charge for hosting • Recurring revenue from feature requests • Get referrals to similar businesses One recent client is now reselling his solution to other companies in the same industry! 6) Why This Beats Building a SaaS Building a traditional SaaS = "nightmare money pit" according to Billy. With vibe coding consulting: • De-risk by getting paid upfront • Learn across multiple projects • No marketing costs • Discover validated problems • Build a portfolio of solutions Six figures on Upwork is VERY doable. 7) The 60-Second Sales Pitch Billy's exact closing technique: • Find job posting • Make mockup in V0 or Replit • Record 1-minute Loom: "I'm Billy, I make apps. I know you wanted Airtable, but I made this custom for you." • Personalize with company name • Send and repeat Simple. Effective. PROFITABLE. The future of coding isn't about knowing every framework—it's about SOLVING PROBLEMS quickly. Anyone can do this with the right tools and approach. Notable Quotes: "The number one thing is how to sell an app that you've built... And the secret is not to market. It's just to sell it." - Billy Howell "We start, we need to find someone to buy the app before we build it. That's where most people get this wrong, is they build something and then try to sell it or try to get users." - Billy Howell LCA helps Fortune 500s and fast-growing startups build their future - from Warner Music to Fortnite to Dropbox. We turn 'what if' into reality with AI, apps, and next-gen products https://latecheckout.agency/ BoringAds — ads agency that will build you profitable ad campaigns http://boringads.com/ BoringMarketing — SEO agency and tools to get your organic customers http://boringmarketing.com/ Startup Empire — a membership for builders who want to build cash-flowing businesses https://www.startupempire.co FIND ME ON SOCIAL X/Twitter: https://twitter.com/gregisenberg Instagram: https://instagram.com/gregisenberg/ LinkedIn: https://www.linkedin.com/in/gisenberg/ FIND BILLY ON SOCIAL X/Twitter: https://x.com/billyjhowell Youtube: https://www.youtube.com/@billyjhowell

The only video you need to Master N8N + AI agents (For complete beginners)
youtube
LLM Vibe Score0.396
Human Vibe Score0.64
Simon Scrapes | AI Agents & AutomationFeb 21, 2025

The only video you need to Master N8N + AI agents (For complete beginners)

Serious about Implementing AI? Shortcut your Path HERE, and connect with +300 entrepreneurs on the same mission: https://www.skool.com/scrapes This is a comprehensive 4hr course with all the secrets I've learned from 8 months of building out N8N workflows for my clients (over 100+ workflows!). During this course we'll cover everything you need to shortcut your journey into building automations with N8N, AI Agents & workflow automation! 🛠️ Links (affiliate) • n8n: https://n8n.partnerlinks.io/scrapesai 📧 Curated roundups of real-world AI implementations 📧 https://scrapes-ai.kit.com/b6b1a73dfd Want more? https://www.youtube.com/@simonscrapes?sub_confirmation=1 🚧 Looking for custom built AI agents for your business? 🚧 https://automake.io 💬 Share in the comments what you learnt during the video! 0:00:00 - Course Overview 0:04:12 - SECTION 1 - Getting started 0:09:57 - 1.1. Setting up N8N 0:15:10 - 1.2. Building blocks of N8N 0:16:52 - 1.3. The N8N Canvas 0:19:02 - 1.4. Triggers & Actions 0:24:55 - 1.5. Connect nodes 0:30:09 - 1.6. Visualising Data 0:32:13 - 1.7. JSON vs Table vs Schema 0:35:12 - 1.8. Mastering Static Data 0:38:10 - 1.9. Dynamic Data 0:43:21 - 1.10. Referencing Nodes (Foolproof) 0:47:05 - 1.11. Pinning Data 0:49:26 - 1.12. Simple Retry Logic 0:52:15 - 1.13. Node Naming 0:57:38 - SECTION 2 - Building Your First Automation with Data From Your Business 0:58:45 - 2.1. Planning Your Workflow 1:02:05 - 2.2. Monitoring Your Gmail 1:04:15 - 2.3. Setting up Google Credentials 1:09:01 - 2.4. Manipulating Data with Set 1:13:11 - 2.5. Data Format Comparison (HTML, Markdown) 1:15:55 - 2.6. Your First Automation 1:20:46 - 2.7. Building an Invoice Parsing System & Tackling File Formats 1:30:42 - 2.8. Cleaning Data with Code Node 1:39:19 - 2.9. Conditionals (IF) 1:44:24 - 2.10. Multiple Inputs 1:46:04 - 2.11. Merging Data 1:50:03 - 2.12. Memory Management 1:51:15 - 2.13. Large Data Sets (Loops) 1:54:52 - 2.14. Rounding Up Our Automation 1:55:16 - SECTION 3 - Agentic Workflows & AI Agents 1:56:07 - 3.1. Agentic vs Non-Agentic Workflows 1:59:28 - 3.2. Agentic Examples You Might Use 2:05:16 - 3.3. N8N AI Nodes 2:12:55 - 3.4. AI Agents - So What Are They? 2:20:42 - 3.5. AI Agents - What Business Use Do They Have? 2:25:05 - 3.6. Setting Up AI in Our Workflow 2:27:58 - 3.7. Prompting for Beginners 2:33:29 - 3.8. Openrouter for AI Models 2:39:10 - 3.9. Getting Consistent Outputs 2:45:53 - 3.10. Rounding Up Your Invoice Parsing Workflow 2:46:49 - 3.11. Mapping Back to Your Database 2:54:00 - SECTION 4 - Data From Outside Your Business 2:59:10 - 4.1. Connecting to an API with N8N 3:01:29 - 4.2. Reading API Docs Made Easy 3:04:24 - 4.3. API Authorisation 3:06:50 - 4.4. POST Request - PDFco 3:12:47 - 4.5. Uploading Our Files via API 3:22:18 - 4.6. Completing Our API Uploads 3:25:37 - 4.7. Connect to ANY API in 2 mins 3:29:30 - 4.8. Push Data Back to Our Table 3:35:03 - SECTION 5 - Making Your Life Easy & Scalable 3:37:27 - 5.1. Naming Workflows & Tagging 3:38:43 - 5.2. Workflow Separation 3:41:11 - 5.3. Modular Design 3:48:12 - 5.4. Error Handling 3:52:31 - 5.5. Debugging (easy Mode!) 3:53:31 - 5.6. Community Nodes 3:56:31 - 5.7. N8N Template Library 3:59:14 - 5.8. Getting Help #N8N #n8ntutorial #N8NBeginner

I ranked every AI Coder: Bolt vs. Cursor vs. Replit vs Lovable
youtube
LLM Vibe Score0.399
Human Vibe Score0.77
Greg IsenbergJan 24, 2025

I ranked every AI Coder: Bolt vs. Cursor vs. Replit vs Lovable

v0 vs windsurf vs replit vs bolt vs lovable vs tempolabs - which one should you use? Ras Mic breaks down the AI coding platforms based on how tech-savvy you are and how much control you want. He splits the tools into three groups: no-code options for non-techies, hybrid platforms for those with a mix of skills, and advanced tools for developers. None of them are quite ready for full-on production yet, but the video highlights what each one does best—whether it’s integrations, teamwork, or deployment features. Timestamps: 00:00 - Intro 01:00 - Overview of Popular Tools 02:29 - Technical vs. non-technical user classification 05:37 - Production readiness discussion 09:50 - Mapping Tools to User Profiles 12:52 - Platform comparisons and strengths 15:15 - Pricing discussion 16:43 - AI agents in coding platforms 19:04 - Final Recommendations and User Alignment Key Points: • Comprehensive comparison of major AI coding platforms (Lovable, Bolt, V0, Replit, Tempo Labs, Onlook, Cursor, Windsurf) • Tools categorized by technical expertise required and level of control offered • None of the platforms are 100% production-ready, but Replit and Tempo Labs are closest • All platforms offer similar base pricing ($20-30/month) with generous free tiers 1) First, understand the 3 MAJOR CATEGORIES of AI coding tools: • No-code (non-technical friendly) • Middle-ground (hybrid) • Technical (developer-focused) Your choice depends on TWO key factors: How much control you want Your technical expertise 2) THE CONTROL SPECTRUM Less Control → More Control • Lovable (basic control) • Bolt/V0 (code tweaking) • Replit (file management) • Tempo/Onlook (design control) • Cursor/Windsurf (full code control) 3) PRODUCTION READINESS STATUS Most honest take: None are 100% there yet, but some are close: Top contenders: • Replit • Tempo Labs Runner-ups: • Bolt • Lovable Pro tip: Start building now to be ready when they mature! 4) BEST TOOLS BY USER TYPE Non-technical: • Lovable • Bolt Product-minded non-technical: • Tempo Labs • Replit Technical folks: • Cursor • Windsurf 5) WINNING FEATURES BY PLATFORM Integrations: Lovable (crushing it!) Replit Tempo Labs Collaboration: Tempo Labs Replit Deployment: All solid, but Tempo needs work 6) PRICING INSIDER TIP All platforms hover around $20-30/month for basic tiers SECRET: They ALL have generous free tiers! Pro tip: Test drive everything before committing to paid plans 7) FINAL ADVICE Build a simple todo app on each platform Use free tiers to test Choose based on: Your technical comfort Desired level of control Specific project needs Remember: There's no "perfect" tool - just the right one for YOU! Notable Quotes: "None of the tools are there yet. I cannot confidently say you can build something to production easily, simply without a ton of roadblocks." - Ras Mic "Control is not for everybody. Did you like the assumptions that AI product was making for you? Or do you want to be able to tell it exactly what to do?" - Ras Mic LCA helps Fortune 500s and fast-growing startups build their future - from Warner Music to Fortnite to Dropbox. We turn 'what if' into reality with AI, apps, and next-gen products https://latecheckout.agency/ BoringAds — ads agency that will build you profitable ad campaigns http://boringads.com/ BoringMarketing — SEO agency and tools to get your organic customers http://boringmarketing.com/ Startup Empire - a membership for builders who want to build cash-flowing businesses https://www.startupempire.co FIND ME ON SOCIAL X/Twitter: https://twitter.com/gregisenberg Instagram: https://instagram.com/gregisenberg/ LinkedIn: https://www.linkedin.com/in/gisenberg/ FIND MIC ON SOCIAL X/Twitter: https://x.com/rasmickyy Youtube: https://www.youtube.com/@rasmic

airflow-tutorial
github
LLM Vibe Score0.508
Human Vibe Score0.13240553426231688
hgrifJan 19, 2025

airflow-tutorial

Airflow tutorial This tutorial is loosely based on the Airflow tutorial in the official documentation. It will walk you through the basics of setting up Airflow and creating an Airflow workflow. This tutorial was published on the blog of GoDataDriven. Setup You can skip this section if Airflow is already set up. Make sure that you can run airflow commands, know where to put your DAGs and have access to the web UI. Install Airflow Airflow is installable with pip via a simple pip install apache-airflow. Either use a separate python virtual environment or install it in your default python environment. To use the conda virtual environment as defined in environment.yml in this git-repo: Install miniconda. Make sure that conda is on your path: Create the virtual environment from environment.yml: Activate the virtual environment: You should now have an (almost) working Airflow installation. Alternatively, install Airflow yourself by running: Airflow used to be packaged as airflow but is packaged as apache-airflow since version 1.8.1. Make sure that you install any extra packages with the right Python package: e.g. use pip install apache-airflow[dask] if you've installed apache-airflow and do not use pip install airflow[dask]. Leaving out the prefix apache- will install an old version of Airflow next to your current version, leading to a world of hurt. You may run into problems if you don't have the right binaries or Python packages installed for certain backends or operators. When specifying support for e.g. PostgreSQL when installing extra Airflow packages, make sure the database is installed; do a brew install postgresql or apt-get install postgresql before the pip install apache-airflow[postgres]. Similarly, when running into HiveOperator errors, do a pip install apache-airflow[hive] and make sure you can use Hive. Run Airflow Before you can use Airflow you have to initialize its database. The database contains information about historical & running workflows, connections to external data sources, user management, etc. Once the database is set up, Airflow's UI can be accessed by running a web server and workflows can be started. The default database is a SQLite database, which is fine for this tutorial. In a production setting you'll probably be using something like MySQL or PostgreSQL. You'll probably want to back it up as this database stores the state of everything related to Airflow. Airflow will use the directory set in the environment variable AIRFLOW_HOME to store its configuration and our SQlite database. This directory will be used after your first Airflow command. If you don't set the environment variable AIRFLOW_HOME, Airflow will create the directory ~/airflow/ to put its files in. Set environment variable AIRFLOW_HOME to e.g. your current directory $(pwd): or any other suitable directory. Next, initialize the database: Now start the web server and go to localhost:8080 to check out the UI: It should look something like this: With the web server running workflows can be started from a new terminal window. Open a new terminal, activate the virtual environment and set the environment variable AIRFLOW_HOME for this terminal as well: Make sure that you're an in the same directory as before when using $(pwd). Run a supplied example: And check in the web UI that it has run by going to Browse -> Task Instances. This concludes all the setting up that you need for this tutorial. Tips Both Python 2 and 3 are be supported by Airflow. However, some of the lesser used parts (e.g. operators in contrib) might not support Python 3. For more information on configuration check the sections on Configuration and Security of the Airflow documentation. Check the Airflow repository for upstart and systemd templates. Airflow logs extensively, so pick your log folder carefully. Set the timezone of your production machine to UTC: Airflow assumes it's UTC. Workflows We'll create a workflow by specifying actions as a Directed Acyclic Graph (DAG) in Python. The tasks of a workflow make up a Graph; the graph is Directed because the tasks are ordered; and we don't want to get stuck in an eternal loop so the graph also has to be Acyclic. The figure below shows an example of a DAG: The DAG of this tutorial is a bit easier. It will consist of the following tasks: print 'hello' wait 5 seconds print 'world and we'll plan daily execution of this workflow. Create a DAG file Go to the folder that you've designated to be your AIRFLOWHOME and find the DAGs folder located in subfolder dags/ (if you cannot find, check the setting dagsfolder in $AIRFLOW_HOME/airflow.cfg). Create a Python file with the name airflow_tutorial.py that will contain your DAG. Your workflow will automatically be picked up and scheduled to run. First we'll configure settings that are shared by all our tasks. Settings for tasks can be passed as arguments when creating them, but we can also pass a dictionary with default values to the DAG. This allows us to share default arguments for all the tasks in our DAG is the best place to set e.g. the owner and start date of our DAG. Add the following import and dictionary to airflow_tutorial.py to specify the owner, start time, and retry settings that are shared by our tasks: Configure common settings These settings tell Airflow that this workflow is owned by 'me', that the workflow is valid since June 1st of 2017, it should not send emails and it is allowed to retry the workflow once if it fails with a delay of 5 minutes. Other common default arguments are email settings on failure and the end time. Create the DAG We'll now create a DAG object that will contain our tasks. Name it airflowtutorialv01 and pass default_args: With schedule_interval='0 0 *' we've specified a run at every hour 0; the DAG will run each day at 00:00. See crontab.guru for help deciphering cron schedule expressions. Alternatively, you can use strings like '@daily' and '@hourly'. We've used a context manager to create a DAG (new since 1.8). All the tasks for the DAG should be indented to indicate that they are part of this DAG. Without this context manager you'd have to set the dag parameter for each of your tasks. Airflow will generate DAG runs from the startdate with the specified scheduleinterval. Once a DAG is active, Airflow continuously checks in the database if all the DAG runs have successfully ran since the start_date. Any missing DAG runs are automatically scheduled. When you initialize on 2016-01-04 a DAG with a startdate at 2016-01-01 and a daily scheduleinterval, Airflow will schedule DAG runs for all the days between 2016-01-01 and 2016-01-04. A run starts after the time for the run has passed. The time for which the workflow runs is called the execution_date. The daily workflow for 2016-06-02 runs after 2016-06-02 23:59 and the hourly workflow for 2016-07-03 01:00 starts after 2016-07-03 01:59. From the ETL viewpoint this makes sense: you can only process the daily data for a day after it has passed. This can, however, ask for some juggling with date for other workflows. For Machine Learning models you may want to use all the data up to a given date, you'll have to add the scheduleinterval to your executiondate somewhere in the workflow logic. Because Airflow saves all the (scheduled) DAG runs in its database, you should not change the startdate and scheduleinterval of a DAG. Instead, up the version number of the DAG (e.g. airflowtutorialv02) and avoid running unnecessary tasks by using the web interface or command line tools Timezones and especially daylight savings can mean trouble when scheduling things, so keep your Airflow machine in UTC. You don't want to skip an hour because daylight savings kicks in (or out). Create the tasks Tasks are represented by operators that either perform an action, transfer data, or sense if something has been done. Examples of actions are running a bash script or calling a Python function; of transfers are copying tables between databases or uploading a file; and of sensors are checking if a file exists or data has been added to a database. We'll create a workflow consisting of three tasks: we'll print 'hello', wait for 10 seconds and finally print 'world'. The first two are done with the BashOperator and the latter with the PythonOperator. Give each operator an unique task ID and something to do: Note how we can pass bash commands in the BashOperator and that the PythonOperator asks for a Python function that can be called. Dependencies in tasks are added by setting other actions as upstream (or downstream). Link the operations in a chain so that sleep will be run after printhello and is followed by printworld; printhello -> sleep -> printworld: After rearranging the code your final DAG should look something like: Test the DAG First check that DAG file contains valid Python code by executing the file with Python: You can manually test a single task for a given execution_date with airflow test: This runs the task locally as if it was for 2017-07-01, ignoring other tasks and without communicating to the database. Activate the DAG Now that you're confident that your dag works, let's set it to run automatically! To do so, the scheduler needs to be turned on; the scheduler monitors all tasks and all DAGs and triggers the task instances whose dependencies have been met. Open a new terminal, activate the virtual environment and set the environment variable AIRFLOW_HOME for this terminal, and type Once the scheduler is up and running, refresh the DAGs page in the web UI. You should see airflowtutorialv01 in the list of DAGs with an on/off switch next to it. Turn on the DAG in the web UI and sit back while Airflow starts backfilling the dag runs! Tips Make your DAGs idempotent: rerunning them should give the same results. Use the the cron notation for schedule_interval instead of @daily and @hourly. @daily and @hourly always run after respectively midnight and the full hour, regardless of the hour/minute specified. Manage your connections and secrets with the Connections and/or Variables. Exercises You now know the basics of setting up Airflow, creating a DAG and turning it on; time to go deeper! Change the interval to every 30 minutes. Use a sensor to add a delay of 5 minutes before starting. Implement templating for the BashOperator: print the executiondate instead of 'hello' (check out the original tutorial and the example DAG). Implement templating for the PythonOperator: print the executiondate with one hour added in the function printworld() (check out the documentation of the PythonOperator). Resources Data Pipelines with Apache Airflow Airflow documentation ETL best practices with Airflow Airflow: Tips, Tricks, and Pitfalls Kubernetes Custom controller for deploying Airflow

10 Best AI Business Ideas 2024
youtube
LLM Vibe Score0.408
Human Vibe Score0.48
AI UncoveredMar 3, 2024

10 Best AI Business Ideas 2024

10 Best AI Business Ideas 2024 🔒 Keep Your Digital Life Private and Be Safe Online: https://nordvpn.com/safetyfirst Are you curious about the future of business in the exciting realm of Artificial Intelligence (AI)? Look no further! In this captivating video, we unveil the top 10 AI business ideas that are set to revolutionize the entrepreneurial landscape in 2024. From cutting-edge technology to innovative solutions, we delve into the most promising ventures that harness the power of AI to drive success and growth. Discover how AI is reshaping traditional business models and opening up endless possibilities for aspiring entrepreneurs. Whether you're a seasoned professional or a budding visionary, these handpicked AI business ideas offer a gateway to prosperity in the ever-evolving digital age. Join us as we explore groundbreaking concepts that blend creativity with computational intelligence, paving the way for unprecedented innovation and profitability. From automated customer service to personalized marketing strategies, AI is poised to transform every aspect of modern business operations. Dive deep into the realm of AI-powered startups and witness firsthand how these groundbreaking ideas are shaping the future of commerce. With our expert insights and comprehensive analysis, you'll gain invaluable knowledge to embark on your own AI-driven entrepreneurial journey. Don't miss out on the opportunity to stay ahead of the curve and capitalize on the transformative potential of AI in business. Join us as we unveil the 10 best AI business ideas for 2024 and embark on a journey towards success in the dynamic world of artificial intelligence. Subscribe now and stay tuned for more cutting-edge content that empowers you to thrive in the digital economy of tomorrow! #ai #artificialintelligence #aibusiness Subscribe for more! Welcome to AI Uncovered, your ultimate destination for exploring the fascinating world of artificial intelligence! Our channel delves deep into the latest AI trends and technology, providing insights into cutting-edge AI tools, AI news, and breakthroughs in artificial general intelligence (AGI). We simplify complex concepts, making AI explained in a way that is accessible to everyone. At AI Uncovered, we're passionate about uncovering the most captivating stories in AI, including the marvels of ChatGPT and advancements by organizations like OpenAI. Our content spans a wide range of topics, from science news and AI innovations to in-depth discussions on the ethical implications of artificial intelligence. Our mission is to enlighten, inspire, and inform our audience about the rapidly evolving technology landscape. Whether you're a tech enthusiast, a professional seeking to stay ahead of AI trends, or someone curious about the future of artificial intelligence, AI Uncovered is the perfect place to expand your knowledge. Join us as we uncover the secrets behind AI tools and their potential to revolutionize our world. Subscribe to AI Uncovered and stay tuned for enlightening content that bridges the gap between AI novices and experts, covering AI news, AGI, ChatGPT, OpenAI, artificial intelligence, and more. Together, let's explore the limitless possibilities of technology and AI. Disclaimer: Some links included in this description might be affiliate links. If you purchase a product or service through the links that we provide, we may receive a small commission. There is no additional charge for you. Thank you for supporting AI Uncovered so we can continue to provide you with free, high-quality content. _ 🌟 Contact: ai.uncovered.ai@gmail.com

Delivering Automations to Clients (How I Do It)
youtube
LLM Vibe Score0.375
Human Vibe Score0.52
Tyler Germain | AI AutomationJul 21, 2023

Delivering Automations to Clients (How I Do It)

In this video, I explain how to deliver AI Automation Services to your clients. Wether you are building a custom chatbot using Botpress & Stack AI, or you are building custom business process automations with Make.com, this video has you covered. Service delivery is a crucial component of running a respectable agency, and perfecting it will help you scale your business. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ ✅ How to Deliver Services for AI Automation Agencies ✅ Subscribe to Stay Up to Date @tylergermain 🤙🏻 ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ Instagram: https://www.instagram.com/itstylergermain Twitter: https://twitter.com/itstylergermain LinkedIn: https://www.linkedin.com/in/tylergermain Business Email: tyler@automationunlocked.co ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ I’m Tyler Germain, the creator of this channel, dedicated to AI automation. I’m sharing knowledge on starting, growing, and managing my automation agency, Automation Unlocked. I worked in the field of data analytics for years, before ultimately quitting my job to pursue AI automation full time. On this channel, I offer insights on securing clients, selling services, and even delving into real-life client projects. I teach viewers the skills to utilize powerful tools like make.com and zapier, as well as a bit of no-code magic. Join me on my channel to unlock the secrets of AI automation and discover a world of endless possibilities. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ Links: Botpress 👉 https://botpress.com/ Make.com 👉 https://www.make.com/en/register?pc=automationunlocked Automation Unlocked Memberships 👉 https://www.automationunlocked.co ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ Join the AI Automation Community 🤘 https://discord.gg/YQzppBm5Vf ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ Work with Our Team 🦾 https://www.automationunlocked.co/agency

airtable-api-proxy
github
LLM Vibe Score0.348
Human Vibe Score0.008293886065546695
danilocJul 10, 2023

airtable-api-proxy

node.js Airtable API Proxy by Future Fluent ================= Here's a project demonstrating the basics of an Airtable API proxy using node.js and Express. Click here to see the source and remix for your own purposes. Why does Airtable need an API Proxy? Airtable's rate limit is five requests per second per base. Anything more than that and the API will lock down for thirty seconds. By implementing an API proxy, it's possible to cache common results for quick responses and enforce a rate limit for requests. Additionally, an API proxy allows you to keep your API key a secret. Since all Airtable API keys allow full CRUD access, using the key in client-side JavaScript code would leave your data subject to outside tampering. Click here for example output. Click here to see the source data. How does it work? Three files drive the proxy: server.js An API route, /api/ai/list/:page, demonstrates how to serve JSON in response to a request. caching.js Simple, file-based caching. readCacheWithPath(path) Returns cached JSON, if it's not too stale. Use cacheInterval to adjust this. writeCacheWithPath(path, object) Writes a JavaScript object to JSON at the specified path, creating intermediate directories as needed. database-connection.js This is the meat of the project. It uses the Airtable node.js client to connect to a base and writes the results out as a JSON response. Base ID and Airtable API key are in 🗝.env. For more on accessing Airtable via the API, see the interactive Airtable documentation. Rate limiting Bottleneck handles rate limiting. The Airtable database interactions are handled using Bottleneck's wrap function.

The 9 AI Skills You Need NOW to Stay Ahead of 97% of People
youtube
LLM Vibe Score0.289
Human Vibe Score0.91
AI UncoveredMay 14, 2023

The 9 AI Skills You Need NOW to Stay Ahead of 97% of People

The 9 AI Skills You Need NOW to Stay Ahead of 97% of People 🔒 Keep Your Digital Life Private: Stay Safe & Secure Online with NordVPN: https://nordvpn.com/safetyfirst Welcome to our latest educational video, "The 9 AI Skills You Need NOW to Stay Ahead of 97% of People." This video is designed for anyone eager to take a deep dive into the world of artificial intelligence and machine learning. Our goal is to provide you with the most essential AI skills needed to excel in this rapidly evolving field, keeping you ahead of the curve and well-positioned in the job market. In this comprehensive guide, we explore nine fundamental AI skills, ranging from understanding algorithms to deep learning, data science, natural language processing, computer vision, robotics, and more. We also provide practical tips on how to apply these skills in real-world scenarios, whether you're an AI enthusiast or a seasoned professional. AI is not just the future; it's here NOW. By acquiring these nine essential AI skills, you can position yourself among the top 3% of people who are ready to shape the future. Don't be left behind as AI transforms industries, from healthcare and finance to entertainment and transportation. This video is a must-watch for anyone interested in AI, machine learning, data analysis, robotics, or any related field. Whether you're just starting out, looking to upskill, or aiming to stay ahead in your career, these nine AI skills are your key to success. Remember to subscribe to our channel for more valuable content and hit the notification bell so you never miss an update. Join the conversation in the comments section - we'd love to hear your thoughts on AI and how you plan to incorporate these skills into your career or studies. So get ready, click play, and let's take a step towards the future together, learning the 9 AI skills you need NOW to stay ahead of 97% of people. Your AI journey starts here. Enjoy the video! #artificialintelligence #ai #airevolution Subscribe for more! Welcome to AI Uncovered, your ultimate destination for exploring the fascinating world of artificial intelligence! Our channel delves deep into the latest AI trends and technology, providing insights into cutting-edge AI tools, AI news, and breakthroughs in artificial general intelligence (AGI). We simplify complex concepts, making AI explained in a way that is accessible to everyone. At AI Uncovered, we're passionate about uncovering the most captivating stories in AI, including the marvels of ChatGPT and advancements by organizations like OpenAI. Our content spans a wide range of topics, from science news and AI innovations to in-depth discussions on the ethical implications of artificial intelligence. Our mission is to enlighten, inspire, and inform our audience about the rapidly evolving technology landscape. Whether you're a tech enthusiast, a professional seeking to stay ahead of AI trends, or someone curious about the future of artificial intelligence, AI Uncovered is the perfect place to expand your knowledge. Join us as we uncover the secrets behind AI tools and their potential to revolutionize our world. Subscribe to AI Uncovered and stay tuned for enlightening content that bridges the gap between AI novices and experts, covering AI news, AGI, ChatGPT, OpenAI, artificial intelligence, and more. Together, let's explore the limitless possibilities of technology and AI. Disclaimer: Some links included in this description might be affiliate links. If you purchase a product or service through the links that we provide, we may receive a small commission. There is no additional charge for you. Thank you for supporting AI Uncovered so we can continue to provide you with free, high-quality content.