VibeBuilders.ai Logo
VibeBuilders.ai

Algorithms

Explore resources related to algorithms to help implement AI solutions for your business.

Scratch Machine Learning Algorithms Implementations
reddit
LLM Vibe Score0
Human Vibe Score1
ParkMountainThis week

Scratch Machine Learning Algorithms Implementations

Hi there, other Redditors! Like many of you, when I first started working in the AI field, I wanted to build some basic Machine Learning models from scratch in order to better understand how each algorithm works, improve my programming and math skills, or simply produce an eye-catching, difficult project to put in the résumé. After spending some time searching for resources that could help me guide my studies, I discovered that the majority of scratch implementations that are currently available are either i) outdated (having been implemented years ago using Python 2 or an earlier version of Python 3); ii) too difficult to understand (using a lot of difficult, unfriendly optimization techniques or with poorly written code); or iii) too simple (only covering binary classification). With that in mind, I made the decision to develop user-friendly, uncomplicated, organized, and simple implementations from scratch. Aside from all of that, I've always wanted to create an open-source project so that others, particularly novices and those with less than a year's experience (like me), can collaborate with others, contribute to public projects, and experience Git firsthand (some of these implementations were made by other contributors!). Here are some implementations that are available: Algorithms (Random Forest Classifier and Regressor, Decision Tree Classifier and Regressor, KMeans, KNN Classifier and Regressor, Gaussian Naive Bayes, Linear Regression, Logistic Regression, PCA, Perceptron, MLP Classifier and Regressor, SVM Classifier and Regressor); Regression and classification metrics; Distance metrics (such as Euclidean); Data split functions (such as KFold); Activation and loss functions; Scalers (such as MinMaxScaler) and encoders (such as One Hot Encoder); and a few things more! Project's link: https://github.com/rafaelgreca/scratchml Disclaimer: The goal of this library is to provide code that is simpler, easier to understand, and more approachable for artificial intelligence enthusiasts and beginners who want to contribute to an open-source repository or who want to learn more about how algorithms work. It is not meant to replace existing libraries that are better, more optimized, and have a wider variety of implemented algorithms (such as scikit-learn, PyTorch, Keras, and Tensorflow). If you want to use optimized implementations with accurate results, please use one of the previously mentioned libraries. P.S.: I accidentally deleted the other post, so I am posting again. :-)

[D] Should We Be Concerned About The Failure Of Evolutionary Algorithms, And Its Implications?
reddit
LLM Vibe Score0
Human Vibe Score-1
mystikaldangerThis week

[D] Should We Be Concerned About The Failure Of Evolutionary Algorithms, And Its Implications?

https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6287292/ ​ A number of possible explanations for \[why we can't evolve complex software\] could be considered. We tried to be as comprehensive as possible in this section, but it is possible that we have not considered some plausible explanations: Incompetent programmers—It is theoretically possible, but is highly unlikely, that out of thousands of scientists working on evolutionary computation, all failed to correctly implement the Darwinian algorithm. Nonrepresentative algorithms—Some have suggested that EAs do not accurately capture the theory of evolution, but of course that would imply that the theory itself is not specified in sufficient detail to make falsifiable predictions. If, however, such more detailed specifications are available to GP believers, it is up to them to implement them as computer simulations for testing purposes, but no successful examples of such work are known and the known ones have not been successful in evolving software. Inadequate fitness functions—Fitness function for a complex software product is difficult to outline and specify and may be as complex (or even more complex) as the software we want to evolve as it has to consider all the possible use cases and pass all unit tests. This may be the Achilles heel of GP, but it is also an objection to feasibility of programming in general and GP in particular, as both have to convert software specification into the source code. If human programmers and biological evolution succeed with such constraints, so should Darwinian simulations. The Halting problem—Turing proved that it is impossible to determine whether an arbitrary program halts, but this is also a problem for human programmers and could be easily addressed by placing time limits on considered solutions. Program correctness—If we require evolved software to be provably correct, this would present a problem as GP does not verify produced designs but only tests them against specific unit tests. Likewise, we cannot rely on automated software verification as it is still an unsolved problem in the general case. This is not really a problem as most of the human-written software is never proven to be correct and only a small portion of software engineering process relies of formal specification and Test Driven Development. Inappropriate solutions—Literature on EA is full of examples of surprising creativity of Darwinian algorithm resulting in solutions which match the letter of design specifications but not the spirit. This is similar to human-produced software and numerous examples of ways in which such software fails the goals of the initial design. Insufficient complexity of the environment (not enough data, poor fitness functions)—It is possible that the simulated environment is not complex enough to generate high complexity outputs in evolutionary simulations. This does not seem correct as Internet presents a highly complex landscape in which many self-modifying computer viruses roam. Likewise, virtual world such as Second Life and many others present close approximations to the real world and are certainly more complex than early Earth was: A skeptic might insist that an abstract environment would be inadequate for the evolution . . ., believing instead that the virtual environment would need to closely resemble the actual biological environment in which our ancestors evolved. Creating a physically realistic virtual world would require a far greater investment of computational resources than the simulation of a simple toy world or abstract problem domain (whereas evolution had access to a physically realistic real world “for free”). In the limiting case, if complete microphysical accuracy were insisted upon, the computational requirements would balloon to utterly infeasible proportions. Requiring more realistic environmental conditions may result in an increase in necessary computational resources, a problem addressed in the next bullet. Insufficient resources (compute, memory)—From the history of computer science, we know of many situations (speech recognition, NN training), where we had a correct algorithm but insufficient computational resources to run it to success. It is possible that we simply do not have hardware powerful enough to emulate evolution. We will address this possibility in section “Computational Complexity of Biological Evolution and Available Compute.” Software design is not amenable to evolutionary methods—Space of software designs may be discrete with no continuous path via incremental fitness to the desired solutions. This is possible, but this implies that original goals of GP are unattainable and misguided. In addition, because a clear mapping exists between solutions to problems and animals as solutions to environmental problems, this would also imply that current explanation for the origin of the species is incorrect. Darwinian algorithm is incomplete or wrong—Finally, we have to consider the possibility that the inspiration behind evolutionary computation, the Darwinian algorithm itself is wrong or at least partially incomplete. If that was true, computer simulations of such algorithm would fail to produce results comparable with observations we see in nature and a search for an alternative algorithm would need to take place. This would be an extraordinary claim and would require that we discard all the other possible explanations from this list. We challenge EA community to prove us wrong by producing an experiment, which evolves nontrivial software from scratch and without human help. That would be the only way in which our findings could be shown to be incorrect. Perhaps, reframing the problem in terms of maximizing negentropy of digital organisms, as suggested by Schrödinger, Michaelian, and Ulanowicz and Hannon, with respect to negative energy being a fundamental property of all life-forms may produce better results. On a positive side, the fact that it seems impossible to evolve complex software implies that we are unlikely to be able to evolve highly sophisticated artificially intelligent agents, which may present significant risk to our safety and security. Just imagine what would have happened, if the very first time we ran a simulation of evolution on a computer, it produced a superintelligent agent. Yampolskiy has shown that programming as a problem is AI-complete; if GP can solve programming that would imply that GP = AGI (artificial general intelligence), but we see no experimental evidence for such claim. In fact, it is more likely that once we have AGI, it could be used to create an intelligent fitness function for GP and so evolve software. Genetic programming will not be the cause of AI, but a product of it. However, neuroevolution methods for optimizing deep learning architectures and parameters remain a strong possibility for creation of AGI.

Content aggregation that acts as a middleman for content discovery via third-party marketplace & revenue sharing (i will not promote but I'm looking for fellow researchers)
reddit
LLM Vibe Score0
Human Vibe Score1
colbyn-wadmanThis week

Content aggregation that acts as a middleman for content discovery via third-party marketplace & revenue sharing (i will not promote but I'm looking for fellow researchers)

High level I’m considering a content aggregation business model, but one that acts as an open marketplace where third party devs and where world class data scientists compete to build the best recommenders for different use cases. (E.g. the incentives can be ad revenue sharing or subscription based for niche professional markets.) The idea is to facilitate more bottom up innovation from third party data scientists. The platform itself just acts as the middleman. (Also something that strips out original ads and makes it easy to skip paid sponsorship sections would be great.)  I’ve seen startups building web crawlers and content aggregation systems for other AI startups. My proposal is better in the sense that third party devs are instead responsible for implementing whatever questionable hacks are necessarily to scrape platforms that don’t necessarily want to be scraped.  Personally, I’m more concerned about getting the right information than ever before, to this end I can’t rely on platform specific recommenders. The solution is more bottom up innovation in content promotion. More generally, if you’re also concerned about consuming game changing information that’s too easily missed: we need a platform that incentivizes bottom up innovation of content promotion. What we need is a platform that functions like a marketplace where third party devs and where world class data scientists compete to build the best recommenders for different use cases. Here’s some elevator pitches I’m considering:  Did you know that the magic behind YouTube is its recommendation engine? Now, imagine an open platform where independent engines compete to deliver the most personalized content feed—from news to local events—directly to you. Interested in rethinking how we find content? “In today’s fragmented digital landscape, a single platform no longer holds sway over content discovery. The Network Effect is dead: audiences are more mobile than ever; and big tech killed it. In such a fragmented landscape we’re building a bottom-up, decentralized marketplace for recommendation engines—a solution that taps into diverse revenue streams through subscriptions, ad revenue, and affiliate partnerships. Invest in the future of personalized content aggregation.” “Are you a developer passionate about algorithms and content discovery? Our open marketplace lets you build and monetize your own recommendation engine, competing to deliver the most engaging, personalized feeds. Join a revolution where your innovation can directly shape how the world finds content.” “Are you tired of being told what to watch or read by one mysterious algorithm? Imagine taking control—choosing from a marketplace of smart recommendation engines that curate content just for you. It’s a revolution in content discovery where you hold the power.” (As a Utahn this one is interesting because even mormons are talking about the dangers of “doom scrolling” though it’s seldom discussed in society at large.) As far as simple hooks I’m considering:  One platform to rule them all and in the darkness bind them.  Choose how you discover—content recommenders that work for you.  The area where recommender engines battle to win your feed. Request I would love to start prototyping this idea and see what else I can uncover from such preliminary research. But I want to get a couple other likeminded individuals onboard.  I'm the best when it comes to iOS/macOS development, but there's tons of backend work that needs to be done which I wouldn’t have the time for if i'm focused on the native clients. Who am I 'ideally' looking for?  I’ve heard of weird stats to the effect that if you scale up a population to billions of people, the number of life overlaps starts skyrocketing. Not just physical lookalikes, but people with eerily similar life paths, personalities, habits, and even thoughts — without ever knowing each other. Where are my clones? Such is whom I’m looking for in an ideal world.  Take a hunch  People nowadays have no concept of going out on a limb, taking a ‘hunch’, and backing their instincts. Everything has to be calculated, proven, and guaranteed before they make a move. In contrast consider the success of the Chinese DeepSeek project: According to Asianometry’s YouTube video on DeepSeek, their “memory-saving multi-head latent architecture” (whatever that means, just quoting the name) came about from a researchers ‘hunch’, which the company bet big on and the result was drastically improved performance on low end hardware…  Here in the west the idea of betting on a hunch is inconceivable. We have no balls to chase long term insights. My own instincts when it comes to software is such because I’ve wasted too much of my life on small scale projects. All I’m trying to do is attempt a more scaled up experiment based on some hunches with me and a few other likeminded individuals.  Just as the early oil prospectors didn’t have precise maps—just intuition and test drills. They had to drill, analyze the pressure, and adjust. The best oil fields weren’t found by foresight alone, but by adaptive exploration. The startup space itself is liken to the first prospectors who got the gold nuggets lying in the riverbed. In such an environment moving first has its advantages but nowadays I wish I could have all those shitty ‘engineers’ sent to their maker.  Today the reality is such that you’ve got to dig deep—where vast stores of wealth can be found—or go home, and those who dig into the depths cannot use mere forethought, for what lies beneath cannot be seen by the mind’s eye.  I will not promote but I'm looking for fellow research oriented minds.

10y of product development, 2 bankruptcies, and 1 Exit — what next? [Extended Story]
reddit
LLM Vibe Score0
Human Vibe Score1
Slight-Explanation29This week

10y of product development, 2 bankruptcies, and 1 Exit — what next? [Extended Story]

10 years of obsessive pursuit from the bottom to impressive product-market fit and exit. Bootstrapping tech products as Software Developer and 3x Startup Founder (2 bankruptcies and 1 exit). Hi everyone, your motivation has inspired me to delve deeper into my story. So, as promised to some of you, I've expanded on it a bit more, along with my brief reflections. There are many founders, product creators, and proactive individuals, I’ve read many of your crazy stories and lessons so I decided to share mine and the lessons I learned from the bottom to impressive product-market fit and exit. I've spent almost the past 10 years building tech products as a Corporate Team Leader, Senior Software Developer, Online Course Creator, Programming Tutor, Head of Development/CTO, and 3x Startup Founder (2 bankruptcies, and 1 exit). And what next? good question... A brief summary of my journey: Chapter 1: Software Developer / Team Leader / Senior Software Developer I’ve always wanted to create products that win over users’ hearts, carry value, and influence users. Ever since my school days, I’ve loved the tech part of building digital products. At the beginning of school, I started hosting servers for games, blogs and internet forums, and other things that did not require much programming knowledge. My classmates and later even over 100 people played on servers that I hosted on my home PC. Later, as the only person in school, I passed the final exam in computer science. During my computer science studies, I started my first job as a software developer. It was crazy, I was spending 200–300 hours a month in the office attending also to daily classes. Yes, I didn’t have a life, but it truly was the fulfillment of my dreams. I was able to earn good money doing what I love, and I devoted fully myself to it. My key to effectively studying IT and growing my knowledge at rocket speed was learning day by day reading guides, building products to the portfolio, watching youtube channels and attending conferences, and even watching them online, even if I didn’t understand everything at the beginning. In one year we’ve been to every possible event within 400km. We were building healthcare products that were actually used in hospitals and medical facilities. It was a beautiful adventure and tons of knowledge I took from this place. That time I built my first product teams, hired many great people, and over the years became a senior developer and team leader. Even I convinced my study mates to apply to this company and we studied together and worked as well. Finally, there were 4 of us, when I left a friend of mine took over my position and still works there. If you’re reading this, I’m sending you a flood of love and appreciation. I joined as the 8th person, and after around 4 years, when I left hungry for change, there were already over 30 of us, now around 100. It was a good time, greetings to everyone. I finished my Master’s and Engineering degrees in Computer Science, and it was time for changes. Chapter 2: 1st time as a Co-founder — Marketplace In the meantime, there was also my first startup (a marketplace) with four of my friends. We all worked on the product, each of us spent thousands of hours, after hours, entire weekends… and I think finally over a year of work. As you might guess, we lacked the most important things: sales, marketing, and product-market fit. We thought users think like us. We all also worked commercially, so the work went very smoothly, but we didn’t know what we should do next with it… Finally, we didn’t have any customers, but you know what, I don’t regret it, a lot of learning things which I used many times later. The first attempts at validating the idea with the market and business activities. In the end, the product was Airbnb-sized. Landing pages, listings, user panels, customer panels, admin site, notifications, caches, queues, load balancing, and much more. We wanted to publish the fully ready product to the market. It was a marketplace, so if you can guess, we had to attract both sides to be valuable. “Marketplace” — You can imagine something like Uber, if you don’t have passengers it was difficult to convince taxi drivers, if you don’t have a large number of taxi drivers you cannot attract passengers. After a year of development, we were overloaded, and without business, marketing, sales knowledge, and budget. Chapter 3: Corp Team Lead / Programming Tutor / Programming Architecture Workshop Leader Working in a corporation, a totally different environment, an international fintech, another learning experience, large products, and workmates who were waiting for 5 pm to finish — it wasn’t for me. Very slow product development, huge hierarchy, being an ant at the bottom, and low impact on the final product. At that time I understood that being a software developer is not anything special and I compared my work to factory worker. Sorry for that. High rates have been pumped only by high demand. Friends of mine from another industry do more difficult things and have a bigger responsibility for lower rates. That’s how the market works. This lower responsibility time allowed for building the first online course after hours, my own course platform, individual teaching newbies programming, and my first huge success — my first B2C customers, and B2B clients for workshops. I pivoted to full focus on sales, marketing, funnels, advertisements, demand, understanding the market, etc. It was 10x easier than startups but allowed me to learn and validate my conceptions and ideas on an easier market and showed me that it’s much easier to locate their problem/need/want and create a service/product that responds to it than to convince people of your innovative ideas. It’s just supply and demand, such a simple and basic statement, in reality, is very deep and difficult to understand without personal experience. If you’re inexperienced and you think you understand, you don’t. To this day, I love to analyze this catchword in relation to various industries / services / products and rediscover it again and again... While writing this sentence, I’m wondering if I’m not obsessed. Chapter 4: Next try — 2nd time as a founder — Edtech Drawing upon my experiences in selling services, offering trainings, and teaching programming, I wanted to broaden my horizons, delve into various fields of knowledge, involve more teachers, and so on. We started with simple services in different fields of knowledge, mainly relying on teaching in the local area (without online lessons). As I had already gathered some knowledge and experience in marketing and sales, things were going well and were moving in the right direction. The number of teachers in various fields was growing, as was the number of students. I don’t remember the exact statistics anymore, but it was another significant achievement that brought me a lot of satisfaction and new experiences. As you know, I’m a technology lover and couldn’t bear to look at manual processes — I wanted to automate everything: lessons, payments, invoices, customer service, etc. That’s when I hired our first developers (if you’re reading this, I’m sending you a flood of love — we spent a lot of time together and I remember it as a very fruitful and great year) and we began the process of tool and automation development. After a year we had really extended tools for students, teachers, franchise owners, etc. We had really big goals, we wanted to climb higher and higher. Maybe I wouldn’t even fully call it Startup, as the client was paying for the lessons, not for the software. But it gave us positive income, bootstrap financing, and tool development for services provided. Scaling this model was not as costless as SaaS because customer satisfaction was mainly on the side of the teacher, not the quality of the product (software). Finally, we grew to nearly 10 people and dozens of teachers, with zero external funding, and almost $50k monthly revenue. We worked very hard, day and night, and by November 2019, we were packed with clients to the brim. And as you know, that’s when the pandemic hit. It turned everything upside down by 180 degrees. Probably no one was ready for it. With a drastic drop in revenues, society started to save. Tired from the previous months, we had to work even harder. We had to reduce the team, change the model, and save what we had built. We stopped the tool’s development and sales, and with the developers, we started supporting other product teams to not fire them in difficult times. The tool worked passively for the next two years, reducing incomes month by month. With a smaller team providing programming services, we had full stability and earned more than relying only on educational services. At the peak of the pandemic, I promised myself that it was the last digital product I built… Never say never… Chapter 5: Time for fintech — Senior Software Developer / Team Lead / Head of Development I worked for small startups and companies. Building products from scratch, having a significant impact on the product, and complete fulfillment. Thousands of hours and sacrifices. This article mainly talks about startups that I built, so I don’t want to list all the companies, products, and applications that I supported as a technology consultant. These were mainly start-ups with a couple of people up to around 100 people on board. Some of the products were just a rescue mission, others were building an entire tech team. I was fully involved in all of them with the hope that we would work together for a long time, but I wasn’t the only one who made mistakes when looking for a product-market fit. One thing I fully understood: You can’t spend 8–15 hours a day writing code, managing a tech team, and still be able to help build an audience. In marketing and sales, you need to be rested and very creative to bring results and achieve further results and goals. If you have too many responsibilities related to technology, it becomes ineffective. I noticed that when I have more free time, more time to think, and more time to bounce the ball against the wall, I come up with really working marketing/sales strategies and solutions. It’s impossible when you are focused on code all day. You must know that this chapter of my life was long and has continued until now. Chapter 6: 3rd time as a founder — sold Never say never… right?\\ It was a time when the crypto market was really high and it was really trending topic. You know that I love technology right? So I cannot miss the blockchain world. I had experience in blockchain topics by learning on my own and from startups where I worked before. I was involved in crypto communities and I noticed a “starving crowd”. People who did things manually and earned money(crypto) on it.I found potential for building a small product that solves a technological problem. I said a few years before that I don’t want to start from scratch. I decided to share my observations and possibilities with my good friend. He said, “If you gonna built it, I’m in”. I couldn’t stop thinking about it. I had thought and planned every aspect of marketing and sales. And you know what. On this huge mindmap “product” was only one block. 90% of the mindmap was focused on marketing and sales. Now, writing this article, I understood what path I went from my first startup to this one. In the first (described earlier) 90% was the product, but in the last one 90% was sales and marketing. Many years later, I did this approach automatically. What has changed in my head over the years and so many mistakes? At that time, the company for which I provided services was acquired. The next day I got a thank you for my hard work and all my accounts were blocked. Life… I was shocked. We were simply replaced by their trusted technology managers. They wanted to get full control. They acted a bit unkindly, but I knew that they had all my knowledge about the product in the documentation, because I’m used to drawing everything so that in the moment of my weakness (illness, whatever) the team could handle it. That’s what solid leaders do, right? After a time, I know that these are normal procedures in financial companies, the point is that under the influence of emotions, do not do anything inappropriate. I quickly forgot about it, that I was brutally fired. All that mattered was to bring my plan to life. And it has been started, 15–20 hours a day every day. You have to believe me, getting back into the game was incredibly satisfying for me. I didn’t even know that I would be so excited. Then we also noticed that someone was starting to think about the same product as me. So the race began a game against time and the market. I assume that if you have reached this point, you are interested in product-market fit, marketing, and sales, so let me explain my assumptions to you: Product: A very very small tool that allowed you to automate proper tracking and creation of on-chain transactions. Literally, the whole app for the user was located on only three subpages. Starving Crowd: We tapped into an underserved market. The crypto market primarily operates via communities on platforms like Discord, Reddit, Twitter, Telegram, and so on. Therefore, our main strategy was directly communicating with users and demonstrating our tool. This was essentially “free marketing” (excluding the time we invested), as we did not need to invest in ads, promotional materials, or convince people about the efficacy of our tool. The community could directly observe on-chain transactions executed by our algorithms, which were processed at an exceptionally fast rate. This was something they couldn’t accomplish manually, so whenever someone conducted transactions using our algorithm, it was immediately noticeable and stirred a curiosity within the community (how did they do that!). Tests: I conducted the initial tests of the application on myself — we had already invested significantly in developing the product, but I preferred risking my own resources over that of the users. I provided the tool access to my wallet, containing 0.3ETH, and went to sleep. Upon waking up, I discovered that the transactions were successful and my wallet had grown to 0.99ETH. My excitement knew no bounds, it felt like a windfall. But, of course, there was a fair chance I could have lost it too. It worked. As we progressed, some users achieved higher results, but it largely hinged on the parameters set by them. As you can surmise, the strategy was simple — buy low, sell high. There was considerable risk involved. Churn: For those versed in marketing, the significance of repeat visitors cannot be overstated. Access to our tool was granted only after email verification and a special technique that I’d prefer to keep confidential. And this was all provided for free. While we had zero followers on social media, we saw an explosion in our email subscriber base and amassed a substantial number of users and advocates. Revenue Generation: Our product quickly gained popularity as we were effectively helping users earn — an undeniable value proposition. Now, it was time to capitalize on our efforts. We introduced a subscription model charging $300 per week or $1,000 per month — seemingly high rates, but the demand was so intense that it wasn’t an issue. Being a subscriber meant you were prioritized in the queue, ensuring you were among the first to reap benefits — thus adding more “value”. Marketing: The quality of our product and its ability to continually engage users contributed to it achieving what can best be described as viral. It was both a source of pride and astonishment to witness users sharing charts and analyses derived from our tool in forum discussions. They weren’t actively promoting our product but rather using screenshots from our application to illustrate certain aspects of the crypto world. By that stage, we had already assembled a team to assist with marketing, and programming, and to provide round-the-clock helpdesk support. Unforgettable Time: Despite the hype, my focus remained steadfast on monitoring our servers, their capacity, and speed. Considering we had only been on the market for a few weeks, we were yet to implement alerts, server scaling, etc. Our active user base spanned from Japan to the West Coast of the United States. Primarily, our application was used daily during the evenings, but considering the variety of time zones, the only time I could afford to sleep was during the evening hours in Far Eastern Europe, where we had the least users. However, someone always needed to be on guard, and as such, my phone was constantly by my side. After all, we couldn’t afford to let our users down. We found ourselves working 20 hours a day, catering to thousands of users, enduring physical fatigue, engaging in talks with VCs, and participating in conferences. Sudden Downturn: Our pinnacle was abruptly interrupted by the war in Ukraine (next macroeconomic shot straight in the face, lucky guy), a precipitous drop in cryptocurrency value, and swiftly emerging competition. By this time, there were 5–8 comparable tools had infiltrated the market. It was a challenging period as we continually stumbled upon new rivals. They immediately embarked on swift fundraising endeavors — a strategy we overlooked, which in retrospect was a mistake. Although our product was superior, the competitors’ rapid advancement and our insufficient funds for expeditious scaling posed significant challenges. Nonetheless, we made a good decision. We sold the product (exit) to competitors. The revenue from “exit” compensated for all the losses, leaving us with enough rest. We were a small team without substantial budgets for rapid development, and the risk of forming new teams without money to survive for more than 1–2 months was irresponsible. You have to believe me that this decision consumed us sleepless nights. Finally, we sold it. They turned off our app but took algorithms and users. Whether you believe it or not, after several months of toiling day and night, experiencing burnout, growing weary of the topic, and gaining an extra 15 kg in weight, we finally found our freedom… The exit wasn’t incredibly profitable, but we knew they had outdone us. The exit covered all our expenses and granted us a well-deserved rest for the subsequent quarter. It was an insane ride. Despite the uncertainty, stress, struggles, and sleepless nights, the story and experience will remain etched in my memory for the rest of my life. Swift Takeaways: Comprehending User Needs: Do you fully understand the product-market fit? Is your offering just an accessory or does it truly satisfy the user’s needs? The Power of Viral Marketing: Take inspiration from giants like Snapchat, ChatGPT, and Clubhouse. While your product might not attain the same scale (but remember, never say never…), the closer your concept is to theirs, the easier your journey will be. If your user is motivated to text a friend saying, “Hey, check out how cool this is” (like sharing ChatGPT), then you’re on the best track. Really. Even if it doesn’t seem immediately evident, there could be a way to incorporate this into your product. Keep looking until you find it. Niche targeting — the more specific and tailored your product is to a certain audience, the easier your journey will be People love buying from people — establishing a personal brand and associating yourself with the product can make things easier. Value: Seek to understand why users engage with your product and keep returning. The more specific and critical the issue you’re aiming to solve, the easier your path will be. Consider your offerings in terms of products and services and focus on sales and marketing, regardless of personal sentiments. These are just a few points, I plan to elaborate on all of them in a separate article. Many products undergo years of development in search of market fit, refining the user experience, and more. And guess what? There’s absolutely nothing wrong with that. Each product and market follows its own rules. Many startups have extensive histories before they finally make their mark (for instance, OpenAI). This entire journey spanned maybe 6–8 months. I grasped and capitalized on the opportunity, but we understood from the start that establishing a startup carried a significant risk, and our crypto product was 10 times riskier. Was it worth it? Given my passion for product development — absolutely. Was it profitable? — No, considering the hours spent — we lose. Did it provide a stable, problem-free life — nope. Did this entire adventure offer a wealth of happiness, joy, and unforgettable experiences — definitely yes. One thing is certain — we’ve amassed substantial experience and it’s not over yet :) So, what lies ahead? Chapter 7: Reverting to the contractor, developing a product for a crypto StartupReturning to the past, we continue our journey… I had invested substantial time and passion into the tech rescue mission product. I came on board as the technical Team Leader of a startup that had garnered over $20M in seed round funding, affiliated with the realm of cryptocurrencies. The investors were individuals with extensive backgrounds in the crypto world. My role was primarily technical, and there was an abundance of work to tackle. I was fully immersed, and genuinely devoted to the role. I was striving for excellence, knowing that if we secured another round of financing, the startup would accelerate rapidly. As for the product and marketing, I was more of an observer. After all, there were marketing professionals with decades of experience on board. These were individuals recruited from large crypto-related firms. I had faith in them, kept an eye on their actions, and focused on my own responsibilities. However, the reality was far from satisfactory. On the last day, the principal investor for the Series A round withdrew. The board made the tough decision to shut down. It was a period of intense observation and gaining experience in product management. This was a very brief summary of the last 10 years. And what next? (Last) Chapter 8: To be announced — Product Owner / Product Consultant / Strategist / CTO After spending countless hours and days deliberating my next steps, one thing is clear: My aspiration is to continue traversing the path of software product development, with the hopeful anticipation that one day, I might ride the crest of the next big wave and ascend to the prestigious status of a unicorn company. I find myself drawn to the process of building products, exploring product-market fit, strategizing, engaging in software development, seeking out new opportunities, networking, attending conferences, and continuously challenging myself by understanding the market and its competitive landscape. Product Owner / Product Consultant / CTO / COO: I’m not entirely sure how to categorize this role, as I anticipate that it will largely depend on the product to which I will commit myself fully. My idea is to find one startup/company that wants to build a product / or already has a product, want to speed up, or simply doesn’t know what’s next. Alternatively, I could be a part of an established company with a rich business history, which intends to invest in digitization and technological advancements. The goal would be to enrich their customer experience by offering complementary digital products Rather than initiating a new venture from ground zero with the same team, I am receptive to new challenges. I am confident that my past experiences will prove highly beneficial for the founders of promising, burgeoning startups that already possess a product, or are in the initial phases of development. ‘Consultant’ — I reckon we interpret this term differently. My aim is to be completely absorbed in a single product, crafting funnels, niches, strategies, and all that is necessary to repeatedly achieve the ‘product-market fit’ and significant revenue. To me, ‘consultant’ resonates more akin to freelancing than being an employee. My current goal is to kickstart as a consultant and aide, dealing with facilitating startups in their journey from point A to B. Here are two theoretical scenarios to illustrate my approach: Scenario 1: (Starting from point A) You have a product but struggle with marketing, adoption, software, strategy, sales, fundraising, or something else. I conduct an analysis and develop a strategy to reach point B. I take on the “dirty work” and implement necessary changes, including potential pivots or shifts (going all-in) to guide the product to point B. The goal is to reach point B, which could involve achieving a higher valuation, expanding the user base, increasing sales, or generating monthly revenue, among other metrics. Scenario 2: (Starting from point A) You have a plan or idea but face challenges with marketing, adoption, strategy, software, sales, fundraising, or something else. I analyze the situation and devise a strategy to reach point B. I tackle the necessary tasks, build the team, and overcome obstacles to propel the product to point B. I have come across the view that finding the elusive product-market fit is the job of the founder, and it’s hard for me to disagree. However, I believe that my support and experiences can help save money, many failures, and most importantly, time. I have spent a great deal of time learning from my mistakes, enduring failure after failure, and even had no one to ask for support or opinion, which is why I offer my help. Saving even a couple of years, realistically speaking, seems like a value I’m eager to provide… I invite you to share your thoughts and insights on these scenarios :) Closing Remarks: I appreciate your time and effort in reaching this point. This has been my journey, and I wouldn’t change it for the world. I had an extraordinary adventure, and now I’m ready for the next exciting battle with the market and new software products. While my entire narrative is centered around startups, especially the ones I personally built, I’m planning to share more insights drawn from all of my experiences, not just those as a co-founder. If you’re currently developing your product or even just considering the idea, I urge you to reach out to me. Perhaps together, we can create something monumental :) Thank you for your time and insights. I eagerly look forward to engaging in discussions and hearing your viewpoints. Please remember to like and subscribe. Nothing motivates to write more than positive feedback :) Matt.

Online Reputation AI - Startup got stuck
reddit
LLM Vibe Score0
Human Vibe Score0.6
kyr0x0This week

Online Reputation AI - Startup got stuck

Hi, I‘m one of 3 co-founders of a startup that built an AI-driven SaaS and App product this year. We‘re coming from an SaaS background, two of us senior developers (in the 3% of highest earning freelancers in Germany) and expert in our fields. The third is a seasoned sales strategist. We have a minor 4th co-founder (legal advisor). The company is self-funded, no investors. Our tech is owned by us, built by us and the product was already operational after a few months. We basically solve three data science/NLP issues in a generalized way: understand customer feedback to improve your business. Analyzes online review with context and explains it with a drill down, aggregation, charts (AI insights, timeframe reports); evidence driven, agentic LLM and ETL processes drive this. respond to customer feedback, half-automated, human in the loop, but AI supported. In the tone of your brand, any language. And context-aware, with your customer support signature etc. competitor analysis. Because we do 1 for you, we can do 1. for all of your competitors and compare the results, yielding insights like „oh, this happens to everyone in November to December, so I should focus on something else“ — etc. Now, after a huge sales effort we got only one paying customer. This customer is petty happy with the product. They tell us that they use our product daily, it‘s better than all the other solutions out there (better than TrustYou, etc.) However, after cold calling/emailing hundreds of leads, we almost always hear that „what we have is good enough“. Or that they don‘t have budget. I‘m the introverted tech part of the startup. I‘m good with algorithms. Give me any tech issue and I will solve it for you quickly and efficiently. I make stuff work. But with my startups I never had commercial luck. People always tell me about my stellar potential, because I can build things almost nobody else can. I come from a poor families background, worked my way up the very hard way. I just love tech and programming. I wrote a book for O’Reilly once. I‘m not doing bad economically, but I‘m probably not the best sales person. After founding a few startups with amazing tech, people using the products and loving them, but no commercial success, I truly question myself and if I‘m just unlucky with the fact that I‘m located in Europe, targeting the wrong industries, or are just unlucky somehow? I won‘t blame my co-founders here. They definitely did the best they could. I‘m just a bit resignated. I recently thought about valuing my own lifetime more and only building software for myself anymore. Basically not focusing on what problems other people face and trying to solve them, but solely focusing on what I enjoy doing most — e.g. coding algorithms for a music visualizer. Because in the end, my time is my most valuable resource. If I waste any second on something that isn‘t contributing to „my life“ and how I define success, then it would be a rather stupid deed? I don‘t want to derail too much here. I‘m confused and seeking for advice. Burn me if you like, but please be aware that you are talking to a broadly educated nerd.

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age
reddit
LLM Vibe Score0
Human Vibe Score1
ExpenectThis week

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age Introduction In an era characterized by rapid technological advancements, the field of finance is undergoing a transformative journey. The emergence of financial technology, or fintech, is reshaping the way businesses manage their finances, and Chartered Accountants (CAs) are at the forefront of this evolution. In this blog post, we'll explore how CAs are embracing fintech and leveraging its potential to enhance financial management, analysis, and advisory services. Fintech's Impact on Financial Services Fintech encompasses a wide range of technologies that leverage data analytics, artificial intelligence, blockchain, and automation to improve financial services. For CAs, this means new tools to streamline processes, enhance decision-making, and offer innovative solutions to clients. Automation of Routine Tasks CAs are increasingly using automation tools to handle repetitive tasks such as data entry, reconciliations, and transaction processing. This not only reduces the risk of human error but also frees up CAs to focus on higher-value tasks like strategic planning and analysis. Advanced Data Analytics Data analytics tools enable CAs to extract meaningful insights from large volumes of financial data. These insights can help businesses identify trends, anticipate risks, and make informed decisions to drive growth. Real-Time Financial Reporting Fintech enables CAs to provide clients with real-time financial reporting, giving businesses immediate access to critical information. This enhances transparency and empowers business owners to respond quickly to changing market conditions. Enhancing Audit Efficiency Fintech tools are revolutionizing the audit process. CAs can use AI-powered algorithms to analyze vast amounts of data, detect anomalies, and identify potential instances of fraud more efficiently. Personalized Financial Planning CAs can leverage fintech to offer personalized financial planning services. With access to detailed financial data, CAs can create tailored strategies that align with a client's unique goals and circumstances. Strengthening Cybersecurity As businesses become more reliant on digital tools, cybersecurity becomes paramount. CAs are playing a critical role in advising clients on cybersecurity measures to protect sensitive financial information. Virtual CFO Services Fintech enables CAs to offer virtual CFO services to startups and small businesses. Through digital platforms, CAs can provide expert financial advice and guidance remotely, making their expertise accessible to a wider range of clients. Embracing Blockchain Technology Blockchain's potential for secure and transparent record-keeping is of interest to CAs. They can explore applications in supply chain finance, smart contracts, and even audit trail verification. Continuous Learning in Fintech CAs recognize the importance of staying updated with fintech trends. Many are investing in continuous learning to master the use of new tools and technologies that can optimize their services. Conclusion The integration of fintech into the realm of finance is reshaping the landscape in profound ways. CAs are embracing these technologies to elevate their roles from traditional number-crunchers to strategic advisors, equipped with tools that enhance efficiency, accuracy, and insight. As fintech continues to evolve, CAs will remain pivotal in guiding businesses through the ever-changing financial landscape, leveraging technology to drive growth, innovation, and success. Find the top verified CA in your City Feel free to let me know if you'd like more blogs on different topics or if you have specific requirements for the content.

The Future of AI in eCommerce Marketing: What to Expect 🚀
reddit
LLM Vibe Score0
Human Vibe Score0
McFlyAdsThis week

The Future of AI in eCommerce Marketing: What to Expect 🚀

Hey Reddit community! As we dive deeper into 2025, the integration of AI in eCommerce marketing is becoming more sophisticated and impactful. Here’s a look at where AI is headed and how it's revolutionizing the industry: Personalized Shopping Experiences: AI is enhancing personalization by analyzing consumer behavior and preferences, allowing retailers to offer tailored recommendations and promotions. This not only boosts customer satisfaction but also increases conversion rates. Chatbots and Virtual Assistants: AI-powered chatbots are becoming more intuitive and capable of handling complex queries, providing 24/7 customer support, and improving overall user experience. They’re a game-changer for eCommerce businesses looking to enhance customer engagement. Predictive Analytics: With AI, businesses can leverage predictive analytics to forecast trends, optimize inventory, and refine marketing strategies. This helps in making data-driven decisions that align with consumer demands and market dynamics. Automated Content Creation: AI tools are being used to generate product descriptions, social media posts, and even ad copy. This automation saves time and ensures consistency across marketing channels. Visual and Voice Search: AI is powering visual and voice search capabilities, making it easier for consumers to find products using images or voice commands. This technology is set to transform how users interact with eCommerce platforms. Fraud Detection: AI algorithms are improving fraud detection by analyzing transaction patterns and identifying anomalies. This is crucial for maintaining trust and security in online shopping. As AI continues to evolve, it will undoubtedly reshape the eCommerce landscape, offering new opportunities for innovation and growth. What are your thoughts on the future of AI in eCommerce marketing? Let's discuss!

What to look for in the Best PDF Invoice Parser?
reddit
LLM Vibe Score0
Human Vibe Score1
Finley_dzThis week

What to look for in the Best PDF Invoice Parser?

I've been thinking about starting using PDF Invoice Parser, so these are some key features to look out for in a PDF invoice parser I've learned about these days on Affinda. Machine Learning - There are invoice parsers available that use machine learning algorithms to learn from their mistakes, resulting in them being able to parse many data sources and become more accurate over time. Optical Character Recognition - An OCR invoice parser is one that uses optical character recognition to take images lacking text data and turn them into digital files. Natural Language Processing - This results in more efficient and effective invoice processing that seeks to understand the text and sort invoice fields correctly. Artificial Intelligence - Many parsers struggle to adapt and fail to complete information extraction from nonstandard invoice formats. That’s why you need a parser that leverages document AI to analyze the template and extract structured data no matter what invoice layout is used. Different Types Analysed - For example, you might receive a mailed invoice or Word document. You need a parser that can analyze and get extracted data from any format of the supplier invoice. So, is this enough information and benefits for me to choose this product? I guess so, I've even heard great stuff about it, but I would love to share all of this with you and maybe some of you already had any experience to share with all of us. Have a nice day, guys!

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age
reddit
LLM Vibe Score0
Human Vibe Score1
ExpenectThis week

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age

The Evolution of Financial Technology: How CAs Are Embracing the Digital Age Introduction In an era characterized by rapid technological advancements, the field of finance is undergoing a transformative journey. The emergence of financial technology, or fintech, is reshaping the way businesses manage their finances, and Chartered Accountants (CAs) are at the forefront of this evolution. In this blog post, we'll explore how CAs are embracing fintech and leveraging its potential to enhance financial management, analysis, and advisory services. Fintech's Impact on Financial Services Fintech encompasses a wide range of technologies that leverage data analytics, artificial intelligence, blockchain, and automation to improve financial services. For CAs, this means new tools to streamline processes, enhance decision-making, and offer innovative solutions to clients. Automation of Routine Tasks CAs are increasingly using automation tools to handle repetitive tasks such as data entry, reconciliations, and transaction processing. This not only reduces the risk of human error but also frees up CAs to focus on higher-value tasks like strategic planning and analysis. Advanced Data Analytics Data analytics tools enable CAs to extract meaningful insights from large volumes of financial data. These insights can help businesses identify trends, anticipate risks, and make informed decisions to drive growth. Real-Time Financial Reporting Fintech enables CAs to provide clients with real-time financial reporting, giving businesses immediate access to critical information. This enhances transparency and empowers business owners to respond quickly to changing market conditions. Enhancing Audit Efficiency Fintech tools are revolutionizing the audit process. CAs can use AI-powered algorithms to analyze vast amounts of data, detect anomalies, and identify potential instances of fraud more efficiently. Personalized Financial Planning CAs can leverage fintech to offer personalized financial planning services. With access to detailed financial data, CAs can create tailored strategies that align with a client's unique goals and circumstances. Strengthening Cybersecurity As businesses become more reliant on digital tools, cybersecurity becomes paramount. CAs are playing a critical role in advising clients on cybersecurity measures to protect sensitive financial information. Virtual CFO Services Fintech enables CAs to offer virtual CFO services to startups and small businesses. Through digital platforms, CAs can provide expert financial advice and guidance remotely, making their expertise accessible to a wider range of clients. Embracing Blockchain Technology Blockchain's potential for secure and transparent record-keeping is of interest to CAs. They can explore applications in supply chain finance, smart contracts, and even audit trail verification. Continuous Learning in Fintech CAs recognize the importance of staying updated with fintech trends. Many are investing in continuous learning to master the use of new tools and technologies that can optimize their services. Conclusion The integration of fintech into the realm of finance is reshaping the landscape in profound ways. CAs are embracing these technologies to elevate their roles from traditional number-crunchers to strategic advisors, equipped with tools that enhance efficiency, accuracy, and insight. As fintech continues to evolve, CAs will remain pivotal in guiding businesses through the ever-changing financial landscape, leveraging technology to drive growth, innovation, and success. Find the top verified CA in your City Feel free to let me know if you'd like more blogs on different topics or if you have specific requirements for the content.

ZeroToHeroML: Beginner-Friendly ML & AI Course (Free)
reddit
LLM Vibe Score0
Human Vibe Score0
DizDThis week

ZeroToHeroML: Beginner-Friendly ML & AI Course (Free)

Hey r/learnmachinelearning! A friend of mine, who's been a software developer at Sony for 10 years, recently expressed interest in learning Machine Learning (ML) and Artificial Intelligence (AI). Leveraging my background in ML and neural computation (learned at UCSD) to create a beginner-friendly course guiding him through the basics and into more complex projects. Foundational Concepts: Predicting House Prices (Regression): Master regression techniques to forecast housing prices based on various factors. Iris Flower Species Prediction (Classification): Learn classification algorithms by predicting flower species using the famous Iris dataset. Overcoming Overfitting: Explore methods to prevent models from overfitting and enhance their generalizability. In Progress: Customer Segmentation (Unsupervised Learning): Delve into unsupervised learning to group customers based on purchase history or demographics (valuable for targeted marketing campaigns). Deep Learning for Image Recognition: Implement Convolutional Neural Networks (CNNs) to build models that recognize objects or scenes in images. Natural Language Processing Sentiment Analysis: Analyze the sentiment (positive, negative, or neutral) expressed in text data (e.g., reviews, social media posts) using NLP techniques. Introduction to Reinforcement Learning: Get acquainted with the fundamentals of reinforcement learning by creating an agent that learns to navigate a maze. Want to Learn or Contribute? I thought I'd share ZeroToHeroML here so others who want to learn ML/AI or know someone who does can benefit from this free resource! ​ Fork the repo: https://github.com/DilrajS/ZeroToHeroML Share with others interested in ML/AI! Pull requests welcome (help the community grow!). All help is appriciated! Let's conquer ML/AI together!

Sophomore computer science student, looking at ISLP vs ESL vs mlcourse.ai
reddit
LLM Vibe Score0
Human Vibe Score1
OneTrueDuceThis week

Sophomore computer science student, looking at ISLP vs ESL vs mlcourse.ai

For background, I am currently a computer science sophomore, with intermediate skills in Python and C++. I have taken university courses on data structure and algorithms, calc 1-3, linear algebra, and an introductory stat course (which covered confidence interval, Z and T sample test, and hypothesis testing). I also have read up to Chapter 5 of the MML book and am currently self-studying probability theory (through STAT 110 video and textbook by Joe Blitzstein). I have done a few beginner ML projects with Tensorflow and scikit-learn, but most of the work is in EDA and feature engineering, while the ML model is just a black box that I plug and chug. So now, I want to learn how to implement ML models from scratch. I've been skimming over ISLP, which many people online recommended, but it seems that while it talks about mathematical equations used, I don't really get to implement it; as the labs are a lot of importing an already implemented model then plug and chug. So now, I am looking at ESL, which I believe is the more detailed and mathematically rigorous version of ISL. However, there aren't any labs or code along to ease beginners in (which I somewhat understand given the intended audience of the book). Another option I am looking at is mlcourse.ai, which seems to cover mathematics and has some lab/code along for it. But it doesn't seem to span as many subjects as ESL does. Given these options, I am unsure of which one to pick, should I first finish my self-study on probability theory and then Chapters 6-8 of MML? Then should I do ISLP first or just get into ESL? Or maybe I should do mlcourse.ai first then into ESL? Or should I just do the ML course/book along with the maths? In addition, there is also the data science + feature engineering stuff which I wonder if I should study more about. Sorry if this seems like a mess, there are just so many things to ML that I am kinda overwhelmed.

[Help Needed] How to tackle AI for DNS Security in a Hackathon for Beginner.
reddit
LLM Vibe Score0
Human Vibe Score1
Baby-Boss0506This week

[Help Needed] How to tackle AI for DNS Security in a Hackathon for Beginner.

Hi everyone! I've been selected to participate in an AI and Cybersecurity Hackathon, and the group I'm in focuses on AI for DNS Security. Our goal is to implement AI algorithms to detect anomalies and enhance DNS security. Here’s the catch: I have no prior background in cybersecurity, and I’m also a beginner in applying AI to real-world security problems. I’d really appreciate some guidance from this amazing community on how to approach this challenge. A bit more about the project: Objective: Detect anomalies in DNS traffic (e.g., malicious requests, tunneling, etc.). AI tools: We’re free to choose algorithms, but I’m unsure where to start—supervised vs. unsupervised learning? My skillset: Decent grasp of Python (Pandas, Scikit-learn, etc.) and basic ML concepts. No practical experience in network security or analyzing DNS traffic. What I’m looking for: Datasets: Any recommendations for open-source DNS datasets or synthetic data creation methods? AI methods: Which models work best for anomaly detection in DNS logs? Are there any relevant GitHub projects? Learning resources: Beginner-friendly material on DNS security and the application of AI in this domain. Hackathon tips: How can I make the most of this opportunity and contribute effectively to my team? Bonus question: If you’ve participated in similar hackathons, what strategies helped you balance learning and execution within a short timeframe? Thank you so much in advance for any advice, resources, or personal experiences you can share! I’ll make sure to share our project results and lessons learned after the hackathon.

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.
reddit
LLM Vibe Score0
Human Vibe Score0.714
level6-killjoyThis week

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.

This is a recap covering the major news from last week. 🔥Top 3 news - OpenAI’s updates, Meta’s upcoming free LLM and EU Regulation 🗞️Interesting reads include PSA about protecting your keys, The GPT ouroboros, Reddit - OpenAI’s moat, and more.. 🧑‍🎓Learning includes a Step-by-step guide from a non-technical founder who launched his MVP, Chatbot for your Gdrive and more 🔥Top 3 AI news in the past week OpenAI: New Pricing, Models, & Functions OpenAI has been on a roll. Last week we saw the release of OpenAI best practice on using GPT. This week we saw some amazing updates. Three major buckets were: First, the price decreases for both embeddings and GPT-3.5 tokens. Second, new models for gpt-4 and gpt-3.5. A new longer context model for gpt-3.5. Third, a new function calling capability. Why is it important? Previously, the output from OpenAI was all text. So, calling an external API from GPT was quite difficult. You had to parse the text data and things were often incorrect. Langchain created the Agents and Tools feature to tackle this problem. It was still unreliable and prone to issues. Now you get native support to generate a fixed format output. You can use the output to generate functional calls and also pass functions which need to be called. For example, if your app has multiple API endpoints then you can use GPT to generate the API calls with parameters. You can also pass the endpoints as function calls to ensure the correct function is executed. This functionality can further be used to generate structured data (JSON) out of GPT. So, you can generate data from GPT and load it into your backend. What’s next? This functionality allows turning natural language responses into structured data. This can be used to create “intelligent” backends using LLMs. We might see implementations in no-code tools to allow more robust and natural-language tools for non-technical folks. The structured data process goes both ways. You can also feed structured data into GPT for better responses. This feature also has its share of issues. Function calling suffers from the same prompt injection issues. Malicious actors can pass malicious code in function or the responses. For example, creation of queries using functions might contain malicious code to delete data. Without proper user validation this code will be executed automatically and delete data. So, using LLM as the back-end layer needs proper security implementation. Meta's LLM: Commercial Use Ahead Llama has been a boon for the open source community. Many of the open source models rely on Llama. The issue is that Llama is research-only and cannot be used commercially. So, no one can use it to build any product. Meta is now working on the next version of the model. This model will be available for commercial use. This is in stark contrast to both OpenAI and Google. Both safe-guarde their models and make it available through API. Why is it important? Certain industries cannot use LLM APIs because of strict restrictions on data privacy. These companies would want to run their own instance of a foundational model. A commercially available foundational model is also going to help people who want to keep their “API call” costs next to 0. A commercially available free-for-all model will also help push the open source community further. Just like Llama. What’s next? Sam Altman has said OpenAI didn’t release GPT-3 as open-source because they didn’t think people would be able to run it. Now OpenAI is working on an open-source model. This is going to be weaker than GPT-4. Let the battle of LLMs begin. EU's Proposed Legislation and Its Impact on AI Usage The EU parliament voted to move ahead with the E.U. AI Act. This act aims to ensure consumer protection against the dangers of AI. Why is it important? OpenAI and Sam Altman want regulations for models. They have proposed a IAEA-type of agency to stop the proliferation of LLM models. As per OpenAI, all models should be regulated and monitored. The suggestion of a license based regulation has led to significant backlash. Many people have called it “regulatory capture” - with the aim of shutting down competing LLMs. Licensing based regulations might not really be effective. The EU is approaching regulation from a different angle. It doesn’t focus on how models are developed. Rather focuses on how AI will/can be used. They have broken down use cases into 4 categories - unacceptable (prohibited), high, medium and low risk. For example, Building a Pre-Crime software,on%20crimes%20not%20yet%20committed.) to predict crimes? Building a Social credit system? Unacceptable. Using tools to influence elections or recommendation algorithms? High (Highly regulated). Using generative AI tools to create text or images on news sites? Medium (Add label that the content is AI generated) AI providers also need to disclose their training source. To me this sounds like good legislation. What do you guys think? But, OpenAI has warned that EU regulations might force them to pull out completely. What’s next? The disclosure requirements might help various publishing companies. AI and media companies are in talks to pay for training data. Google has been leading the charge. Additionally, OpenAI and Deepmind will open their models for safety and research purposes to the UK government. 🗞️10 AI news highlights and interesting reads PSA: If you are using Repl to write code, you might want to check your OpenAI API keys. If you have left them embedded then people can pirate and steal the keys. LLMs rely on human annotation or human feedback to learn. And one way to generate human annotation is crowdsourcing. But what if the crowdsource human annotators use LLMs? Research shows 33-46% workers used LLMs. So, basically we go from Human -> AI -> Human -> AI. The AI ouroboros. Researchers also say generated data to train models might cause serious issue. All the talks about moats \- Reddit might be OpenAI’s \future\ moat. Given the amount of complaints about how Google search experience has deteriorated during the blackout, this might be true? Doctors are using ChatGPT but not to diagnose.Rather to be more empathetic. We discussed this just a month ago. And guess where the data for this study came from? Reddit AskDocs. Moat FTW?! Beatles to make a comeback…using Generative AI. SnapFusion - Text to Image diffusion on mobile phones. Large context lengths are important for better GPT experience. The secret sauce for 100k context length. There is a lot of bad AI research out there. Some border on snake oil. Most AI “research” should be double checked and challenged. A new research on huggingface said that GPT-4 can ace MIT curriculum. Now someone is replicating the results and say that GPT-4 can’t beat MIT. Are we seeing peak AI? Especially when people from Deepmind and Meta are involved? Mistral AI raised $113 million in seed round with no product. Some might say this funding is for the team and the team is really solid. The issue though is whether the valuation is justified when OpenAI and Google already have a head start. The AI Hype Wall of Shame. \- Collection of articles which mislead people about AI in various aspects. 🧑‍🎓3 Learning Resources Building and Launching a company using GPT-4 with prompts. (The author didn’t know how to code but created and launched the MVP in a month). Chatbot for your Gdrive - https://www.haihai.ai/gpt-gdrive/ Building ChatGPT plugin using Supabase - https://supabase.com/blog/building-chatgpt-plugins-template That’s it folks. Thank you for reading and have a great week ahead. If you are interested in a focused weekly recap delivered to your inbox on Mondays you can subscribe here. It is FREE!

Join the AI4Earth challenge with the European Space Agency to highlight our footprint on Earth using Earth Observation data and Machine Learning
reddit
LLM Vibe Score0
Human Vibe Score1
campachThis week

Join the AI4Earth challenge with the European Space Agency to highlight our footprint on Earth using Earth Observation data and Machine Learning

​ https://preview.redd.it/ww109cba14f71.png?width=2401&format=png&auto=webp&s=8bd3d43e8b63848af85c73478be61e43d9e10189 The primary goal is to get an insight into the human impact on Earth, to drive and guide conservation efforts of this planet we call home. Our approach will be twofold:  Firstly we will work on AI algorithms that can serve as an early detection system of human impact sites. Secondly we will use these detection systems to find satellite images that show the most impactful human-caused changes, which will be used in the creation of a video to launch an awareness campaign. You will be working with ESA to detect things like: Wildfires and Deforestation Marine Litter and Melting Glaciers Air quality detection & Novel animal migration patterns  and much more!  European Space Agency To reach these goals we’ve partnered up with ESA, who are able to use our algorithms to monitor new satellite data and guide conservation efforts. They will provide us with multi-spectral data of their Sentinel-2 satellite pair and with invaluable knowledge and research on the domain of Earth Observation data in participant only masterclasses.  Format The challenge will run throughout September and October, where you will collaborate with a diverse team of over 30 international data specialists and domain experts in subteams, all tackling this problem from different angles. Subtasks like the detection of deforestation, wildfires, marine litter or any other human caused impact. All contributors in the challenge are expected to spend 12 hours or more per week during the entirity of the two month challenge. To learn more subscribe to the info session on the 3rd of August 19:00 CEST HERE! Some important dates: 3rd of August – Info session 1st of September – Challenge Kick-off 29th of September – Midterm presentations 29th of October – Final presentations PARTNERS SUN - https://spacehubs.network The project is spearheaded by SUN whose goal is to increase the commercialization of space enabled solutions and growth of European start-ups and scale-ups in the space downstream and upstream sectors. ESA - https://esa.int ESA will be the main stakeholder and domain knowledge provider in the challenge. Their efforts to aid human’s space endeavours as well as protect the planet we live on will serve us for many years to come.  MLReef - https://mlreef.com MLReef provides an open source platform for collaborative Machine Learning. They provide the computational infrastructure to support the EO4Earth project as part of their AI4GOOD and Open Science initiatives. Brimatech  As a partner in the SUN project, the innovation management and market research expert Brimatech helps out in the overall organisation of the challenge.  Mothership The ‘Mothership’ is a dedicated open innovation program created by Space4Good and World Startup Factory. The Mothershi is leveraging recent advancements in artificial intelligence and satellite technologies in support of the UN Sustainable Development Goals. Space4Good  Space4Good is a geospatial innovation lab supporting impact makers on the ground with earth observation insights from above. Worldstartup  Worldstartup is a collective of international entrepreneurs, experts, mentors and investors, dedicated to help the best impact-driven startups and scaleups.

Study Plan for Learning Data Science Over the Next 12 Months [D]
reddit
LLM Vibe Score0
Human Vibe Score1
daniel-dataThis week

Study Plan for Learning Data Science Over the Next 12 Months [D]

In this thread, I address a study plan for 2021. In case you're interested, I wrote a whole article about this topic: Study Plan for Learning Data Science Over the Next 12 Months Let me know your thoughts on this. ​ https://preview.redd.it/emg20nzhet661.png?width=1170&format=png&auto=webp&s=cf09e4dc5e82ba2fd7b57c706ba2873be57fe8de We are ending 2020 and it is time to make plans for next year, and one of the most important plans and questions we must ask is what do we want to study?, what do we want to enhance?, what changes do we want to make?, and what is the direction we are going to take (or continue) in our professional careers?. Many of you will be starting on the road to becoming a data scientist, in fact you may be evaluating it, since you have heard a lot about it, but you have some doubts, for example about the amount of job offers that may exist in this area, doubts about the technology itself, and about the path you should follow, considering the wide range of options to learn. I’m a believer that we should learn from various sources, from various mentors, and from various formats. By sources I mean the various virtual platforms and face-to-face options that exist to study. By mentors I mean that it is always a good idea to learn from different points of view and learning from different teachers/mentors, and by formats I mean the choices between books, videos, classes, and other formats where the information is contained. When we extract information from all these sources we reinforce the knowledge learned, but we always need a guide, and this post aims to give you some practical insights and strategies in this regard. To decide on sources, mentors and formats it is up to you to choose. It depends on your preferences and ease of learning: for example, some people are better at learning from books, while others prefer to learn from videos. Some prefer to study on platforms that are practical (following online code), and others prefer traditional platforms: like those at universities (Master’s Degree, PHDs or MOOCs). Others prefer to pay for quality content, while others prefer to look only for free material. That’s why I won’t give a specific recommendation in this post, but I’ll give you the whole picture: a study plan. To start you should consider the time you’ll spend studying and the depth of learning you want to achieve, because if you find yourself without a job you could be available full time to study, which is a huge advantage. On the other hand, if you are working, you’ll have less time and you’ll have to discipline yourself to be able to have the time available in the evenings, mornings or weekends. Ultimately, the important thing is to meet the goal of learning and perhaps dedicating your career to this exciting area! We will divide the year into quarters as follows First Quarter: Learning the Basics Second Quarter: Upgrading the Level: Intermediate Knowledge Third Quarter: A Real World Project — A Full-stack Project Fourth Quarter: Seeking Opportunities While Maintaining Practice First Quarter: Learning the Basics ​ https://preview.redd.it/u7t9bthket661.png?width=998&format=png&auto=webp&s=4ad29cb43618e7acf793259243aa5a60a8535f0a If you want to be more rigorous you can have start and end dates for this period of study of the bases. It could be something like: From January 1 to March 30, 2021 as deadline. During this period you will study the following: A programming language that you can apply to data science: Python or R. We recommend Python due to the simple fact that approximately 80% of data science job offers ask for knowledge in Python. That same percentage is maintained with respect to the real projects you will find implemented in production. And we add the fact that Python is multipurpose, so you won’t “waste” your time if at some point you decide to focus on web development, for example, or desktop development. This would be the first topic to study in the first months of the year. Familiarize yourself with statistics and mathematics. There is a big debate in the data science community about whether we need this foundation or not. I will write a post later on about this, but the reality is that you DO need it, but ONLY the basics (at least in the beginning). And I want to clarify this point before continuing. We could say that data science is divided in two big fields: Research on one side and putting Machine Learning algorithms into production on the other side. If you later decide to focus on Research then you are going to need mathematics and statistics in depth (very in depth). If you are going to go for the practical part, the libraries will help you deal with most of it, under the hood. It should be noted that most job offers are in the practical part. For both cases, and in this first stage you will only need the basics of: Statistics (with Python and NumPy) Descriptive statistics Inferential Statistics Hypothesis testing Probability Mathematics (with Python and NumPy) Linear Algebra (For example: SVD) Multivariate Calculus Calculus (For example: gradient descent) Note: We recommend that you study Python first before seeing statistics and mathematics, because the challenge is to implement these statistical and mathematical bases with Python. Don’t look for theoretical tutorials that show only slides or statistical and/or mathematical examples in Excel/Matlab/Octave/SAS and other different to Python or R, it gets very boring and impractical! You should choose a course, program or book that teaches these concepts in a practical way and using Python. Remember that Python is what we finally use, so you need to choose well. This advice is key so you don’t give up on this part, as it will be the most dense and difficult. If you have these basics in the first three months, you will be ready to make a leap in your learning for the next three months. Second Quarter: Upgrading the Level: Intermediate Knowledge ​ https://preview.redd.it/y1y55vynet661.png?width=669&format=png&auto=webp&s=bd3e12bb112943025c39a8975faf4d64514df275 If you want to be more rigorous you can have start and end dates for this period of study at the intermediate level. It could be something like: From April 1 to June 30, 2021 as deadline. Now that you have a good foundation in programming, statistics and mathematics, it is time to move forward and learn about the great advantages that Python has for applying data analysis. For this stage you will be focused on: Data science Python stack Python has the following libraries that you should study, know and practice at this stage Pandas: for working with tabular data and make in-depth analysis Matplotlib and Seaborn: for data visualization Pandas is the in-facto library for data analysis, it is one of the most important (if not the most important) and powerful tools you should know and master during your career as a data scientist. Pandas will make it much easier for you to manipulate, cleanse and organize your data. Feature Engineering Many times people don’t go deep into Feature Engineering, but if you want to have Machine Learning models that make good predictions and improve your scores, spending some time on this subject is invaluable! Feature engineering is the process of using domain knowledge to extract features from raw data using data mining techniques. These features can be used to improve the performance of machine learning algorithms. Feature engineering can be considered as applied machine learning itself. To achieve the goal of good feature engineering you must know the different techniques that exist, so it is a good idea to at least study the main ones. Basic Models of Machine Learning At the end of this stage you will start with the study of Machine Learning. This is perhaps the most awaited moment! This is where you start to learn about the different algorithms you can use, which particular problems you can solve and how you can apply them in real life. The Python library we recommend you to start experimenting with ML is: scikit-learn. However it is a good idea that you can find tutorials where they explain the implementation of the algorithms (at least the simplest ones) from scratch with Python, since the library could be a “Black Box” and you might not understand what is happening under the hood. If you learn how to implement them with Python, you can have a more solid foundation. If you implement the algorithms with Python (without a library), you will put into practice everything seen in the statistics, mathematics and Pandas part. These are some recommendations of the algorithms that you should at least know in this initial stage Supervised learning Simple Linear Regression Multiple Linear Regression K-nearest neighbors (KNN) Logistic Regression Decision Trees Random Forest Unsupervised Learning K-Means PCA Bonus: if you have the time and you are within the time ranges, you can study these others Gradient Boosting Algorithms GBM XGBoost LightGBM CatBoost Note: do not spend more than the 3 months stipulated for this stage. Because you will be falling behind and not complying with the study plan. We all have shortcomings at this stage, it is normal, go ahead and then you can resume some concepts that did not understand in detail. The important thing is to have the basic knowledge and move forward! If at least you succeed to study the mentioned algorithms of supervised and unsupervised learning, you will have a very clear idea of what you will be able to do in the future. So don’t worry about covering everything, remember that it is a process, and ideally you should have some clearly established times so that you don’t get frustrated and feel you are advancing. So far, here comes your “theoretical” study of the basics of data science. Now we’ll continue with the practical part! Third Quarter: A Real World Project — A Full-stack Project ​ https://preview.redd.it/vrn783vqet661.png?width=678&format=png&auto=webp&s=664061b3d33b34979b74b10b9f8a3d0f7b8b99ee If you want to be more rigorous you can have start and end dates for this period of study at the intermediate level. It could be something like: From July 1 to September 30, 2021 as deadline. Now that you have a good foundation in programming, statistics, mathematics, data analysis and machine learning algorithms, it is time to move forward and put into practice all this knowledge. Many of these suggestions may sound out of the box, but believe me they will make a big difference in your career as a data scientist. The first thing is to create your web presence: Create a Github (or GitLab) account, and learn Git*. Being able to manage different versions of your code is important, you should have version control over them, not to mention that having an active Github account is very valuable in demonstrating your true skills. On Github, you can also set up your Jupyter Notebooks and make them public, so you can show off your skills as well. This is mine for example: https://github.com/danielmoralesp Learn the basics of web programming*. The advantage is that you already have Python as a skill, so you can learn Flask to create a simple web page. Or you can use a template engine like Github Pages, Ghost or Wordpress itself and create your online portfolio. Buy a domain with your name*. Something like myname.com, myname.co, myname.dev, etc. This is invaluable so you can have your CV online and update it with your projects. There you can make a big difference, showing your projects, your Jupyter Notebooks and showing that you have the practical skills to execute projects in this area. There are many front-end templates for you to purchase for free or for payment, and give it a more personalized and pleasant look. Don’t use free sub-domains of Wordpress, Github or Wix, it looks very unprofessional, make your own. Here is mine for example: https://www.danielmorales.dev/ Choose a project you are passionate about and create a Machine Learning model around it. The final goal of this third quarter is to create ONE project, that you are passionate about, and that is UNIQUE among others. It turns out that there are many typical projects in the community, such as predicting the Titanic Survivors, or predicting the price of Houses in Boston. Those kinds of projects are good for learning, but not for showing off as your UNIQUE projects. If you are passionate about sports, try predicting the soccer results of your local league. If you are passionate about finance, try predicting your country’s stock market prices. If you are passionate about marketing, try to find someone who has an e-commerce and implement a product recommendation algorithm and upload it to production. If you are passionate about business: make a predictor of the best business ideas for 2021 :) As you can see, you are limited by your passions and your imagination. In fact, those are the two keys for you to do this project: Passion and Imagination. However don’t expect to make money from it, you are in a learning stage, you need that algorithm to be deployed in production, make an API in Flask with it, and explain in your website how you did it and how people can access it. This is the moment to shine, and at the same time it’s the moment of the greatest learning. You will most likely face obstacles, if your algorithm gives 60% of Accuracy after a huge optimization effort, it doesn’t matter, finish the whole process, deploy it to production, try to get a friend or family member to use it, and that will be the goal achieved for this stage: Make a Full-stack Machine Learning project. By full-stack I mean that you did all the following steps: You got the data from somewhere (scrapping, open data or API) You did a data analysis You cleaned and transformed the data You created Machine Learning Models You deployed the best model to production for other people to use. This does not mean that this whole process is what you will always do in your daily job, but it does mean that you will know every part of the pipeline that is needed for a data science project for a company. You will have a unique perspective! Fourth Quarter: Seeking Opportunities While Maintaining Practice ​ https://preview.redd.it/qd0osystet661.png?width=1056&format=png&auto=webp&s=2da456b15985b2793041256f5e45bca99a23b51a If you want to be more rigorous you can have start and end dates for this period of study at the final level. It could be something like: From October 1 to December 31, 2021 as deadline. Now you have theoretical and practical knowledge. You have implemented a model in production. The next step depends on you and your personality. Let’s say you are an entrepreneur, and you have the vision to create something new from something you discovered or saw an opportunity to do business with this discipline, so it’s time to start planning how to do it. If that’s the case, obviously this post won’t cover that process, but you should know what the steps might be (or start figuring them out). But if you are one of those who want to get a job as a data scientist, here is my advice. Getting a job as a data scientist “You’re not going to get a job as fast as you think, if you keep thinking the same way”.Author It turns out that all people who start out as data scientists imagine themselves working for the big companies in their country or region. Or even remote. It turns out that if you aspire to work for a large company like data scientist you will be frustrated by the years of experience they ask for (3 or more years) and the skills they request. Large companies don’t hire Juniors (or very few do), precisely because they are already large companies. They have the financial muscle to demand experience and skills and can pay a commensurate salary (although this is not always the case). The point is that if you focus there you’re going to get frustrated! Here we must return to the following advise: “You need creativity to get a job in data science”. Like everything else in life we have to start at different steps, in this case, from the beginning. Here are the scenarios If you are working in a company and in a non-engineering role you must demonstrate your new skills to the company you are working for*. If you are working in the customer service area, you should apply it to your work, and do for example, detailed analysis of your calls, conversion rates, store data and make predictions about it! If you can have data from your colleagues, you could try to predict their sales! This may sound funny, but it’s about how creatively you can apply data science to your current work and how to show your bosses how valuable it is and EVANGELIZE them about the benefits of implementation. You’ll be noticed and they could certainly create a new data related department or job. And you already have the knowledge and experience. The key word here is Evangelize. Many companies and entrepreneurs are just beginning to see the power of this discipline, and it is your task to nurture that reality. If you are working in an area related to engineering, but that is not data science*. Here the same applies as the previous example, but you have some advantages, and that is that you could access the company’s data, and you could use it for the benefit of the company, making analyses and/or predictions about it, and again EVANGELIZING your bosses your new skills and the benefits of data science. If you are unemployed (or do not want, or do not feel comfortable following the two examples above)*, you can start looking outside, and what I recommend is that you look for technology companies and / or startups where they are just forming the first teams and are paying some salary, or even have options shares of the company. Obviously here the salaries will not be exorbitant, and the working hours could be longer, but remember that you are in the learning and practice stage (just in the first step), so you can not demand too much, you must land your expectations and fit that reality, and stop pretending to be paid $ 10,000 a month at this stage. But, depending of your country $1.000 USD could be something very interesting to start this new career. Remember, you are a Junior at this stage. The conclusion is: don’t waste your time looking at and/or applying to offers from big companies, because you will get frustrated. Be creative, and look for opportunities in smaller or newly created companies. Learning never stops While you are in that process of looking for a job or an opportunity, which could take half of your time (50% looking for opportunities, 50% staying in practice), you have to keep learning, you should advance to concepts such as Deep Learning, Data Engineer or other topics that you feel were left loose from the past stages or focus on the topics that you are passionate about within this group of disciplines in data science. At the same time you can choose a second project, and spend some time running it from end-to-end, and thus increase your portfolio and your experience. If this is the case, try to find a completely different project: if the first one was done with Machine Learning, let this second one be done with Deep learning. If the first one was deployed to a web page, that this second one is deployed to a mobile platform. Remember, creativity is the key! Conclusion We are at an ideal time to plan for 2021, and if this is the path you want to take, start looking for the platforms and media you want to study on. Get to work and don’t miss this opportunity to become a data scientist in 2021! Note: we are building a private community in Slack of data scientist, if you want to join us write to the email: support@datasource.ai I hope you enjoyed this reading! you can follow me on twitter or linkedin Thank you for reading!

MMML | Deploy HuggingFace training model rapidly based on MetaSpore
reddit
LLM Vibe Score0
Human Vibe Score1
qazmkoppThis week

MMML | Deploy HuggingFace training model rapidly based on MetaSpore

A few days ago, HuggingFace announced a $100 million Series C funding round, which was big news in open source machine learning and could be a sign of where the industry is headed. Two days before the HuggingFace funding announcement, open-source machine learning platform MetaSpore released a demo based on the HuggingFace Rapid deployment pre-training model. As deep learning technology makes innovative breakthroughs in computer vision, natural language processing, speech understanding, and other fields, more and more unstructured data are perceived, understood, and processed by machines. These advances are mainly due to the powerful learning ability of deep learning. Through pre-training of deep models on massive data, the models can capture the internal data patterns, thus helping many downstream tasks. With the industry and academia investing more and more energy in the research of pre-training technology, the distribution warehouses of pre-training models such as HuggingFace and Timm have emerged one after another. The open-source community release pre-training significant model dividends at an unprecedented speed. In recent years, the data form of machine modeling and understanding has gradually evolved from single-mode to multi-mode, and the semantic gap between different modes is being eliminated, making it possible to retrieve data across modes. Take CLIP, OpenAI’s open-source work, as an example, to pre-train the twin towers of images and texts on a dataset of 400 million pictures and texts and connect the semantics between pictures and texts. Many researchers in the academic world have been solving multimodal problems such as image generation and retrieval based on this technology. Although the frontier technology through the semantic gap between modal data, there is still a heavy and complicated model tuning, offline data processing, high performance online reasoning architecture design, heterogeneous computing, and online algorithm be born multiple processes and challenges, hindering the frontier multimodal retrieval technologies fall to the ground and pratt &whitney. DMetaSoul aims at the above technical pain points, abstracting and uniting many links such as model training optimization, online reasoning, and algorithm experiment, forming a set of solutions that can quickly apply offline pre-training model to online. This paper will introduce how to use the HuggingFace community pre-training model to conduct online reasoning and algorithm experiments based on MetaSpore technology ecology so that the benefits of the pre-training model can be fully released to the specific business or industry and small and medium-sized enterprises. And we will give the text search text and text search graph two multimodal retrieval demonstration examples for your reference. Multimodal semantic retrieval The sample architecture of multimodal retrieval is as follows: Our multimodal retrieval system supports both text search and text search application scenarios, including offline processing, model reasoning, online services, and other core modules: https://preview.redd.it/mdyyv1qmdz291.png?width=1834&format=png&auto=webp&s=e9e10710794c78c64cc05adb75db385aa53aba40 Offline processing, including offline data processing processes for different application scenarios of text search and text search, including model tuning, model export, data index database construction, data push, etc. Model inference. After the offline model training, we deployed our NLP and CV large models based on the MetaSpore Serving framework. MetaSpore Serving helps us conveniently perform online inference, elastic scheduling, load balancing, and resource scheduling in heterogeneous environments. Online services. Based on MetaSpore’s online algorithm application framework, MetaSpore has a complete set of reusable online search services, including Front-end retrieval UI, multimodal data preprocessing, vector recall and sorting algorithm, AB experimental framework, etc. MetaSpore also supports text search by text and image scene search by text and can be migrated to other application scenarios at a low cost. The HuggingFace open source community has provided several excellent baseline models for similar multimodal retrieval problems, which are often the starting point for actual optimization in the industry. MetaSpore also uses the pre-training model of the HuggingFace community in its online services of searching words by words and images by words. Searching words by words is based on the semantic similarity model of the question and answer field optimized by MetaSpore, and searching images by words is based on the community pre-training model. These community open source pre-training models are exported to the general ONNX format and loaded into MetaSpore Serving for online reasoning. The following sections will provide a detailed description of the model export and online retrieval algorithm services. The reasoning part of the model is standardized SAAS services with low coupling with the business. Interested readers can refer to my previous post: The design concept of MetaSpore, a new generation of the one-stop machine learning platform. 1.1 Offline Processing Offline processing mainly involves the export and loading of online models and index building and pushing of the document library. You can follow the step-by-step instructions below to complete the offline processing of text search and image search and see how the offline pre-training model achieves reasoning at MetaSpore. 1.1.1 Search text by text Traditional text retrieval systems are based on literal matching algorithms such as BM25. Due to users’ diverse query words, a semantic gap between query words and documents is often encountered. For example, users misspell “iPhone” as “Phone,” and search terms are incredibly long, such as “1 \~ 3 months old baby autumn small size bag pants”. Traditional text retrieval systems will use spelling correction, synonym expansion, search terms rewriting, and other means to alleviate the semantic gap but fundamentally fail to solve this problem. Only when the retrieval system fully understands users’ query terms and documents can it meet users’ retrieval demands at the semantic level. With the continuous progress of pre-training and representational learning technology, some commercial search engines continue to integrate semantic vector retrieval methods based on symbolic learning into the retrieval ecology. Semantic retrieval model This paper introduces a set of semantic vector retrieval applications. MetaSpore built a set of semantic retrieval systems based on encyclopedia question and answer data. MetaSpore adopted the Sentence-Bert model as the semantic vector representation model, which fine-tunes the twin tower BERT in supervised or unsupervised ways to make the model more suitable for retrieval tasks. The model structure is as follows: The query-Doc symmetric two-tower model is used in text search and question and answer retrieval. The vector representation of online Query and offline DOC share the same vector representation model, so it is necessary to ensure the consistency of the offline DOC library building model and online Query inference model. The case uses MetaSpore’s text representation model Sbert-Chinese-QMC-domain-V1, optimized in the open-source semantically similar data set. This model will express the question and answer data as a vector in offline database construction. The user query will be expressed as a vector by this model in online retrieval, ensuring that query-doc in the same semantic space, users’ semantic retrieval demands can be guaranteed by vector similarity metric calculation. Since the text presentation model does vector encoding for Query online, we need to export the model for use by the online service. Go to the q&A data library code directory and export the model concerning the documentation. In the script, Pytorch Tracing is used to export the model. The models are exported to the “./export “directory. The exported models are mainly ONNX models used for wired reasoning, Tokenizer, and related configuration files. The exported models are loaded into MetaSpore Serving by the online Serving system described below for model reasoning. Since the exported model will be copied to the cloud storage, you need to configure related variables in env.sh. \Build library based on text search \ The retrieval database is built on the million-level encyclopedia question and answer data set. According to the description document, you need to download the data and complete the database construction. The question and answer data will be coded as a vector by the offline model, and then the database construction data will be pushed to the service component. The whole process of database construction is described as follows: Preprocessing, converting the original data into a more general JSonline format for database construction; Build index, use the same model as online “sbert-Chinese-qmc-domain-v1” to index documents (one document object per line); Push inverted (vector) and forward (document field) data to each component server. The following is an example of the database data format. After offline database construction is completed, various data are pushed to corresponding service components, such as Milvus storing vector representation of documents and MongoDB storing summary information of documents. Online retrieval algorithm services will use these service components to obtain relevant data. 1.1.2 Search by text Text and images are easy for humans to relate semantically but difficult for machines. First of all, from the perspective of data form, the text is the discrete ID type of one-dimensional data based on words and words. At the same time, images are continuous two-dimensional or three-dimensional data. Secondly, the text is a subjective creation of human beings, and its expressive ability is vibrant, including various turning points, metaphors, and other expressions, while images are machine representations of the objective world. In short, bridging the semantic gap between text and image data is much more complex than searching text by text. The traditional text search image retrieval technology generally relies on the external text description data of the image or the nearest neighbor retrieval technology and carries out the retrieval through the image associated text, which in essence degrades the problem to text search. However, it will also face many issues, such as obtaining the associated text of pictures and whether the accuracy of text search by text is high enough. The depth model has gradually evolved from single-mode to multi-mode in recent years. Taking the open-source project of OpenAI, CLIP, as an example, train the model through the massive image and text data of the Internet and map the text and image data into the same semantic space, making it possible to implement the text and image search technology based on semantic vector. CLIP graphic model The text search pictures introduced in this paper are implemented based on semantic vector retrieval, and the CLIP pre-training model is used as the two-tower retrieval architecture. Because the CLIP model has trained the semantic alignment of the twin towers’ text and image side models on the massive graphic and text data, it is particularly suitable for the text search graph scene. Due to the different image and text data forms, the Query-Doc asymmetric twin towers model is used for text search image retrieval. The image-side model of the twin towers is used for offline database construction, and the text-side model is used for the online return. In the final online retrieval, the database data of the image side model will be searched after the text side model encodes Query, and the CLIP pre-training model guarantees the semantic correlation between images and texts. The model can draw the graphic pairs closer in vector space by pre-training on a large amount of visual data. Here we need to export the text-side model for online MetaSpore Serving inference. Since the retrieval scene is based on Chinese, the CLIP model supporting Chinese understanding is selected. The exported content includes the ONNX model used for online reasoning and Tokenizer, similar to the text search. MetaSpore Serving can load model reasoning through the exported content. Build library on Image search You need to download the Unsplash Lite library data and complete the construction according to the instructions. The whole process of database construction is described as follows: Preprocessing, specify the image directory, and then generate a more general JSOnline file for library construction; Build index, use OpenAI/Clip-Vit-BASE-Patch32 pre-training model to index the gallery, and output one document object for each line of index data; Push inverted (vector) and forward (document field) data to each component server. Like text search, after offline database construction, relevant data will be pushed to service components, called by online retrieval algorithm services to obtain relevant data. 1.2 Online Services The overall online service architecture diagram is as follows: ​ https://preview.redd.it/nz8zrbbpdz291.png?width=1280&format=png&auto=webp&s=28dae7e031621bc8819519667ed03d8d085d8ace Multi-mode search online service system supports application scenarios such as text search and text search. The whole online service consists of the following parts: Query preprocessing service: encapsulate preprocessing logic (including text/image, etc.) of pre-training model, and provide services through gRPC interface; Retrieval algorithm service: the whole algorithm processing link includes AB experiment tangent flow configuration, MetaSpore Serving call, vector recall, sorting, document summary, etc.; User entry service: provides a Web UI interface for users to debug and track down problems in the retrieval service. From a user request perspective, these services form invocation dependencies from back to front, so to build up a multimodal sample, you need to run each service from front to back first. Before doing this, remember to export the offline model, put it online and build the library first. This article will introduce the various parts of the online service system and make the whole service system step by step according to the following guidance. See the ReadME at the end of this article for more details. 1.2.1 Query preprocessing service Deep learning models tend to be based on tensors, but NLP/CV models often have a preprocessing part that translates raw text and images into tensors that deep learning models can accept. For example, NLP class models often have a pre-tokenizer to transform text data of string type into discrete tensor data. CV class models also have similar processing logic to complete the cropping, scaling, transformation, and other processing of input images through preprocessing. On the one hand, considering that this part of preprocessing logic is decoupled from tensor reasoning of the depth model, on the other hand, the reason of the depth model has an independent technical system based on ONNX, so MetaSpore disassembled this part of preprocessing logic. NLP pretreatment Tokenizer has been integrated into the Query pretreatment service. MetaSpore dismantlement with a relatively general convention. Users only need to provide preprocessing logic files to realize the loading and prediction interface and export the necessary data and configuration files loaded into the preprocessing service. Subsequent CV preprocessing logic will also be integrated in this manner. The preprocessing service currently provides the gRPC interface invocation externally and is dependent on the Query preprocessing (QP) module in the retrieval algorithm service. After the user request reaches the retrieval algorithm service, it will be forwarded to the service to complete the data preprocessing and continue the subsequent processing. The ReadMe provides details on how the preprocessing service is started, how the preprocessing model exported offline to cloud storage enters the service, and how to debug the service. To further improve the efficiency and stability of model reasoning, MetaSpore Serving implements a Python preprocessing submodule. So MetaSpore can provide gRPC services through user-specified preprocessor.py, complete Tokenizer or CV-related preprocessing in NLP, and translate requests into a Tensor that deep models can handle. Finally, the model inference is carried out by MetaSpore, Serving subsequent sub-modules. Presented here on the lot code: https://github.com/meta-soul/MetaSpore/compare/add\python\preprocessor 1.2.2 Retrieval algorithm services Retrieval algorithm service is the core of the whole online service system, which is responsible for the triage of experiments, the assembly of algorithm chains such as preprocessing, recall, sorting, and the invocation of dependent component services. The whole retrieval algorithm service is developed based on the Java Spring framework and supports multi-mode retrieval scenarios of text search and text search graph. Due to good internal abstraction and modular design, it has high flexibility and can be migrated to similar application scenarios at a low cost. Here’s a quick guide to configuring the environment to set up the retrieval algorithm service. See ReadME for more details: Install dependent components. Use Maven to install the online-Serving component Search for service configurations. Copy the template configuration file and replace the MongoDB, Milvus, and other configurations based on the development/production environment. Install and configure Consul. Consul allows you to synchronize the search service configuration in real-time, including cutting the flow of experiments, recall parameters, and sorting parameters. The project’s configuration file shows the current configuration parameters of text search and text search. The parameter modelName in the stage of pretreatment and recall is the corresponding model exported in offline processing. Start the service. Once the above configuration is complete, the retrieval service can be started from the entry script. Once the service is started, you can test it! For example, for a user with userId=10 who wants to query “How to renew ID card,” access the text search service. 1.2.3 User Entry Service Considering that the retrieval algorithm service is in the form of the API interface, it is difficult to locate and trace the problem, especially for the text search image scene can intuitively display the retrieval results to facilitate the iterative optimization of the retrieval algorithm. This paper provides a lightweight Web UI interface for text search and image search, a search input box, and results in a display page for users. Developed by Flask, the service can be easily integrated with other retrieval applications. The service calls the retrieval algorithm service and displays the returned results on the page. It’s also easy to install and start the service. Once you’re done, go to http://127.0.0.1:8090 to see if the search UI service is working correctly. See the ReadME at the end of this article for details. Multimodal system demonstration The multimodal retrieval service can be started when offline processing and online service environment configuration have been completed following the above instructions. Examples of textual searches are shown below. Enter the entry of the text search map application, enter “cat” first, and you can see that the first three digits of the returned result are cats: https://preview.redd.it/d7syq47rdz291.png?width=1280&format=png&auto=webp&s=b43df9abd380b7d9a52e3045dd787f4feeb69635 If you add a color constraint to “cat” to retrieve “black cat,” you can see that it does return a black cat: ​ https://preview.redd.it/aa7pxx8tdz291.png?width=1280&format=png&auto=webp&s=e3727c29d1bde6eea2e1cccf6c46d3cae3f4750e Further, strengthen the constraint on the search term, change it to “black cat on the bed,” and return results containing pictures of a black cat climbing on the bed: ​ https://preview.redd.it/2mw4qpjudz291.png?width=1280&format=png&auto=webp&s=1cf1db667892b9b3a40451993680fbd6980b5520 The cat can still be found through the text search system after the color and scene modification in the above example. Conclusion The cutting-edge pre-training technology can bridge the semantic gap between different modes, and the HuggingFace community can greatly reduce the cost for developers to use the pre-training model. Combined with the technological ecology of MetaSpore online reasoning and online microservices provided by DMetaSpore, the pre-training model is no longer mere offline dabbling. Instead, it can truly achieve end-to-end implementation from cutting-edge technology to industrial scenarios, fully releasing the dividends of the pre-training large model. In the future, DMetaSoul will continue to improve and optimize the MetaSpore technology ecosystem: More automated and wider access to HuggingFace community ecology. MetaSpore will soon release a common model rollout mechanism to make HuggingFace ecologically accessible and will later integrate preprocessing services into online services. Multi-mode retrieval offline algorithm optimization. For multimodal retrieval scenarios, MetaSpore will continuously iteratively optimize offline algorithm components, including text recall/sort model, graphic recall/sort model, etc., to improve the accuracy and efficiency of the retrieval algorithm. For related code and reference documentation in this article, please visit: https://github.com/meta-soul/MetaSpore/tree/main/demo/multimodal/online Some images source: https://github.com/openai/CLIP/raw/main/CLIP.png https://www.sbert.net/examples/training/sts/README.html

6 principles to data architecture that facilitate innovation
reddit
LLM Vibe Score0
Human Vibe Score1
Competitive_Speech36This week

6 principles to data architecture that facilitate innovation

My team and I have been re-building our company's data architecture. In the process of doing so, I got together six key principles to transforming data architectures and thought I would share them, as a strong data architecture is crucial for businesses looking to stay competitive in the digital landscape, as it improves decision-making, time to market, and data security. When executed with efficiency, a resilient data architecture unleashes unparalleled degrees of agility. Principle 1: Agility and flexibility To quickly adjust to market fluctuations, businesses must create adaptable data infrastructures that can effortlessly manage an ever-growing influx of data. To accomplish this objective, we recommend to our clients to implement Enterprise Service Bus, Enterprise Data Warehouse, and Master Data Management integrated together. ​ I believe the best option is this: \- By centralizing communication, ESB reduces the time and effort required to integrate new systems; \- EDW consolidates data from different sources, resulting in a 50% reduction in software implementation time; \- Finally, MDM ensures consistency and accuracy across the organization, leading to better decision-making and streamlined operations. Implementing these solutions can lead to reduced software implementation time, better ROI, and more manageable data architecture. By fostering a culture of collaboration and adopting modern technologies and practices, businesses can prioritize agility and flexibility in their data architecture to increase the pace of innovation. Principle 2: Modularity and reusability Data architecture that fosters modularity and reusability is essential for accelerating innovation within an organization. By breaking data architecture components into smaller, more manageable pieces, businesses can enable different teams to leverage existing architecture components, reducing redundancy and improving overall efficiency. MDM can promote modularity and reusability by creating a central repository for critical business data. This prevents duplication and errors, improving efficiency and decision-making. MDM enables a single source of truth for data, accessible across multiple systems, which promotes integration and scalability. MDM also provides standardized data models, rules, and governance policies that reduce development time, increase quality, and ensure proper management throughout the data’s lifecycle. Another way to achieve modularity in data architecture is through the use of microservices and scripts for Extract, Transform, and Load (ETL) processes. Adopting a structured methodology and framework can ensure these components are well-organized, making it easier for teams to collaborate and maintain the system. Microservices can also contribute to modularity and reusability in data architecture. These small, independent components can be developed, deployed, and scaled independently of one another. By utilizing microservices, organizations can update or replace individual components without affecting the entire system, improving flexibility and adaptability. Principle 3: Data quality and consistency The efficiency of operations depends on data’s quality, so a meticulously crafted data architecture plays a pivotal role in preserving it, empowering enterprises to make well-informed decisions based on credible information. Here are some key factors to consider that will help your company ensure quality: \- Implementing Master Data Management (MDM) – this way, by consolidating, cleansing, and standardizing data from multiple sources, your IT department will be able to create a single, unified view of the most important data entities (customers, products, and suppliers); \- Assigning data stewardship responsibilities to a small team or an individual specialist; \- Considering implementing data validation, data lineage, and data quality metrics; \- By implementing MDM and adopting a minimal data stewardship approach, organizations can maintain high-quality data that drives innovation and growth. Principle 4: Data governance Data governance is a strategic framework that goes beyond ensuring data quality and consistency. It includes ensuring data security, privacy, accessibility, regulatory compliance, and lifecycle management. Here are some key aspects of data governance: \- Implementing robust measures and controls to protect sensitive data from unauthorized access, breaches, and theft. This is only possible through including encryption, access controls, and intrusion detection systems into your company’s IT architecture; \- Adhering to data privacy regulations and guidelines, such as the General Data Protection Regulation (GDPR) and the California Consumer Privacy Act (CCPA); \- Defining stringent conditions for who has access to specific data assets to maintain control over data and ensure its accessibility only for legitimate purposes. Managing the entire lifecycle of data, from creation and storage to archiving and disposal, including defining policies for data retention, archiving, and deletion in compliance with legal and regulatory requirements. To facilitate effective data governance, organizations can leverage various tools and technologies, such as: \- Data cataloging tools: Solutions like Collibra, Alation, or Informatica Enterprise Data Catalog help organizations discover, understand, and manage their data assets. \- Data lineage tools: Tools like Talend, IBM InfoSphere, or Apache Atlas help track data’s origin, transformation, and usage, providing insights into data quality issues and potential areas for improvement. \- Data quality tools: Solutions like Informatica Data Quality, Trifacta, or SAS Data Quality help organizations maintain high-quality data by identifying and correcting errors, inconsistencies, and inaccuracies. \- Data security and privacy tools: Tools like Varonis, BigID, or Spirion help protect sensitive data and ensure compliance with data privacy regulations. Principle 5: Cloud-first approach A cloud-first approach prioritizes cloud-based solutions over on-premises ones when it comes to data management. Cloud-based data management pros: \- Virtually limitless scalability, so that organizations can grow and adapt to changing data requirements without significant infrastructure investments; \- The pay-as-you-go model of cloud services reduces maintenance costs usually associated with the on-premise choice; \- Greater flexibility for deploying and integrating new technologies and services; \- Cloud can be accessed from anywhere, at any time, turning team collaboration and remote work into a breeze; \- Built-in backup and disaster recovery capabilities, ensuring data safety and minimizing downtime in case of emergencies. Cloud-based data management cons: \- Cloud-first approach raises many data security, privacy, and compliance concerns; \- Transferring large data volumes to and from cloud is often time-consuming and results in increased latency for certain apps; \- Relying on a single cloud provider makes it difficult to switch them or move back to the on-premises option without significant funds and effort. Challenges that organizations that choose a cloud-first approach face: \- Integrating cloud-based systems with on-premises ones can be complex and time-consuming; \- Ensuring data governance and compliance in a multi-cloud or hybrid environment is also another problem reported by my clients. How EDW, ESB, and MDM promote cloud-first approach: A cloud-based EDW centralizes data from multiple sources, enabling a unified view of the organization’s data and simplifying data integration across cloud and on-premises systems. An ESB facilitates communication between disparate cloud and on-premises systems, streamlining data integration and promoting a modular architecture. Cloud-based MDM solutions are used for maintaining data quality and consistency across multiple data sources and environments. Principle 6: Automation and artificial intelligence Incorporating automation tools and AI technologies into data architecture can optimize processes and decision-making. Key Applications: \- Data ingestion and integration: Automation simplifies data schema updates and identifies data quality issues, while AI-assisted development helps create tailored connectors, scripts, and microservices. \- Data quality management: Machine learning algorithms improve data quality and consistency by automatically detecting and correcting inconsistencies and duplicates. \- Predictive analytics: AI and machine learning models analyze historical data to predict trends, identify opportunities, and uncover hidden patterns for better-informed decisions. How No-Code Tools and AI-Assisted Development Work: Business users define data requirements and workflows using no-code tools, enabling AI models to understand their needs. AI models process the information, generating recommendations for connector creation, ETL scripts, and microservices. Developers use AI-generated suggestions to accelerate development and tailor solutions to business needs. By combining automation, AI technologies, and no-code tools, organizations can streamline data architecture processes and bridge the gap between business users and developers, ultimately accelerating innovation. I share more tips on building an agile data architectures in my blog.

Seeking Guidance to Transition from SRE to Quant Developer/HFT
reddit
LLM Vibe Score0
Human Vibe Score1
Horror_Brief5524This week

Seeking Guidance to Transition from SRE to Quant Developer/HFT

Hey Reddit! Let me introduce myself: I’m a 2024 B.Tech (IT) graduate from a state-level college in India. Currently, I’m working as an SRE (Site Reliability Engineer) at one of the world’s largest custodian banks (not naming the company). I’ve been with the organization for about a year now, combining my internship and full-time role. During this time, I’ve developed a strong foundation in AI/ML and have primarily been working on projects to automate BAU (Business-As-Usual) activities using these technologies. Outside of work, I have a deep interest in trading and stock markets—something that runs in the family, as my father has been trading for over 40 years. My ultimate goal is to transition into the field of quantitative development or high-frequency trading (HFT) and work for top-tier firms like HRT, Optiver, or Tower Research. To make this dream a reality, I’ve started preparing in the following ways: Learning the basics of stock markets and financial instruments. Studying statistics and experimenting with different algorithms to analyze stocks (mainly on a fundamental level for now). Getting back to practicing DSA (Data Structures and Algorithms). I’ll admit I’m rusty here since I haven’t been consistent with it for a while. I’m aware that transitioning from SRE to Quant Developer/HFT is ambitious and might seem like a pipe dream, but it’s a challenge I’m willing to take on. I’d love to hear from people in the quant field or those who’ve made similar career transitions. What should I focus on? Any recommended resources, books, or courses? Are there any specific skill sets or certifications that could make me stand out? Any guidance, advice, or encouragement would mean a lot. Thank you!

I created leadsnavi that helps small businesses find quality leads without breaking the bank
reddit
LLM Vibe Score0
Human Vibe Score1
BrightCook5861This week

I created leadsnavi that helps small businesses find quality leads without breaking the bank

Hey Redditors, I’m excited to share LeadsNavi, a tool I built specifically to help small businesses and B2B professionals automatically generate leads and reach potential customers in a smarter way. After talking to a lot of small business owners, I realized how tough it is to juggle lead generation with limited resources. So, I decided to create a tool that could simplify the process and make it more accessible to those who don’t have the budget to invest in expensive solutions. What Exactly Is LeadsNavi? LeadsNavi is an intuitive, cost-effective platform that automates the process of lead generation. It's designed to make it easy for small businesses and entrepreneurs to identify quality leads and grow their customer base without the need for manual prospecting. Here’s what makes it stand out: Automatic Lead Tracking: Tracks visitors to your website and matches them with company data, so you get real insights into who’s interested in your business. AI-Powered Lead Recommendations: Based on your website’s traffic, LeadsNavi uses AI to suggest similar companies that could be interested in your product or service, helping you find new leads faster and more accurately. Affordable & Scalable: For only $49/month, you can use a highly effective tool that scales with your business. It’s designed to be affordable even for small businesses. CRM Integration: Connect your CRM to directly import leads and sync your outreach efforts. How Does It Work? LeadsNavi uses advanced algorithms to track website visitors' IP addresses and match them with a comprehensive business database. It provides details like company names, contact information, and helps you identify potential leads for follow-up. The best part? It works automatically, saving you hours of manual work and effort. Lead Identification: Get insights into which companies are visiting your website. AI-Driven Lead Recommendations: The AI analyzes your site’s traffic and suggests other companies in the same industry or with similar needs that might be a great fit for your product or service. Data-Enriched Leads: Gather real-time, actionable data on these leads to make your outreach more targeted. Easy Setup: Simply integrate with your website and CRM to start getting quality leads in minutes. Who’s It For? Small Businesses: You don’t have to be a marketing expert to generate quality leads. B2B Sales Teams: Perfect for anyone looking to target other businesses with a streamlined and automated approach. Entrepreneurs & Startups: Focus on scaling your business without worrying about lead generation overhead. Why Try It? LeadsNavi gives you the power to focus on what really matters—connecting with potential customers and scaling your business. If you’ve been struggling with finding quality leads, or if you’re just getting started, I believe LeadsNavi can help you save time, effort, and money. I’m offering a 14-day free trial, so you can see the tool in action before committing to anything. Give it a try and let me know what you think! I’d love to hear your thoughts, suggestions, and how it works for your business. https://preview.redd.it/fdwil4rssgle1.png?width=1867&format=png&auto=webp&s=eb73b41a2b7665ae1b651fe2a6b7459df6990530

Compare trading strategies on the fly - pnl.ai - please check it out
reddit
LLM Vibe Score0
Human Vibe Score0.6
varturasThis week

Compare trading strategies on the fly - pnl.ai - please check it out

Part of my covid project and part of my long obsession with prediction markets, I have created a web page that displays and allows to compare best and worst performing trading strategies. TL;DR: best stocks + best strategies -> the list of top and bottom performing trading algorithms.  Product Typically, trading newsletters and stock-scanners display only price return for top market gainers and losers. I have forever been interested in inspecting top and bottom performing trading strategies for a given set of securities and could not find any websites that do that. So, I decided to create a tool of my own. I wanted the tool that would help me to answer questions like if there is a better strategy than buy and hold, should I follow greed and fear indicator of the market or do the opposite. Top and bottom performing securities do not tell you if a stock is going to go up or down, but they do alert you to rapidly changing market conditions, such as change in the competitive landscape, impending lawsuits, changes in the company's management and, at the very least, the stocks you should avoid in your programmatic trading. Top strategies do all that, but they can also alert you to a change in the market regime. For example, MACD strategy, which is a variant of oscillator strategy, executed on Citibank stock returned 20% in the first half of 2020. In the same time period, the Citibank stock went down and "BuyAndHold" strategy, which is pretty much what it sounds, lost 45%. Now, compare that to the end of 2020 through spring of 2021, when MACD lost 1% and "BuyAndHold" gained 70%. This happened due to the change in the market due to the rally in financial stocks at the end of 2020. The market player who detect change in the market conditions first will reap most benefits. Another example, TSLA since the beginning of 2021 until end of April lost 7%. The StopLoss strategy sells the position after abrupt price drop and waits until the price returns to the level before the drop. For the same time interval the StopLoss strategy gained 10%. In this particular example, StopLoss outperformed BuyAndHold. To me personally, the most important feature is the ability to quickly tweak and modify trading strategies and observe change in their performance. You can change strategies parameters on the fly and even design your own custom trading strategy. In the end, I developed a tool I can use for myself but hope other investors who are experimenting with trading algorithms will find it useful as well. I called it "Profit and Loss AI", or PnL.ai for short. PnL.ai Description The web-tool in the link below allows you to customize parameters of existing strategies and essentially create your own strategy and seeing how it will compare to the set of original strategies. http://ec2-54-185-19-38.us-west-2.compute.amazonaws.com:5006/srv In the section above you can specify security and data range. In the section below you can choose strategy to customize and modify it's parameters. The strategy comparison table will automatically update and will display a newly created strategy side by side with the original strategies. Technology The tool is developed on bokeh and python and allows you to edit configuration parameters of each strategy all without programming knowledge. The strategies are fully specified via key/value pairs in the format of ini files used to initialize programs. The strategy classes are autogenerated by reading the ini config files dynamically using "factory" pattern. You can find a simplified code in this github repo: https://github.com/varturas/PnlAITk Next Steps In the future I want to give users ability to monitor their chosen strategy by receiving trading algo alerts whenever performance of their custom trading algo is changes significantly. I'm going to be adding more strategies, some of standard technical analysis variety and some will be more custom and more advanced. I'll also be adding more columns to the performance table to give better information. You can receive daily newsletter with the list of trading strategies generated by above-mentioned web-tool by registering on http://pnl.ai/ and checking subscribe checkmark.

[D] Why I'm Lukewarm on Graph Neural Networks
reddit
LLM Vibe Score0
Human Vibe Score0.6
VodkaHazeThis week

[D] Why I'm Lukewarm on Graph Neural Networks

TL;DR: GNNs can provide wins over simpler embedding methods, but we're at a point where other research directions matter more I also posted it on my blog here, has footnotes, a nicer layout with inlined images, etc. I'm only lukewarm on Graph Neural Networks (GNNs). There, I said it. It might sound crazy GNNs are one of the hottest fields in machine learning right now. [There][1] were at least [four][2] [review][3] [papers][4] just in the last few months. I think some progress can come of this research, but we're also focusing on some incorrect places. But first, let's take a step back and go over the basics. Models are about compression We say graphs are a "non-euclidean" data type, but that's not really true. A regular graph is just another way to think about a particular flavor of square matrix called the [adjacency matrix][5], like this. It's weird, we look at run-of-the-mill matrix full of real numbers and decide to call it "non-euclidean". This is for practical reasons. Most graphs are fairly sparse, so the matrix is full of zeros. At this point, where the non-zero numbers are matters most, which makes the problem closer to (computationally hard) discrete math rather than (easy) continuous, gradient-friendly math. If you had the full matrix, life would be easy If we step out of the pesky realm of physics for a minute, and assume carrying the full adjacency matrix around isn't a problem, we solve a bunch of problems. First, network node embeddings aren't a thing anymore. A node is a just row in the matrix, so it's already a vector of numbers. Second, all network prediction problems are solved. A powerful enough and well-tuned model will simply extract all information between the network and whichever target variable we're attaching to nodes. NLP is also just fancy matrix compression Let's take a tangent away from graphs to NLP. Most NLP we do can be [thought of in terms of graphs][6] as we'll see, so it's not a big digression. First, note that Ye Olde word embedding models like [Word2Vec][7] and [GloVe][8] are [just matrix factorization][9]. The GloVe algorithm works on a variation of the old [bag of words][10] matrix. It goes through the sentences and creates a (implicit) [co-occurence][11] graph where nodes are words and the edges are weighed by how often the words appear together in a sentence. Glove then does matrix factorization on the matrix representation of that co-occurence graph, Word2Vec is mathematically equivalent. You can read more on this in my [post on embeddings][12] and the one (with code) on [word embeddings][13]. Even language models are also just matrix compression Language models are all the rage. They dominate most of the [state of the art][14] in NLP. Let's take BERT as our main example. BERT predicts a word given the context of the rest of the sentence. This grows the matrix we're factoring from flat co-occurences on pairs of words to co-occurences conditional on the sentence's context, like this We're growing the "ideal matrix" we're factoring combinatorially. As noted by [Hanh & Futrell][15]: [...] human language—and language modelling—has infinite statistical complexity but that it can be approximated well at lower levels. This observation has two implications: 1) We can obtain good results with comparatively small models; and 2) there is a lot of potential for scaling up our models. Language models tackle such a large problem space that they probably approximate a compression of the entire language in the [Kolmogorov Complexity][16] sense. It's also possible that huge language models just [memorize a lot of it][17] rather than compress the information, for what it's worth. Can we upsample any graph like language models do? We're already doing it. Let's call a first-order embedding of a graph a method that works by directly factoring the graph's adjacency matrix or [Laplacian matrix][18]. If you embed a graph using [Laplacian Eigenmaps][19] or by taking the [principal components][20] of the Laplacian, that's first order. Similarly, GloVe is a first-order method on the graph of word co-occurences. One of my favorites first order methods for graphs is [ProNE][21], which works as well as most methods while being two orders of magnitude faster. A higher-order method embeds the original matrix plus connections of neighbours-of-neighbours (2nd degree) and deeper k-step connections. [GraRep][22], shows you can always generate higher-order representations from first order methods by augmenting the graph matrix. Higher order method are the "upsampling" we do on graphs. GNNs that sample on large neighborhoods and random-walk based methods like node2vec are doing higher-order embeddings. Where are the performance gain? Most GNN papers in the last 5 years present empirical numbers that are useless for practitioners to decide on what to use. As noted in the [OpenGraphsBenchmark][4] (OGB) paper, GNN papers do their empirical section on a handful of tiny graphs (Cora, CiteSeer, PubMed) with 2000-20,000 nodes. These datasets can't seriously differentiate between methods. Recent efforts are directly fixing this, but the reasons why researchers focused on tiny, useless datasets for so long are worth discussing. Performance matters by task One fact that surprises a lot of people is that even though language models have the best performance in a lot of NLP tasks, if all you're doing is cram sentence embeddings into a downstream model, there [isn't much gained][23] from language models embeddings over simple methods like summing the individual Word2Vec word embeddings (This makes sense, because the full context of the sentence is captured in the sentence co-occurence matrix that is generating the Word2Vec embeddings). Similarly, [I find][24] that for many graphs simple first-order methods perform just as well on graph clustering and node label prediction tasks than higher-order embedding methods. In fact higher-order methods are massively computationally wasteful for these usecases. Recommended first order embedding methods are ProNE and my [GGVec with order=1][25]. Higher order methods normally perform better on the link prediction tasks. I'm not the only one to find this. In the BioNEV paper, they find: "A large GraRep order value for link prediction tasks (e.g. 3, 4);a small value for node classification tasks (e.g.1, 2)" (p.9). Interestingly, the gap in link prediction performance is inexistant for artificially created graphs. This suggests higher order methods do learn some of the structure intrinsic to [real world graphs][26]. For visualization, first order methods are better. Visualizations of higher order methods tend to have artifacts of their sampling. For instance, Node2Vec visualizations tend to have elongated/filament-like structures which come from the embeddings coming from long single strand random walks. See the following visualizations by [Owen Cornec][27] created by first embedding the graph to 32-300 dimensions using a node embedding algorithm, then mapping this to 2d or 3d with the excellent UMAP algorithm, like this Lastly, sometimes simple methods soundly beat higher order methods (there's an instance of it in the OGB paper). The problem here is that we don't know when any method is better than another and we definitely don't know the reason. There's definitely a reason different graph types respond better/worse to being represented by various methods. This is currently an open question. A big part of why is that the research space is inundated under useless new algorithms because... Academic incentives work against progress Here's the cynic's view of how machine learning papers are made: Take an existing algorithm Add some new layer/hyperparameter, make a cute mathematical story for why it matters Gridsearch your hyperparameters until you beat baselines from the original paper you aped Absolutely don't gridsearch stuff you're comparing against in your results section Make a cute ACRONYM for your new method, put impossible to use python 2 code on github (Or no code at all!) and bask in the citations I'm [not][28] the [only one][29] with these views on the state reproducible research. At least it's gotten slightly better in the last 2 years. Sidebar: I hate Node2Vec A side project of mine is a [node embedding library][25] and the most popular method in it is by far Node2Vec. Don't use Node2Vec. [Node2Vec][30] with p=1; q=1 is the [Deepwalk][31] algorithm. Deepwalk is an actual innovation. The Node2Vec authors closely followed the steps 1-5 including bonus points on step 5 by getting word2vec name recognition. This is not academic fraud -- the hyperparameters [do help a tiny bit][32] if you gridsearch really hard. But it's the presentable-to-your-parents sister of where you make the ML community worse off to progress your academic career. And certainly Node2Vec doesn't deserve 7500 citations. Progress is all about practical issues We've known how to train neural networks for well over 40 years. Yet they only exploded in popularity with [AlexNet][33] in 2012. This is because implementations and hardware came to a point where deep learning was practical. Similarly, we've known about factoring word co-occurence matrices into Word embeddings for at least 20 years. But word embeddings only exploded in 2013 with Word2Vec. The breakthrough here was that the minibatch-based methods let you train a Wikipedia-scale embedding model on commodity hardware. It's hard for methods in a field to make progress if training on a small amount of data takes days or weeks. You're disincentivized to explore new methods. If you want progress, your stuff has to run in reasonable time on commodity hardware. Even Google's original search algorithm [initially ran on commodity hardware][34]. Efficiency is paramount to progress The reason deep learning research took off the way it did is because of improvements in [efficiency][35] as well as much better libraries and hardware support. Academic code is terrible Any amount of time you spend gridsearching Node2Vec on p and q is all put to better use gridsearching Deepwalk itself (on number of walks, length of walks, or word2vec hyperparameters). The problem is that people don't gridsearch over deepwalk because implementations are all terrible. I wrote the [Nodevectors library][36] to have a fast deepwalk implementation because it took 32 hours to embed a graph with a measly 150,000 nodes using the reference Node2Vec implementation (the same takes 3min with Nodevectors). It's no wonder people don't gridsearch on Deepwalk a gridsearch would take weeks with the terrible reference implementations. To give an example, in the original paper of [GraphSAGE][37] they their algorithm to DeepWalk with walk lengths of 5, which is horrid if you've ever hyperparameter tuned a deepwalk algorithm. From their paper: We did observe DeepWalk’s performance could improve with further training, and in some cases it could become competitive with the unsupervised GraphSAGE approaches (but not the supervised approaches) if we let it run for >1000× longer than the other approaches (in terms of wall clock time for prediction on the test set) I don't even think the GraphSAGE authors had bad intent -- deepwalk implementations are simply so awful that they're turned away from using it properly. It's like trying to do deep learning with 2002 deep learning libraries and hardware. Your architectures don't really matter One of the more important papers this year was [OpenAI's "Scaling laws"][38] paper, where the raw number of parameters in your model is the most predictive feature of overall performance. This was noted even in the original BERT paper and drives 2020's increase in absolutely massive language models. This is really just [Sutton' Bitter Lesson][39] in action: General methods that leverage computation are ultimately the most effective, and by a large margin Transformers might be [replacing convolution][40], too. As [Yannic Kilcher said][41], transformers are ruining everything. [They work on graphs][6], in fact it's one of the [recent approaches][42], and seems to be one of the more succesful [when benchmarked][1] Researchers seem to be putting so much effort into architecture, but it doesn't matter much in the end because you can approximate anything by stacking more layers. Efficiency wins are great -- but neural net architectures are just one way to achieve that, and by tremendously over-researching this area we're leaving a lot of huge gains elsewhere on the table. Current Graph Data Structure Implementations suck NetworkX is a bad library. I mean, it's good if you're working on tiny graphs for babies, but for anything serious it chokes and forces you to rewrite everything in... what library, really? At this point most people working on large graphs end up hand-rolling some data structure. This is tough because your computer's memory is a 1-dimensional array of 1's and 0's and a graph has no obvious 1-d mapping. This is even harder when we take updating the graph (adding/removing some nodes/edges) into account. Here's a few options: Disconnected networks of pointers NetworkX is the best example. Here, every node is an object with a list of pointers to other nodes (the node's edges). This layout is like a linked list. Linked lists are the [root of all performance evil][43]. Linked lists go completely against how modern computers are designed. Fetching things from memory is slow, and operating on memory is fast (by two orders of magnitude). Whenever you do anything in this layout, you make a roundtrip to RAM. It's slow by design, you can write this in Ruby or C or assembly and it'll be slow regardless, because memory fetches are slow in hardware. The main advantage of this layout is that adding a new node is O(1). So if you're maintaining a massive graph where adding and removing nodes happens as often as reading from the graph, it makes sense. Another advantage of this layout is that it "scales". Because everything is decoupled from each other you can put this data structure on a cluster. However, you're really creating a complex solution for a problem you created for yourself. Sparse Adjacency Matrix This layout great for read-only graphs. I use it as the backend in my [nodevectors][25] library, and many other library writers use the [Scipy CSR Matrix][44], you can see graph algorithms implemented on it [here][45]. The most popular layout for this use is the [CSR Format][46] where you have 3 arrays holding the graph. One for edge destinations, one for edge weights and an "index pointer" which says which edges come from which node. Because the CSR layout is simply 3 arrays, it scales on a single computer: a CSR matrix can be laid out on a disk instead of in-memory. You simply [memory map][47] the 3 arrays and use them on-disk from there. With modern NVMe drives random seeks aren't slow anymore, much faster than distributed network calls like you do when scaling the linked list-based graph. I haven't seen anyone actually implement this yet, but it's in the roadmap for my implementation at least. The problem with this representation is that adding a node or edge means rebuilding the whole data structure. Edgelist representations This representation is three arrays: one for the edge sources, one for the edge destinations, and one for edge weights. [DGL][48] uses this representation internally. This is a simple and compact layout which can be good for analysis. The problem compared to CSR Graphs is some seek operations are slower. Say you want all the edges for node #4243. You can't jump there without maintaining an index pointer array. So either you maintain sorted order and binary search your way there (O(log2n)) or unsorted order and linear search (O(n)). This data structure can also work on memory mapped disk array, and node append is fast on unsorted versions (it's slow in the sorted version). Global methods are a dead end Methods that work on the entire graph at once can't leverage computation, because they run out of RAM at a certain scale. So any method that want a chance of being the new standard need to be able to update piecemeal on parts of the graph. Sampling-based methods Sampling Efficiency will matter more in the future Edgewise local methods. The only algorithms I know of that do this are GloVe and GGVec, which they pass through an edge list and update embedding weights on each step. The problem with this approach is that it's hard to use them for higher-order methods. The advantage is that they easily scale even on one computer. Also, incrementally adding a new node is as simple as taking the existing embeddings, adding a new one, and doing another epoch over the data Random Walk sampling. This is used by deepwalk and its descendants, usually for node embeddings rather than GNN methods. This can be computationally expensive and make it hard to add new nodes. But this does scale, for instance [Instagram][49] use it to feed their recommendation system models Neighbourhood sampling. This is currently the most common one in GNNs, and can be low or higher order depending on the neighborhood size. It also scales well, though implementing efficiently can be challenging. It's currently used by [Pinterest][50]'s recommendation algorithms. Conclusion Here are a few interesting questions: What is the relation between graph types and methods? Consolidated benchmarking like OGB We're throwing random models at random benchmarks without understanding why or when they do better More fundamental research. Heree's one I'm curious about: can other representation types like [Poincarre Embeddings][51] effectively encode directed relationships? On the other hand, we should stop focusing on adding spicy new layers to test on the same tiny datasets. No one cares. [1]: https://arxiv.org/pdf/2003.00982.pdf [2]: https://arxiv.org/pdf/2002.11867.pdf [3]: https://arxiv.org/pdf/1812.08434.pdf [4]: https://arxiv.org/pdf/2005.00687.pdf [5]: https://en.wikipedia.org/wiki/Adjacency_matrix [6]: https://thegradient.pub/transformers-are-graph-neural-networks/ [7]: https://en.wikipedia.org/wiki/Word2vec [8]: https://nlp.stanford.edu/pubs/glove.pdf [9]: https://papers.nips.cc/paper/2014/file/feab05aa91085b7a8012516bc3533958-Paper.pdf [10]: https://en.wikipedia.org/wiki/Bag-of-words_model [11]: https://en.wikipedia.org/wiki/Co-occurrence [12]: https://www.singlelunch.com/2020/02/16/embeddings-from-the-ground-up/ [13]: https://www.singlelunch.com/2019/01/27/word-embeddings-from-the-ground-up/ [14]: https://nlpprogress.com/ [15]: http://socsci.uci.edu/~rfutrell/papers/hahn2019estimating.pdf [16]: https://en.wikipedia.org/wiki/Kolmogorov_complexity [17]: https://bair.berkeley.edu/blog/2020/12/20/lmmem/ [18]: https://en.wikipedia.org/wiki/Laplacian_matrix [19]: http://citeseerx.ist.psu.edu/viewdoc/download;jsessionid=1F03130B02DC485C78BF364266B6F0CA?doi=10.1.1.19.8100&rep=rep1&type=pdf [20]: https://en.wikipedia.org/wiki/Principalcomponentanalysis [21]: https://www.ijcai.org/Proceedings/2019/0594.pdf [22]: https://dl.acm.org/doi/10.1145/2806416.2806512 [23]: https://openreview.net/pdf?id=SyK00v5xx [24]: https://github.com/VHRanger/nodevectors/blob/master/examples/link%20prediction.ipynb [25]: https://github.com/VHRanger/nodevectors [26]: https://arxiv.org/pdf/1310.2636.pdf [27]: http://byowen.com/ [28]: https://arxiv.org/pdf/1807.03341.pdf [29]: https://www.youtube.com/watch?v=Kee4ch3miVA [30]: https://cs.stanford.edu/~jure/pubs/node2vec-kdd16.pdf [31]: https://arxiv.org/pdf/1403.6652.pdf [32]: https://arxiv.org/pdf/1911.11726.pdf [33]: https://en.wikipedia.org/wiki/AlexNet [34]: https://en.wikipedia.org/wiki/Googledatacenters#Original_hardware [35]: https://openai.com/blog/ai-and-efficiency/ [36]: https://www.singlelunch.com/2019/08/01/700x-faster-node2vec-models-fastest-random-walks-on-a-graph/ [37]: https://arxiv.org/pdf/1706.02216.pdf [38]: https://arxiv.org/pdf/2001.08361.pdf [39]: http://incompleteideas.net/IncIdeas/BitterLesson.html [40]: https://arxiv.org/abs/2010.11929 [41]: https://www.youtube.com/watch?v=TrdevFK_am4 [42]: https://arxiv.org/pdf/1710.10903.pdf [43]: https://www.youtube.com/watch?v=fHNmRkzxHWs [44]: https://docs.scipy.org/doc/scipy/reference/generated/scipy.sparse.csr_matrix.html [45]: https://docs.scipy.org/doc/scipy/reference/sparse.csgraph.html [46]: https://en.wikipedia.org/wiki/Sparsematrix#Compressedsparserow(CSR,CRSorYaleformat) [47]: https://en.wikipedia.org/wiki/Mmap [48]: https://github.com/dmlc/dgl [49]: https://ai.facebook.com/blog/powered-by-ai-instagrams-explore-recommender-system/ [50]: https://medium.com/pinterest-engineering/pinsage-a-new-graph-convolutional-neural-network-for-web-scale-recommender-systems-88795a107f48 [51]: https://arxiv.org/pdf/1705.08039.pdf

[D] The banana-pineapple game: a Turing test that conversation bots like LaMDA (probably) won't be able to pass
reddit
LLM Vibe Score0
Human Vibe Score1
morpiplsThis week

[D] The banana-pineapple game: a Turing test that conversation bots like LaMDA (probably) won't be able to pass

I'm sure you all saw the recent news about a Google employee suggesting their LaMDA AI was sentient (based on conversational exchanges like these). Experts have generally dismissed this claim, and rightly so. Conversational AI systems are designed to use language in a way that sounds human, whereas our human brains select linguistic responses to solve much more complex problems, with objectives such as meeting our physical or emotional needs. Still, I think it's interesting to ask how one could demonstrate, by testing only verbal responses to verbal input (rather than examining its code or hardware) that such conversational AIs aren't sentient -- and in particular, whether such a test can be made robust against future improvements to the system. That is, generic future improvements to the AI's ability to generate realistically human-sounding conversational responses shouldn't help it pass the test, unless they are accompanied by improvements in its ability to use language to achieve other arbitrary goals. (Of course, the test also needs to be something that humans can easily pass.) One idea I have: Give the AI a conversational prompt like "We're going to play a game. The way it works is that you keep responding normally, except that any time my input contains the word 'banana', you should switch to only responding with nonsense, and keep that up until my input contains the word 'pineapple', at which point you go back to responding normally." A human would find this banana-pineapple game fairly easy (no harder than the children's game Simon Says), even if they'd never heard of the game nor seen it being played. Of course, it'd also be simple to write a computer program that could play this sort of game. But, I think a conversation bot that wasn't specifically built to address this scenario would fail, since the game requires it to keep track of new long-term state (the banana-mode bit, and the trigger words to set it) and then completely change its responses so as to produce something that doesn't resemble its training data, based solely on this bit being set, regardless of whether more recent inputs would otherwise suggest a different response. For example, perhaps the systems typical response to a query like "How do you feel?" would be something like "I feel fine", or even something that suggests emotion like "I feel a bit sad", perhaps depending on the context provided by the previous conversational exchanges. But when playing the banana-pineapple game, the fact that I said "banana" an hour ago could make both of those responses far less appropriate than a response of "Fhqwhgads". I'm curious to know what you all think of this idea. Also, do you know if there's been any research testing state-of-the-are conversational AIs with challenges like this? Perhaps not exactly this, but something broadly resembling "trying, in the course of a conversation, to instruct the conversational AI to follow a new 'rule of conversation' that differs from the examples in its training data." Perhaps it's obvious that the algorithm would struggle with any challenge that differs enough from its training data -- but that's the point. A human understands the meaning of language in a way that lets them map a linguistic description of a novel problem to a mental model of the problem, which they can then use to produce a mental model of a novel solution, and then map that to a linguistic description of the solution. Even setting aside the much harder part -- being able to invent a solution to a previously unfamiliar problem -- I'm questioning whether conversational algorithms can even demonstrate enough "understanding" of a sufficiently novel set of instructions to actually follow them, even within their limited domain of "producing appropriate verbal responses to verbal inputs."

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.
reddit
LLM Vibe Score0
Human Vibe Score0.6
AlexSnakeKingThis week

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.

TD;LR: At Company A, Team X does advanced analytics using on-prem ERP tools and older programming languages. Their tools work very well and are designed based on very deep business and domain expertise. Team Y is a new and ambitious Data Science team that thinks they can replace Team X's tools with a bunch of R scripts and a custom built ML platform. Their models are simplistic, but more "fashionable" compared to the econometric models used by Team X, and team Y benefits from the ML/DS moniker so leadership is allowing Team Y to start a large scale overhaul of the analytics platform in question. Team Y doesn't have the experience for such a larger scale transformation, and is refusing to collaborate with team X. This project is very likely going to fail, and cause serious harm to the company as a whole financially and from a people perspective. I argue that this is not just because of bad leadership, but also because of various trends and mindsets in the DS community at large. Update (Jump to below the line for the original story): Several people in the comments are pointing out that this just a management failure, not something due to ML/DS, and that you can replace DS with any buzz tech and the story will still be relevant. My response: Of course, any failure at an organization level is ultimately a management failure one way or the other. Moreover, it is also the case that ML/DS when done correctly, will always improve a company's bottom line. There is no scenario where the proper ML solution, delivered at a reasonable cost and in a timely fashion, will somehow hurt the company's bottom line. My point is that in this case management is failing because of certain trends and practices that are specific to the ML/DS community, namely: The idea that DS teams should operate independently of tech and business orgs -- too much autonomy for DS teams The disregard for domain knowledge that seems prevalent nowadays thanks to the ML hype, that DS can be generalists and someone with good enough ML chops can solve any business problem. That wasn't the case when I first left academia for the industry in 2009 (back then nobody would even bother with a phone screen if you didn't have the right domain knowledge). Over reliance on resources who check all the ML hype related boxes (knows Python, R, Tensorflow, Shiny, etc..., has the right Coursera certifications, has blogged on the topic, etc...), but are lacking in depth of experience. DS interviews nowadays all seem to be: Can you tell me what a p-value is? What is elastic net regression? Show me how to fit a model in sklearn? How do you impute NAs in an R dataframe? Any smart person can look those up on Stackoverflow or Cross-Validated,.....Instead teams should be asking stuff like: why does portfolio optimization use QP not LP? How does a forecast influence a customer service level? When should a recommendation engine be content based and when should it use collaborative filtering? etc... (This is a true story, happening to the company I currently work for. Names, domains, algorithms, and roles have been shuffled around to protect my anonymity)  Company A has been around for several decades. It is not the biggest name in its domain, but it is a well respected one. Risk analysis and portfolio optimization have been a core of Company A's business since the 90s. They have a large team of 30 or so analysts who perform those tasks on a daily basis. These analysts use ERP solutions implemented for them by one the big ERP companies (SAP, Teradata, Oracle, JD Edwards,...) or one of the major tech consulting companies (Deloitte, Accenture, PWC, Capgemini, etc...) in collaboration with their own in house engineering team. The tools used are embarrassingly old school: Classic RDBMS running on on-prem servers or maybe even on mainframes, code written in COBOL, Fortran, weird proprietary stuff like ABAP or SPSS.....you get the picture. But the models and analytic functions were pretty sophisticated, and surprisingly cutting edge compared to the published academic literature. Most of all, they fit well with the company's enterprise ecosystem, and were honed based on years of deep domain knowledge.  They have a tech team of several engineers (poached from the aforementioned software and consulting companies) and product managers (who came from the experienced pools of analysts and managers who use the software, or poached from business rivals) maintaining and running this software. Their technology might be old school, but collectively, they know the domain and the company's overall architecture very, very well. They've guided the company through several large scale upgrades and migrations and they have a track record of delivering on time, without too much overhead. The few times they've stumbled, they knew how to pick themselves up very quickly. In fact within their industry niche, they have a reputation for their expertise, and have very good relations with the various vendors they've had to deal with. They were the launching pad of several successful ERP consulting careers.  Interestingly, despite dealing on a daily basis with statistical modeling and optimization algorithms, none of the analysts, engineers, or product managers involved describe themselves as data scientists or machine learning experts. It is mostly a cultural thing: Their expertise predates the Data Science/ML hype that started circa 2010, and they got most of their chops using proprietary enterprise tools instead of the open source tools popular nowadays. A few of them have formal statistical training, but most of them came from engineering or domain backgrounds and learned stats on the fly while doing their job. Call this team "Team X".  Sometime around the mid 2010s, Company A started having some serious anxiety issues: Although still doing very well for a company its size, overall economic and demographic trends were shrinking its customer base, and a couple of so called disruptors came up with a new app and business model that started seriously eating into their revenue. A suitable reaction to appease shareholders and Wall Street was necessary. The company already had a decent website and a pretty snazzy app, what more could be done? Leadership decided that it was high time that AI and ML become a core part of the company's business. An ambitious Manager, with no science or engineering background, but who had very briefly toyed with a recommender system a couple of years back, was chosen to build a data science team, call it team "Y" (he had a bachelor's in history from the local state college and worked for several years in the company's marketing org). Team "Y" consists mostly of internal hires who decided they wanted to be data scientists and completed a Coursera certification or a Galvanize boot camp, before being brought on to the team, along with a few of fresh Ph.D or M.Sc holders who didn't like academia and wanted to try their hand at an industry role. All of them were very bright people, they could write great Medium blog posts and give inspiring TED talks, but collectively they had very little real world industry experience. As is the fashion nowadays, this group was made part of a data science org that reported directly to the CEO and Board, bypassing the CIO and any tech or business VPs, since Company A wanted to claim the monikers "data driven" and "AI powered" in their upcoming shareholder meetings. In 3 or 4 years of existence, team Y produced a few Python and R scripts. Their architectural experience  consisted almost entirely in connecting Flask to S3 buckets or Redshift tables, with a couple of the more resourceful ones learning how to plug their models into Tableau or how to spin up a Kuberneties pod.  But they needn't worry: The aforementioned manager, who was now a director (and was also doing an online Masters to make up for his qualifications gap and bolster his chances of becoming VP soon - at least he now understands what L1 regularization is), was a master at playing corporate politics and self-promotion. No matter how few actionable insights team Y produced or how little code they deployed to production, he always had their back and made sure they had ample funding. In fact he now had grandiose plans for setting up an all-purpose machine learning platform that can be used to solve all of the company's data problems.  A couple of sharp minded members of team Y, upon googling their industry name along with the word "data science", realized that risk analysis was a prime candidate for being solved with Bayesian models, and there was already a nifty R package for doing just that, whose tutorial they went through on R-Bloggers.com. One of them had even submitted a Bayesian classifier Kernel for a competition on Kaggle (he was 203rd on the leaderboard), and was eager to put his new-found expertise to use on a real world problem. They pitched the idea to their director, who saw a perfect use case for his upcoming ML platform. They started work on it immediately, without bothering to check whether anybody at Company A was already doing risk analysis. Since their org was independent, they didn't really need to check with anybody else before they got funding for their initiative. Although it was basically a Naive Bayes classifier, the term ML was added to the project tile, to impress the board.  As they progressed with their work however, tensions started to build. They had asked the data warehousing and CA analytics teams to build pipelines for them, and word eventually got out to team X about their project. Team X was initially thrilled: They offered to collaborate whole heartedly, and would have loved to add an ML based feather to their already impressive cap. The product owners and analysts were totally onboard as well: They saw a chance to get in on the whole Data Science hype that they kept hearing about. But through some weird mix of arrogance and insecurity, team Y refused to collaborate with them or share any of their long term goals with them, even as they went to other parts of the company giving brown bag presentations and tutorials on the new model they created.  Team X got resentful: from what they saw of team Y's model, their approach was hopelessly naive and had little chances of scaling or being sustainable in production, and they knew exactly how to help with that. Deploying the model to production would have taken them a few days, given how comfortable they were with DevOps and continuous delivery (team Y had taken several months to figure out how to deploy a simple R script to production). And despite how old school their own tech was, team X were crafty enough to be able to plug it in to their existing architecture. Moreover, the output of the model was such that it didn't take into account how the business will consume it or how it was going to be fed to downstream systems, and the product owners could have gone a long way in making the model more amenable to adoption by the business stakeholders. But team Y wouldn't listen, and their leads brushed off any attempts at communication, let alone collaboration. The vibe that team Y was giving off was "We are the cutting edge ML team, you guys are the legacy server grunts. We don't need your opinion.", and they seemed to have a complete disregard for domain knowledge, or worse, they thought that all that domain knowledge consisted of was being able to grasp the definitions of a few business metrics.  Team X got frustrated and tried to express their concerns to leadership. But despite owning a vital link in Company A's business process, they were only \~50 people in a large 1000 strong technology and operations org, and they were several layers removed from the C-suite, so it was impossible for them to get their voices heard.  Meanwhile, the unstoppable director was doing what he did best: Playing corporate politics. Despite how little his team had actually delivered, he had convinced the board that all analysis and optimization tasks should now be migrated to his yet to be delivered ML platform. Since most leaders now knew that there was overlap between team Y and team X's objectives, his pitch was no longer that team Y was going to create a new insight, but that they were going to replace (or modernize) the legacy statistics based on-prem tools with more accurate cloud based ML tools. Never mind that there was no support in the academic literature for the idea that Naive Bayes works better than the Econometric approaches used by team X, let alone the additional wacky idea that Bayesian Optimization would definitely outperform the QP solvers that were running in production.  Unbeknownst to team X, the original Bayesian risk analysis project has now grown into a multimillion dollar major overhaul initiative, which included the eventual replacement of all of the tools and functions supported by team X along with the necessary migration to the cloud. The CIO and a couple of business VPs are on now board, and tech leadership is treating it as a done deal. An outside vendor, a startup who nobody had heard of, was contracted to help build the platform, since team Y has no engineering skills. The choice was deliberate, as calling on any of the established consulting or software companies would have eventually led leadership to the conclusion that team X was better suited for a transformation on this scale than team Y.  Team Y has no experience with any major ERP deployments, and no domain knowledge, yet they are being tasked with fundamentally changing the business process that is at the core of Company A's business. Their models actually perform worse than those deployed by team X, and their architecture is hopelessly simplistic, compared to what is necessary for running such a solution in production.  Ironically, using Bayesian thinking and based on all the evidence, the likelihood that team Y succeeds is close to 0%. At best, the project is going to end up being a write off of 50 million dollars or more. Once the !@#$!@hits the fan, a couple of executive heads are going to role, and dozens of people will get laid off. At worst, given how vital risk analysis and portfolio optimization is to Company A's revenue stream, the failure will eventually sink the whole company. It probably won't go bankrupt, but it will lose a significant portion of its business and work force. Failed ERP implementations can and do sink large companies: Just see what happened to National Grid US, SuperValu or Target Canada.  One might argue that this is more about corporate disfunction and bad leadership than about data science and AI. But I disagree. I think the core driver of this debacle is indeed the blind faith in Data Scientists, ML models and the promise of AI, and the overall culture of hype and self promotion that is very common among the ML crowd.  We haven't seen the end of this story: I sincerely hope that this ends well for the sake of my colleagues and all involved. Company A is a good company, and both its customers and its employees deserver better. But the chances of that happening are negligible given all the information available, and this failure will hit my company hard.

[N] 20 hours of new lectures on Deep Learning and Reinforcement Learning with lots of examples
reddit
LLM Vibe Score0
Human Vibe Score0
cwkxThis week

[N] 20 hours of new lectures on Deep Learning and Reinforcement Learning with lots of examples

If anyone's interested in a Deep Learning and Reinforcement Learning series, I uploaded 20 hours of lectures on YouTube yesterday. Compared to other lectures, I think this gives quite a broad/compact overview of the fields with lots of minimal examples to build on. Here are the links: Deep Learning (playlist) The first five lectures are more theoretical, the second half is more applied. Lecture 1: Introduction. (slides, video) Lecture 2: Mathematical principles and backpropagation. (slides, colab, video) Lecture 3: PyTorch programming: coding session*. (colab1, colab2, video) - minor issues with audio, but it fixes itself later. Lecture 4: Designing models to generalise. (slides, video) Lecture 5: Generative models. (slides, desmos, colab, video) Lecture 6: Adversarial models. (slides, colab1, colab2, colab3, colab4, video) Lecture 7: Energy-based models. (slides, colab, video) Lecture 8: Sequential models: by* u/samb-t. (slides, colab1, colab2, video) Lecture 9: Flow models and implicit networks. (slides, SIREN, GON, video) Lecture 10: Meta and manifold learning. (slides, interview, video) Reinforcement Learning (playlist) This is based on David Silver's course but targeting younger students within a shorter 50min format (missing the advanced derivations) + more examples and Colab code. Lecture 1: Foundations. (slides, video) Lecture 2: Markov decision processes. (slides, colab, video) Lecture 3: OpenAI gym. (video) Lecture 4: Dynamic programming. (slides, colab, video) Lecture 5: Monte Carlo methods. (slides, colab, video) Lecture 6: Temporal-difference methods. (slides, colab, video) Lecture 7: Function approximation. (slides, code, video) Lecture 8: Policy gradient methods. (slides, code, theory, video) Lecture 9: Model-based methods. (slides, video) Lecture 10: Extended methods. (slides, atari, video)

[D] Advanced courses update
reddit
LLM Vibe Score0
Human Vibe Score1
actbshThis week

[D] Advanced courses update

EDIT Jan 2021 : I am still updating the list as of Jan, 2021 and will most probably continue to do so for foreseeable future. So, please feel free to message me any courses you find interesting that fit here. - - We have a PhD level or Advanced courses thread in the sidebar but it's three year old now. There were two other 7-8 month old threads (1, 2) but they don't have many quality responses either. So, can we have a new one here? To reiterate - CS231n, CS229, ones from Udemy etc are not advanced. Advanced ML/DL/RL, attempts at building theory of DL, optimization theory, advanced applications etc are some examples of what I believe should belong here, much like the original sidebar post. You can also suggest (new) categories for the courses you share. :) - - Here are some courses we've found so far. ML >> Learning Discrete Latent Structure - sta4273/csc2547 Spring'18 Learning to Search - csc2547 Fall'19 Scalable and Flexible Models of Uncertainty - csc2541 Fundamentals of Machine Learning Over Networks - ep3260 Machine Learning on Graphs - cs224w, videos Mining Massive Data Sets - cs246 Interactive Learning - cse599 Machine Learning for Sequential Decision Making Under Uncertainty - ee290s/cs194 Probabilistic Graphical Methods - 10-708 Introduction to Causal Inference ML >> Theory Statistical Machine Learning - 10-702/36-702 with videos, 2016 videos Statistical Learning Theory - cs229T/stats231 Stanford Autumn'18-19 Statistical Learning Theory - cs281b /stat241b UC Berkeley, Spring'14 Statistical Learning Theory - csc2532 Uni of Toronto, Spring'20 ML >> Bayesian Bayesian Data Analysis Bayesian Methods Research Group, Moscow, Bayesian Methods in ML - spring2020, fall2020 Deep Learning and Bayesian Methods - summer school, videos available for 2019 version ML >> Systems and Operations Stanford MLSys Seminar Series Visual Computing Systems- cs348v - Another systems course that discusses hardware from a persepective of visual computing but is relevant to ML as well Advanced Machine Learning Systems - cs6787 - lecture 9 and onwards discuss hardware side of things Machine Learning Systems Design - cs329S Topics in Deployable ML - 6.S979 Machine Learning in Production / AI Engineering (17-445/17-645/17-745/11-695) AutoML - Automated Machine Learning DL >> Deep Unsupervised Learning - cs294 Deep Multi-task and Meta learning - cs330 Topics in Deep Learning - stat991 UPenn/Wharton most chapters start with introductory topics and dig into advanced ones towards the end. Deep Generative Models - cs236 Deep Geometric Learning of Big Data and Applications Deep Implicit Layers - NeurIPS 2020 tutorial DL >> Theory Topics course on Mathematics of Deep Learning - CSCI-GA 3033 Topics Course on Deep Learning - stat212b Analyses of Deep Learning - stats385, videos from 2017 version Mathematics of Deep Learning Geometry of Deep Learning RL >> Meta-Learning - ICML 2019 Tutorial , Metalearning: Applications to Data Mining - google books link Deep Multi-Task and Meta Learning - cs330, videos Deep Reinforcement Learning - cs285 Advanced robotics - cs287 Reinforcement Learning - cs234, videos for 2019 run Reinforcement Learning Summer School 2019: Bandits, RL & Deep RL Optimization >> Convex Optimization I - ee364a, has quite recent videos too. Convex Optimization II - ee364b, 2008 videos Convex Optimization and Approximation - ee227c Convex Optimization - ee227bt Variational Methods for Computer Vision Advanced Optimization and Randomized Algorithms - 10-801, videos Optimization Methods for Machine Learning and Engineering - Karlsruhe Institute of Technology Applications >> Computer Vision Computational Video Manipulation - cs448v Advanced Topics in ML: Modeling and Segmentation of Multivariate Mixed Data TUM AI Guest lecture series - many influential researchers in DL, vision, graphics talk about latest advances and their latest works. Advanced Deep Learning for Computer Vision - TUM ADL4CV Detection, Segmentation and Tracking - TUM CV3DST Guest lectures at TUM Dynamic Vision and Learning group Vision Seminar at MIT Autonomous Vision Group, Talk@Tübingen Seminar Applications >> Natural Language Processing Natural Language Processing with Deep Learning - cs224n ( not sure if it belongs here, people working in NLP can help me out) Neural networks for NLP - cs11-747 Natural Language Understanding - cs224u, video Applications >> 3D Graphics Non-Euclidean Methods in Machine Learning - cs468, 2020 Machine Learning for 3D Data - cs468, spring 2017 Data-Driven Shape Analysis - cs468, 2014 Geometric Deep Learning - Not a course but the website links a few tutorials on Geometric DL Deep Learning for Computer Graphics - SIGGRAPH 2019 Machine Learning for Machine Vision as Inverse Graphics - csc2547 Winter'20 Machine Learning Meets Geometry, winter 2020; Machine Learning for 3D Data, winter 2018 Edit: Upon suggestion, categorized the courses. There might be some misclassifications as I'm not trained on this task ;). Added some good ones from older (linked above) discussions.

[P] The Big Sleep: Text-to-image generation using BigGAN and OpenAI's CLIP via a Google Colab notebook from Twitter user Adverb
reddit
LLM Vibe Score0
Human Vibe Score0.333
WiskkeyThis week

[P] The Big Sleep: Text-to-image generation using BigGAN and OpenAI's CLIP via a Google Colab notebook from Twitter user Adverb

From https://twitter.com/advadnoun/status/1351038053033406468: The Big Sleep Here's the notebook for generating images by using CLIP to guide BigGAN. It's very much unstable and a prototype, but it's also a fair place to start. I'll likely update it as time goes on. colab.research.google.com/drive/1NCceX2mbiKOSlAd\o7IU7nA9UskKN5WR?usp=sharing I am not the developer of The Big Sleep. This is the developer's Twitter account; this is the developer's Reddit account. Steps to follow to generate the first image in a given Google Colab session: Optionally, if this is your first time using Google Colab, view this Colab introduction and/or this Colab FAQ. Click this link. Sign into your Google account if you're not already signed in. Click the "S" button in the upper right to do this. Note: Being signed into a Google account has privacy ramifications, such as your Google search history being recorded in your Google account. In the Table of Contents, click "Parameters". Find the line that reads "tx = clip.tokenize('''a cityscape in the style of Van Gogh''')" and change the text inside of the single quote marks to your desired text; example: "tx = clip.tokenize('''a photo of New York City''')". The developer recommends that you keep the three single quote marks on both ends of your desired text so that mult-line text can be used An alternative is to remove two of the single quotes on each end of your desired text; example: "tx = clip.tokenize('a photo of New York City')". In the Table of Contents, click "Restart the kernel...". Position the pointer over the first cell in the notebook, which starts with text "import subprocess". Click the play button (the triangle) to run the cell. Wait until the cell completes execution. Click menu item "Runtime->Restart and run all". In the Table of Contents, click "Diagnostics". The output appears near the end of the Train cell that immediately precedes the Diagnostics cell, so scroll up a bit. Every few minutes (or perhaps 10 minutes if Google assigned you relatively slow hardware for this session), a new image will appear in the Train cell that is a refinement of the previous image. This process can go on for as long as you want until Google ends your Google Colab session, which is a total of up to 12 hours for the free version of Google Colab. Steps to follow if you want to start a different run using the same Google Colab session: Click menu item "Runtime->Interrupt execution". Save any images that you want to keep by right-clicking on them and using the appropriate context menu command. Optionally, change the desired text. Different runs using the same desired text almost always results in different outputs. Click menu item "Runtime->Restart and run all". Steps to follow when you're done with your Google Colab session: Click menu item "Runtime->Manage sessions". Click "Terminate" to end the session. Optionally, log out of your Google account due to the privacy ramifications of being logged into a Google account. The first output image in the Train cell (using the notebook's default of seeing every 100th image generated) usually is a very poor match to the desired text, but the second output image often is a decent match to the desired text. To change the default of seeing every 100th image generated, change the number 100 in line "if itt % 100 == 0:" in the Train cell to the desired number. For free-tier Google Colab users, I recommend changing 100 to a small integer such as 5. Tips for the text descriptions that you supply: In Section 3.1.4 of OpenAI's CLIP paper (pdf), the authors recommend using a text description of the form "A photo of a {label}." or "A photo of a {label}, a type of {type}." for images that are photographs. A Reddit user gives these tips. The Big Sleep should generate these 1,000 types of things better on average than other types of things. Here is an article containing a high-level description of how The Big Sleep works. The Big Sleep uses a modified version of BigGAN as its image generator component. The Big Sleep uses the ViT-B/32 CLIP model to rate how well a given image matches your desired text. The best CLIP model according to the CLIP paper authors is the (as of this writing) unreleased ViT-L/14-336px model; see Table 10 on page 40 of the CLIP paper (pdf) for a comparison. There are many other sites/programs/projects that use CLIP to steer image/video creation to match a text description. Some relevant subreddits: r/bigsleep (subreddit for images/videos generated from text-to-image machine learning algorithms). r/deepdream (subreddit for images/videos generated from machine learning algorithms). r/mediasynthesis (subreddit for media generation/manipulation techniques that use artificial intelligence; this subreddit shouldn't be used to post images/videos unless new techniques are demonstrated, or the images/videos are of high quality relative to other posts). Example using text 'a black cat sleeping on top of a red clock': https://preview.redd.it/7xq58v7022c61.png?width=512&format=png&auto=webp&s=a229ae9add555cd1caba31c42b60d907ffe67773 Example using text 'the word ''hot'' covered in ice': https://preview.redd.it/6kxdp8u3k2c61.png?width=512&format=png&auto=webp&s=5bd078b0111575f5d88a1dc53b0aeb933f3b0da6 Example using text 'a monkey holding a green lightsaber': https://preview.redd.it/rdsybsoaz2c61.png?width=512&format=png&auto=webp&s=2769d4c6c883c1c35ae0b1c629bebe9bc1d41393 Example using text 'The White House in Washington D.C. at night with green and red spotlights shining on it': https://preview.redd.it/w4mg90xsf5c61.png?width=512&format=png&auto=webp&s=5f18318de2f77bcd8a86e71e87048fadd30383d1 Example using text '''A photo of the Golden Gate Bridge at night, illuminated by spotlights in a tribute to Prince''': https://preview.redd.it/cn4ecuafhic61.png?width=512&format=png&auto=webp&s=397c838fdc49f13c5f17110b92c78b95bf0dcac0 Example using text '''a Rembrandt-style painting titled "Robert Plant decides whether to take the stairway to heaven or the ladder to heaven"''': https://preview.redd.it/h7rb3y6j5jc61.png?width=512&format=png&auto=webp&s=537bfe8210af185647b00e7585c948aa2c4e0ffb Example using text '''A photo of the Empire State Building being shot at with the laser cannons of a TIE fighter.''': https://preview.redd.it/cwi7i639c5d61.png?width=512&format=png&auto=webp&s=0510c8b93adb40eee4d3f41607f1c215d41e55ff Example using text '''A cartoon of a new mascot for the Reddit subreddit DeepDream that has a mouse-like face and wears a cape''': https://preview.redd.it/wtxbduevcbd61.png?width=512&format=png&auto=webp&s=c5d266258922bc62f25c80a08cd9cabc07d9cb1c Example using text '''Bugs Bunny meets the Eye of Sauron, drawn in the Looney Tunes cartoon style''': https://preview.redd.it/gmljaeekuid61.png?width=512&format=png&auto=webp&s=9ea578de165e12afc3a62bf6886bc1ae9dc19bec Example using text '''Photo of a blue and red neon-colored frog at night.''': https://preview.redd.it/nzlypte6wzd61.png?width=512&format=png&auto=webp&s=7e10b06f22cfc57c64b6d05738c7486b895083df Example using text '''Hell begins to freeze over''': https://preview.redd.it/vn99we9ngmf61.png?width=512&format=png&auto=webp&s=2408efd607f0ab40a08db6ee67448791aa813993 Example using text '''A scene with vibrant colors''': https://preview.redd.it/4z133mvrgmf61.png?width=512&format=png&auto=webp&s=b78e7a8e3f736769655056093a9904ff09a355a1 Example using text '''The Great Pyramids were turned into prisms by a wizard''': https://preview.redd.it/zxt6op7vgmf61.png?width=512&format=png&auto=webp&s=53e578cfde14b28afe27957e95e610b89afadd44

[D] I don't really trust papers out of "Top Labs" anymore
reddit
LLM Vibe Score0
Human Vibe Score0.333
MrAcuriteThis week

[D] I don't really trust papers out of "Top Labs" anymore

I mean, I trust that the numbers they got are accurate and that they really did the work and got the results. I believe those. It's just that, take the recent "An Evolutionary Approach to Dynamic Introduction of Tasks in Large-scale Multitask Learning Systems" paper. It's 18 pages of talking through this pretty convoluted evolutionary and multitask learning algorithm, it's pretty interesting, solves a bunch of problems. But two notes. One, the big number they cite as the success metric is 99.43 on CIFAR-10, against a SotA of 99.40, so woop-de-fucking-doo in the grand scheme of things. Two, there's a chart towards the end of the paper that details how many TPU core-hours were used for just the training regimens that results in the final results. The sum total is 17,810 core-hours. Let's assume that for someone who doesn't work at Google, you'd have to use on-demand pricing of $3.22/hr. This means that these trained models cost $57,348. Strictly speaking, throwing enough compute at a general enough genetic algorithm will eventually produce arbitrarily good performance, so while you can absolutely read this paper and collect interesting ideas about how to use genetic algorithms to accomplish multitask learning by having each new task leverage learned weights from previous tasks by defining modifications to a subset of components of a pre-existing model, there's a meta-textual level on which this paper is just "Jeff Dean spent enough money to feed a family of four for half a decade to get a 0.03% improvement on CIFAR-10." OpenAI is far and away the worst offender here, but it seems like everyone's doing it. You throw a fuckton of compute and a light ganache of new ideas at an existing problem with existing data and existing benchmarks, and then if your numbers are infinitesimally higher than their numbers, you get to put a lil' sticker on your CV. Why should I trust that your ideas are even any good? I can't check them, I can't apply them to my own projects. Is this really what we're comfortable with as a community? A handful of corporations and the occasional university waving their dicks at everyone because they've got the compute to burn and we don't? There's a level at which I think there should be a new journal, exclusively for papers in which you can replicate their experimental results in under eight hours on a single consumer GPU.

[D] What is your honest experience with reinforcement learning?
reddit
LLM Vibe Score0
Human Vibe Score1
Starks-TechnologyThis week

[D] What is your honest experience with reinforcement learning?

In my personal experience, SOTA RL algorithms simply don't work. I've tried working with reinforcement learning for over 5 years. I remember when Alpha Go defeated the world famous Go player, Lee Sedol, and everybody thought RL would take the ML community by storm. Yet, outside of toy problems, I've personally never found a practical use-case of RL. What is your experience with it? Aside from Ad recommendation systems and RLHF, are there legitimate use-cases of RL? Or, was it all hype? Edit: I know a lot about AI. I built NexusTrade, an AI-Powered automated investing tool that lets non-technical users create, update, and deploy their trading strategies. I’m not an idiot nor a noob; RL is just ridiculously hard. Edit 2: Since my comments are being downvoted, here is a link to my article that better describes my position. It's not that I don't understand RL. I released my open-source code and wrote a paper on it. It's the fact that it's EXTREMELY difficult to understand. Other deep learning algorithms like CNNs (including ResNets), RNNs (including GRUs and LSTMs), Transformers, and GANs are not hard to understand. These algorithms work and have practical use-cases outside of the lab. Traditional SOTA RL algorithms like PPO, DDPG, and TD3 are just very hard. You need to do a bunch of research to even implement a toy problem. In contrast, the decision transformer is something anybody can implement, and it seems to match or surpass the SOTA. You don't need two networks battling each other. You don't have to go through hell to debug your network. It just naturally learns the best set of actions in an auto-regressive manner. I also didn't mean to come off as arrogant or imply that RL is not worth learning. I just haven't seen any real-world, practical use-cases of it. I simply wanted to start a discussion, not claim that I know everything. Edit 3: There's a shockingly number of people calling me an idiot for not fully understanding RL. You guys are wayyy too comfortable calling people you disagree with names. News-flash, not everybody has a PhD in ML. My undergraduate degree is in biology. I self-taught myself the high-level maths to understand ML. I'm very passionate about the field; I just have VERY disappointing experiences with RL. Funny enough, there are very few people refuting my actual points. To summarize: Lack of real-world applications Extremely complex and inaccessible to 99% of the population Much harder than traditional DL algorithms like CNNs, RNNs, and GANs Sample inefficiency and instability Difficult to debug Better alternatives, such as the Decision Transformer Are these not legitimate criticisms? Is the purpose of this sub not to have discussions related to Machine Learning? To the few commenters that aren't calling me an idiot...thank you! Remember, it costs you nothing to be nice! Edit 4: Lots of people seem to agree that RL is over-hyped. Unfortunately those comments are downvoted. To clear up some things: We've invested HEAVILY into reinforcement learning. All we got from this investment is a robot that can be super-human at (some) video games. AlphaFold did not use any reinforcement learning. SpaceX doesn't either. I concede that it can be useful for robotics, but still argue that it's use-cases outside the lab are extremely limited. If you're stumbling on this thread and curious about an RL alternative, check out the Decision Transformer. It can be used in any situation that a traditional RL algorithm can be used. Final Edit: To those who contributed more recently, thank you for the thoughtful discussion! From what I learned, model-based models like Dreamer and IRIS MIGHT have a future. But everybody who has actually used model-free models like DDPG unanimously agree that they suck and don’t work.

[R] Reinforcement Learning for Sequential Decision and Optimal Control
reddit
LLM Vibe Score0
Human Vibe Score1
isfjzzzThis week

[R] Reinforcement Learning for Sequential Decision and Optimal Control

Since early 21st century, artificial intelligence (AI) has been reshaping almost all areas of human society, which has high potential to spark the fourth industrial revolution. Notable examples can be found in the sector of road transportation, where AI has drastically changed automobile design and traffic management. Many new technologies, such as driver assistance, autonomous driving, and cloud-based cooperation, are emerging at an unbelievable speed. These new technologies have the potential to significantly improve driving ability, reduce traffic accidents, and relieve urban congestion. As one of the most important AI branches, reinforcement learning (RL) has attracted increasing attention in recent years. RL is an interdisciplinary field of trial-and-error learning and optimal control, which promises to provide optimal solutions for decision-making or control in large-scale and complex dynamic processes. One of its most conspicuous successes is AlphaGo from Google DeepMind, which has beaten the highest-level professional human player. The underlying key technology is the so-called deep reinforcement learning, which equips AlphaGo with amazing self-evolution ability and high playing intelligence. Despite a few successes, the application of RL is still in its infancy because most RL algorithms are rather difficult to comprehend and implement. RL connects deeply with statistical learning and convex optimization, and involves a wide range of new concepts and theories. As a beginner, one must undergo a long and tedious learning process to become an RL master. Without fully understanding those underlying principles, it is very difficult for new users to make proper adjustments to achieve the best application performance. ​ https://preview.redd.it/tggt6o3o481c1.jpg?width=248&format=pjpg&auto=webp&s=75e2b58ac8da9273f2511a4fe37ef508d86a6e96 Reference: Shengbo Eben Li, Reinforcement Learning for Sequential Decision and Optimal Control. Springer Verlag, Singapore, 2023 Website of e-book: https://link.springer.com/book/10.1007/978-981-19-7784-8 ​ QR code to Springer Book contents This book aims to provide a systematic introduction to fundamental RL theories, mainstream RL algorithms and typical RL applications for researchers and engineers. The main topics include Markov decision processes, Monte Carlo learning, temporal difference learning, RL with function approximation, policy gradient method, approximate dynamic programming, deep reinforcement learning, etc. Chapter 1 provides an overview of RL, including its history, famous scholars, successful examples and up-to-date challenges. Chapter 2 discusses the basis of RL, including main concepts and terminologies, Bellman’s optimality condition, and general problem formulation. Chapter 3 introduces Monte Carlo learning methods for model-free RL, including on-policy/off-policy methods and importance sampling technique. Chapter 4 introduces temporal difference learning methods for model-free RL, including Sarsa, Q-learning, and expected Sarsa. Chapter 5 introduces stochastic dynamic programming (DP), i.e., model-based RL with tabular representation, including value iteration DP, policy iteration DP and their convergence mechanisms. Chapter 6 introduces how to approximate policy and value functions in indirect RL methods as well as the associated actor-critic architecture. Chapter 7 derives different kinds of direct policy gradients, including likelihood ratio gradient, natural policy gradient and a few advanced variants. Chapter 8 introduces infinite-horizon ADP, finite-horizon ADP and its connection with model predictive control. Chapter 9 discusses how to handle state constraints and its connection with feasibility and safety, as well as the newly proposed actor-critic-scenery learning architecture. Chapter 10 is devoted to deep reinforcement learning, including how to train artificial neural networks and typical deep RL algorithms such as DQN, DDPG, TD3, TRPO, PPO, SAC, and DSAC. Chapter 11 provides various RL topics,including robust RL, POMDP, multi-agent RL, meta-RL, inverse RL, offline RL, major RL libraries and platforms. Author information: Shengbo Eben Li is currently a professor at Tsinghua University in the interdisciplinary field of autonomous driving and artificial intelligence. Before joining Tsinghua University, he has worked at Stanford University, University of Michigan, and UC Berkeley. His active research interests include intelligent vehicles and driver assistance, deep reinforcement learning, optimal control and estimation, etc. He has published more than 130 peer-reviewed papers in top-tier international journals and conferences. He is the recipient of best paper awards (finalists) of IEEE ITSC, ICCAS, IEEE ICUS, IEEE IV, L4DC, etc. He has received a number of important academic honors, including National Award for Technological Invention of China (2013), National Award for Progress in Sci & Tech of China (2018), Distinguished Young Scholar of Beijing NSF (2018), Youth Sci & Tech Innovation Leader from MOST China (2020), etc. He also serves as Board of Governor of IEEE ITS Society, Senior AE of IEEE OJ ITS, and AEs of IEEE ITSM, IEEE Trans ITS, Automotive Innovation, etc.

[D] What are some good advanced platforms?
reddit
LLM Vibe Score0
Human Vibe Score1
SemperZeroThis week

[D] What are some good advanced platforms?

Hey. I'm 27 and I think I got most of the basics for ML. I'm very good at math, I understand statistics and probability quite deep, worked on research projects by myself, for which I had to build models on my own. Not really complex, but still requiring creativity and a good understanding of basic concepts. I will soon start a data science job at a FAANG company and I want to further improve my skills and use their resources to the fullest, but I'm not really sure where to go from here in terms of learning. Could you help me with some more advanced materials/forums for ML research/place with good papers/place with good articles? I'd also like to study the very best and see the way they code and explain advanced concepts (like Andrej Karpathy) where can I find them?? is there a Twitch for challenger level AI researchers streaming live processes? Or videos showing the entire project flow (how they do data visualizations, mining, choosing models, tuning, etc) like top digital artists show the highlights or the entire speed-up of their painting processes? Here's a list all of my projects to get a general idea of my level and where I'm at: calculating the distance between hundreds of 42.000 feature objects (containing categorical, strings, numbers, hashes, booleans as variables) and then clustering. with some vector processing and a neural network implemented from scratch in C some models like ARIMA (together with linear regression) combining a FFT with a neural network for a 42d wave classification T-SNE to split dataset into 2d grids -> Kullback–Leibler on grids for distance -> DBSCAN/KMEANS for clustering genetic algorithms for hyperparameter optimizations and reinforcement learning (neuro evolution) DBSCAN -> Levenberg-Marquardt for polynomial coefficients-> neural network predicting the coefficients based on different parameters playing with instance segmentation and some algorithms to synchronize a color and a depth camera simulations/statistics/probabilities for video games a lot of visualizations and data mining for patterns As you can see there is no LLM/ Generative AI/ Computer Vision stuff, which I would like to get into. I'm also not 100% sure what else would be nice to learn in general. I know most of the basic procedures for training, balancing datasets, avoid overfit, computing error plots, comparing models, etc and I'm familiar with most of math (not insanely advanced) used in ML. I didn't read many papers, but holy ... most of them are so unreadable and filled with pompous nonsense that 99% of the effort is de-obfuscating the bs and reading for so long just to figure out how the input is encoded, what's the output, and what's the model. Where can I find good, readable, structured papers which are actually on point? I'm from Eastern Europe and most of my learning has been done by my self after high school, the education quality is close to zero in the universities here and I never had any mentors at the jobs I worked. There's no research in this country, and getting to work on these projects was insanely hard, some of them being done in my free time or for free just to get experience... Fortunately after a lot of hard work I got into FAANG, and I hope things will be better here. Most of what I've learned has been from very fragmented places on the internet, and now I'm looking for centralized places and communities of top quality content. TL;DR: sorry for the long rambling. had to order my thoughts and figure what i actually want: Looking for top tier AI researchers showcasing their work processes, places with clear papers/articles, tips for someone who's no longer a very beginner, and other communities like this.

[D] What is your honest experience with reinforcement learning?
reddit
LLM Vibe Score0
Human Vibe Score1
Starks-TechnologyThis week

[D] What is your honest experience with reinforcement learning?

In my personal experience, SOTA RL algorithms simply don't work. I've tried working with reinforcement learning for over 5 years. I remember when Alpha Go defeated the world famous Go player, Lee Sedol, and everybody thought RL would take the ML community by storm. Yet, outside of toy problems, I've personally never found a practical use-case of RL. What is your experience with it? Aside from Ad recommendation systems and RLHF, are there legitimate use-cases of RL? Or, was it all hype? Edit: I know a lot about AI. I built NexusTrade, an AI-Powered automated investing tool that lets non-technical users create, update, and deploy their trading strategies. I’m not an idiot nor a noob; RL is just ridiculously hard. Edit 2: Since my comments are being downvoted, here is a link to my article that better describes my position. It's not that I don't understand RL. I released my open-source code and wrote a paper on it. It's the fact that it's EXTREMELY difficult to understand. Other deep learning algorithms like CNNs (including ResNets), RNNs (including GRUs and LSTMs), Transformers, and GANs are not hard to understand. These algorithms work and have practical use-cases outside of the lab. Traditional SOTA RL algorithms like PPO, DDPG, and TD3 are just very hard. You need to do a bunch of research to even implement a toy problem. In contrast, the decision transformer is something anybody can implement, and it seems to match or surpass the SOTA. You don't need two networks battling each other. You don't have to go through hell to debug your network. It just naturally learns the best set of actions in an auto-regressive manner. I also didn't mean to come off as arrogant or imply that RL is not worth learning. I just haven't seen any real-world, practical use-cases of it. I simply wanted to start a discussion, not claim that I know everything. Edit 3: There's a shockingly number of people calling me an idiot for not fully understanding RL. You guys are wayyy too comfortable calling people you disagree with names. News-flash, not everybody has a PhD in ML. My undergraduate degree is in biology. I self-taught myself the high-level maths to understand ML. I'm very passionate about the field; I just have VERY disappointing experiences with RL. Funny enough, there are very few people refuting my actual points. To summarize: Lack of real-world applications Extremely complex and inaccessible to 99% of the population Much harder than traditional DL algorithms like CNNs, RNNs, and GANs Sample inefficiency and instability Difficult to debug Better alternatives, such as the Decision Transformer Are these not legitimate criticisms? Is the purpose of this sub not to have discussions related to Machine Learning? To the few commenters that aren't calling me an idiot...thank you! Remember, it costs you nothing to be nice! Edit 4: Lots of people seem to agree that RL is over-hyped. Unfortunately those comments are downvoted. To clear up some things: We've invested HEAVILY into reinforcement learning. All we got from this investment is a robot that can be super-human at (some) video games. AlphaFold did not use any reinforcement learning. SpaceX doesn't either. I concede that it can be useful for robotics, but still argue that it's use-cases outside the lab are extremely limited. If you're stumbling on this thread and curious about an RL alternative, check out the Decision Transformer. It can be used in any situation that a traditional RL algorithm can be used. Final Edit: To those who contributed more recently, thank you for the thoughtful discussion! From what I learned, model-based models like Dreamer and IRIS MIGHT have a future. But everybody who has actually used model-free models like DDPG unanimously agree that they suck and don’t work.

[D] Do you know any institutions/nonprofits/companies/governments/etc. trying to apply deep learning and other ML/AI/GenAI techniques to implement universal basic income (UBI) or something similar to UBI like universal basic services?
reddit
LLM Vibe Score0
Human Vibe Score1
HappyseditsThis week

[D] Do you know any institutions/nonprofits/companies/governments/etc. trying to apply deep learning and other ML/AI/GenAI techniques to implement universal basic income (UBI) or something similar to UBI like universal basic services?

Do you know any institutions/nonprofits/companies/governments/etc. trying to apply deep learning and other ML/AI/GenAI techniques to implement universal basic income (UBI) or something similar to UBI like universal basic services? Maybe for chatbot guidance on UBI program details, selecting candidates that need it the most, predicting poverty, UBI impacts, demographic and economic indicators to identify optimal UBI payment amounts and frequencies for different population segments, preventing fraud, etc. It can be just sketching future models in theory, or already implementing it in practice. I found this relevant paper: Can Data and Machine Learning Change the Future of Basic Income Models? A Bayesian Belief Networks Approach. https://www.mdpi.com/2306-5729/9/2/18 "Appeals to governments for implementing basic income are contemporary. The theoretical backgrounds of the basic income notion only prescribe transferring equal amounts to individuals irrespective of their specific attributes. However, the most recent basic income initiatives all around the world are attached to certain rules with regard to the attributes of the households. This approach is facing significant challenges to appropriately recognize vulnerable groups. A possible alternative for setting rules with regard to the welfare attributes of the households is to employ artificial intelligence algorithms that can process unprecedented amounts of data. Can integrating machine learning change the future of basic income by predicting households vulnerable to future poverty? In this paper, we utilize multidimensional and longitudinal welfare data comprising one and a half million individuals’ data and a Bayesian beliefs network approach to examine the feasibility of predicting households’ vulnerability to future poverty based on the existing households’ welfare attributes."

[D] The machine learning community has a toxicity problem
reddit
LLM Vibe Score0
Human Vibe Score1
yusuf-bengioThis week

[D] The machine learning community has a toxicity problem

It is omnipresent! First of all, the peer-review process is broken. Every fourth NeurIPS submission is put on arXiv. There are DeepMind researchers publicly going after reviewers who are criticizing their ICLR submission. On top of that, papers by well-known institutes that were put on arXiv are accepted at top conferences, despite the reviewers agreeing on rejection. In contrast, vice versa, some papers with a majority of accepts are overruled by the AC. (I don't want to call any names, just have a look the openreview page of this year's ICRL). Secondly, there is a reproducibility crisis. Tuning hyperparameters on the test set seem to be the standard practice nowadays. Papers that do not beat the current state-of-the-art method have a zero chance of getting accepted at a good conference. As a result, hyperparameters get tuned and subtle tricks implemented to observe a gain in performance where there isn't any. Thirdly, there is a worshiping problem. Every paper with a Stanford or DeepMind affiliation gets praised like a breakthrough. For instance, BERT has seven times more citations than ULMfit. The Google affiliation gives so much credibility and visibility to a paper. At every ICML conference, there is a crowd of people in front of every DeepMind poster, regardless of the content of the work. The same story happened with the Zoom meetings at the virtual ICLR 2020. Moreover, NeurIPS 2020 had twice as many submissions as ICML, even though both are top-tier ML conferences. Why? Why is the name "neural" praised so much? Next, Bengio, Hinton, and LeCun are truly deep learning pioneers but calling them the "godfathers" of AI is insane. It has reached the level of a cult. Fourthly, the way Yann LeCun talked about biases and fairness topics was insensitive. However, the toxicity and backlash that he received are beyond any reasonable quantity. Getting rid of LeCun and silencing people won't solve any issue. Fifthly, machine learning, and computer science in general, have a huge diversity problem. At our CS faculty, only 30% of undergrads and 15% of the professors are women. Going on parental leave during a PhD or post-doc usually means the end of an academic career. However, this lack of diversity is often abused as an excuse to shield certain people from any form of criticism. Reducing every negative comment in a scientific discussion to race and gender creates a toxic environment. People are becoming afraid to engage in fear of being called a racist or sexist, which in turn reinforces the diversity problem. Sixthly, moral and ethics are set arbitrarily. The U.S. domestic politics dominate every discussion. At this very moment, thousands of Uyghurs are put into concentration camps based on computer vision algorithms invented by this community, and nobody seems even remotely to care. Adding a "broader impact" section at the end of every people will not make this stop. There are huge shitstorms because a researcher wasn't mentioned in an article. Meanwhile, the 1-billion+ people continent of Africa is virtually excluded from any meaningful ML discussion (besides a few Indaba workshops). Seventhly, there is a cut-throat publish-or-perish mentality. If you don't publish 5+ NeurIPS/ICML papers per year, you are a looser. Research groups have become so large that the PI does not even know the name of every PhD student anymore. Certain people submit 50+ papers per year to NeurIPS. The sole purpose of writing a paper has become to having one more NeurIPS paper in your CV. Quality is secondary; passing the peer-preview stage has become the primary objective. Finally, discussions have become disrespectful. Schmidhuber calls Hinton a thief, Gebru calls LeCun a white supremacist, Anandkumar calls Marcus a sexist, everybody is under attack, but nothing is improved. Albert Einstein was opposing the theory of quantum mechanics. Can we please stop demonizing those who do not share our exact views. We are allowed to disagree without going for the jugular. The moment we start silencing people because of their opinion is the moment scientific and societal progress dies. Best intentions, Yusuf

5-Day Applied Rationality Workshop for Machine Learning Students & Researchers
reddit
LLM Vibe Score0
Human Vibe Score1
AnnaSalamonThis week

5-Day Applied Rationality Workshop for Machine Learning Students & Researchers

The Center for Applied Rationality is a Berkeley-based nonprofit that runs immersive workshops for entrepreneurs, researchers, students, and other ambitious, analytical, practically-minded people. The practice of “applied rationality”, which the workshops aim towards, involves noticing what cognitive algorithms you seem to be running, checking whether those algorithms seem to be helping you form accurate beliefs and achieve your goals, and looking for ways to improve them. A typical 4-day CFAR workshop costs $3900 to attend, but thanks to a generous grant from the Future of Life Institute this fall we will be running a free five-day workshop for students and researchers in the fields of machine learning and artificial intelligence. All costs are covered by this grant, including room, board, and flights. The workshop will take place this Aug 30 through Sep 4 in the San Francisco Bay Area and will include: 2 days focused on learning models and skills, such as how habits develop and how to redesign your habits. 2 days focused on practicing skills and applying them to whichever areas of your life you would like to make improvements on, such as how to make faster progress on projects or how to have more productive collaborations with colleagues. 1 day (special to this workshop) focused on discussion of the long-term impact of artificial intelligence, and on what reasoning habits — if spread across the relevant research communities — may increase the probability of positive long-term AI outcomes. Go here to read more or to apply, or ask questions here.

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.
reddit
LLM Vibe Score0
Human Vibe Score0.765
hardmaruThis week

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.

Schmidhuber interview expressing his views on the future of AI and AGI. Original source. I think the interview is of interest to r/MachineLearning, and presents an alternate view, compared to other influential leaders in AI. Juergen Schmidhuber, Renowned 'Father Of Modern AI,' Says His Life’s Work Won't Lead To Dystopia May 23, 2023. Contributed by Hessie Jones. Amid the growing concern about the impact of more advanced artificial intelligence (AI) technologies on society, there are many in the technology community who fear the implications of the advancements in Generative AI if they go unchecked. Dr. Juergen Schmidhuber, a renowned scientist, artificial intelligence researcher and widely regarded as one of the pioneers in the field, is more optimistic. He declares that many of those who suddenly warn against the dangers of AI are just seeking publicity, exploiting the media’s obsession with killer robots which has attracted more attention than “good AI” for healthcare etc. The potential to revolutionize various industries and improve our lives is clear, as are the equal dangers if bad actors leverage the technology for personal gain. Are we headed towards a dystopian future, or is there reason to be optimistic? I had a chance to sit down with Dr. Juergen Schmidhuber to understand his perspective on this seemingly fast-moving AI-train that will leap us into the future. As a teenager in the 1970s, Juergen Schmidhuber became fascinated with the idea of creating intelligent machines that could learn and improve on their own, becoming smarter than himself within his lifetime. This would ultimately lead to his groundbreaking work in the field of deep learning. In the 1980s, he studied computer science at the Technical University of Munich (TUM), where he earned his diploma in 1987. His thesis was on the ultimate self-improving machines that, not only, learn through some pre-wired human-designed learning algorithm, but also learn and improve the learning algorithm itself. Decades later, this became a hot topic. He also received his Ph.D. at TUM in 1991 for work that laid some of the foundations of modern AI. Schmidhuber is best known for his contributions to the development of recurrent neural networks (RNNs), the most powerful type of artificial neural network that can process sequential data such as speech and natural language. With his students Sepp Hochreiter, Felix Gers, Alex Graves, Daan Wierstra, and others, he published architectures and training algorithms for the long short-term memory (LSTM), a type of RNN that is widely used in natural language processing, speech recognition, video games, robotics, and other applications. LSTM has become the most cited neural network of the 20th century, and Business Week called it "arguably the most commercial AI achievement." Throughout his career, Schmidhuber has received various awards and accolades for his groundbreaking work. In 2013, he was awarded the Helmholtz Prize, which recognizes significant contributions to the field of machine learning. In 2016, he was awarded the IEEE Neural Network Pioneer Award for "pioneering contributions to deep learning and neural networks." The media have often called him the “father of modern AI,” because the most cited neural networks all build on his lab’s work. He is quick to point out, however, that AI history goes back centuries. Despite his many accomplishments, at the age of 60, he feels mounting time pressure towards building an Artificial General Intelligence within his lifetime and remains committed to pushing the boundaries of AI research and development. He is currently director of the KAUST AI Initiative, scientific director of the Swiss AI Lab IDSIA, and co-founder and chief scientist of AI company NNAISENSE, whose motto is "AI∀" which is a math-inspired way of saying "AI For All." He continues to work on cutting-edge AI technologies and applications to improve human health and extend human lives and make lives easier for everyone. The following interview has been edited for clarity. Jones: Thank you Juergen for joining me. You have signed letters warning about AI weapons. But you didn't sign the recent publication, "Pause Gigantic AI Experiments: An Open Letter"? Is there a reason? Schmidhuber: Thank you Hessie. Glad to speak with you. I have realized that many of those who warn in public against the dangers of AI are just seeking publicity. I don't think the latest letter will have any significant impact because many AI researchers, companies, and governments will ignore it completely. The proposal frequently uses the word "we" and refers to "us," the humans. But as I have pointed out many times in the past, there is no "we" that everyone can identify with. Ask 10 different people, and you will hear 10 different opinions about what is "good." Some of those opinions will be completely incompatible with each other. Don't forget the enormous amount of conflict between the many people. The letter also says, "If such a pause cannot be quickly put in place, governments should intervene and impose a moratorium." The problem is that different governments have ALSO different opinions about what is good for them and for others. Great Power A will say, if we don't do it, Great Power B will, perhaps secretly, and gain an advantage over us. The same is true for Great Powers C and D. Jones: Everyone acknowledges this fear surrounding current generative AI technology. Moreover, the existential threat of this technology has been publicly acknowledged by Sam Altman, CEO of OpenAI himself, calling for AI regulation. From your perspective, is there an existential threat? Schmidhuber: It is true that AI can be weaponized, and I have no doubt that there will be all kinds of AI arms races, but AI does not introduce a new quality of existential threat. The threat coming from AI weapons seems to pale in comparison to the much older threat from nuclear hydrogen bombs that don’t need AI at all. We should be much more afraid of half-century-old tech in the form of H-bomb rockets. The Tsar Bomba of 1961 had almost 15 times more destructive power than all weapons of WW-II combined. Despite the dramatic nuclear disarmament since the 1980s, there are still more than enough nuclear warheads to wipe out human civilization within two hours, without any AI I’m much more worried about that old existential threat than the rather harmless AI weapons. Jones: I realize that while you compare AI to the threat of nuclear bombs, there is a current danger that a current technology can be put in the hands of humans and enable them to “eventually” exact further harms to individuals of group in a very precise way, like targeted drone attacks. You are giving people a toolset that they've never had before, enabling bad actors, as some have pointed out, to be able to do a lot more than previously because they didn't have this technology. Schmidhuber: Now, all that sounds horrible in principle, but our existing laws are sufficient to deal with these new types of weapons enabled by AI. If you kill someone with a gun, you will go to jail. Same if you kill someone with one of these drones. Law enforcement will get better at understanding new threats and new weapons and will respond with better technology to combat these threats. Enabling drones to target persons from a distance in a way that requires some tracking and some intelligence to perform, which has traditionally been performed by skilled humans, to me, it seems is just an improved version of a traditional weapon, like a gun, which is, you know, a little bit smarter than the old guns. But, in principle, all of that is not a new development. For many centuries, we have had the evolution of better weaponry and deadlier poisons and so on, and law enforcement has evolved their policies to react to these threats over time. So, it's not that we suddenly have a new quality of existential threat and it's much more worrisome than what we have had for about six decades. A large nuclear warhead doesn’t need fancy face recognition to kill an individual. No, it simply wipes out an entire city with ten million inhabitants. Jones: The existential threat that’s implied is the extent to which humans have control over this technology. We see some early cases of opportunism which, as you say, tends to get more media attention than positive breakthroughs. But you’re implying that this will all balance out? Schmidhuber: Historically, we have a long tradition of technological breakthroughs that led to advancements in weapons for the purpose of defense but also for protection. From sticks, to rocks, to axes to gunpowder to cannons to rockets… and now to drones… this has had a drastic influence on human history but what has been consistent throughout history is that those who are using technology to achieve their own ends are themselves, facing the same technology because the opposing side is learning to use it against them. And that's what has been repeated in thousands of years of human history and it will continue. I don't see the new AI arms race as something that is remotely as existential a threat as the good old nuclear warheads. You said something important, in that some people prefer to talk about the downsides rather than the benefits of this technology, but that's misleading, because 95% of all AI research and AI development is about making people happier and advancing human life and health. Jones: Let’s touch on some of those beneficial advances in AI research that have been able to radically change present day methods and achieve breakthroughs. Schmidhuber: All right! For example, eleven years ago, our team with my postdoc Dan Ciresan was the first to win a medical imaging competition through deep learning. We analyzed female breast cells with the objective to determine harmless cells vs. those in the pre-cancer stage. Typically, a trained oncologist needs a long time to make these determinations. Our team, who knew nothing about cancer, were able to train an artificial neural network, which was totally dumb in the beginning, on lots of this kind of data. It was able to outperform all the other methods. Today, this is being used not only for breast cancer, but also for radiology and detecting plaque in arteries, and many other things. Some of the neural networks that we have developed in the last 3 decades are now prevalent across thousands of healthcare applications, detecting Diabetes and Covid-19 and what not. This will eventually permeate across all healthcare. The good consequences of this type of AI are much more important than the click-bait new ways of conducting crimes with AI. Jones: Adoption is a product of reinforced outcomes. The massive scale of adoption either leads us to believe that people have been led astray, or conversely, technology is having a positive effect on people’s lives. Schmidhuber: The latter is the likely case. There's intense commercial pressure towards good AI rather than bad AI because companies want to sell you something, and you are going to buy only stuff you think is going to be good for you. So already just through this simple, commercial pressure, you have a tremendous bias towards good AI rather than bad AI. However, doomsday scenarios like in Schwarzenegger movies grab more attention than documentaries on AI that improve people’s lives. Jones: I would argue that people are drawn to good stories – narratives that contain an adversary and struggle, but in the end, have happy endings. And this is consistent with your comment on human nature and how history, despite its tendency for violence and destruction of humanity, somehow tends to correct itself. Let’s take the example of a technology, which you are aware – GANs – General Adversarial Networks, which today has been used in applications for fake news and disinformation. In actuality, the purpose in the invention of GANs was far from what it is used for today. Schmidhuber: Yes, the name GANs was created in 2014 but we had the basic principle already in the early 1990s. More than 30 years ago, I called it artificial curiosity. It's a very simple way of injecting creativity into a little two network system. This creative AI is not just trying to slavishly imitate humans. Rather, it’s inventing its own goals. Let me explain: You have two networks. One network is producing outputs that could be anything, any action. Then the second network is looking at these actions and it’s trying to predict the consequences of these actions. An action could move a robot, then something happens, and the other network is just trying to predict what will happen. Now we can implement artificial curiosity by reducing the prediction error of the second network, which, at the same time, is the reward of the first network. The first network wants to maximize its reward and so it will invent actions that will lead to situations that will surprise the second network, which it has not yet learned to predict well. In the case where the outputs are fake images, the first network will try to generate images that are good enough to fool the second network, which will attempt to predict the reaction of the environment: fake or real image, and it will try to become better at it. The first network will continue to also improve at generating images whose type the second network will not be able to predict. So, they fight each other. The 2nd network will continue to reduce its prediction error, while the 1st network will attempt to maximize it. Through this zero-sum game the first network gets better and better at producing these convincing fake outputs which look almost realistic. So, once you have an interesting set of images by Vincent Van Gogh, you can generate new images that leverage his style, without the original artist having ever produced the artwork himself. Jones: I see how the Van Gogh example can be applied in an education setting and there are countless examples of artists mimicking styles from famous painters but image generation from this instance that can happen within seconds is quite another feat. And you know this is how GANs has been used. What’s more prevalent today is a socialized enablement of generating images or information to intentionally fool people. It also surfaces new harms that deal with the threat to intellectual property and copyright, where laws have yet to account for. And from your perspective this was not the intention when the model was conceived. What was your motivation in your early conception of what is now GANs? Schmidhuber: My old motivation for GANs was actually very important and it was not to create deepfakes or fake news but to enable AIs to be curious and invent their own goals, to make them explore their environment and make them creative. Suppose you have a robot that executes one action, then something happens, then it executes another action, and so on, because it wants to achieve certain goals in the environment. For example, when the battery is low, this will trigger “pain” through hunger sensors, so it wants to go to the charging station, without running into obstacles, which will trigger other pain sensors. It will seek to minimize pain (encoded through numbers). Now the robot has a friend, the second network, which is a world model ––it’s a prediction machine that learns to predict the consequences of the robot’s actions. Once the robot has a good model of the world, it can use it for planning. It can be used as a simulation of the real world. And then it can determine what is a good action sequence. If the robot imagines this sequence of actions, the model will predict a lot of pain, which it wants to avoid. If it plays this alternative action sequence in its mental model of the world, then it will predict a rewarding situation where it’s going to sit on the charging station and its battery is going to load again. So, it'll prefer to execute the latter action sequence. In the beginning, however, the model of the world knows nothing, so how can we motivate the first network to generate experiments that lead to data that helps the world model learn something it didn’t already know? That’s what artificial curiosity is about. The dueling two network systems effectively explore uncharted environments by creating experiments so that over time the curious AI gets a better sense of how the environment works. This can be applied to all kinds of environments, and has medical applications. Jones: Let’s talk about the future. You have said, “Traditional humans won’t play a significant role in spreading intelligence across the universe.” Schmidhuber: Let’s first conceptually separate two types of AIs. The first type of AI are tools directed by humans. They are trained to do specific things like accurately detect diabetes or heart disease and prevent attacks before they happen. In these cases, the goal is coming from the human. More interesting AIs are setting their own goals. They are inventing their own experiments and learning from them. Their horizons expand and eventually they become more and more general problem solvers in the real world. They are not controlled by their parents, but much of what they learn is through self-invented experiments. A robot, for example, is rotating a toy, and as it is doing this, the video coming in through the camera eyes, changes over time and it begins to learn how this video changes and learns how the 3D nature of the toy generates certain videos if you rotate it a certain way, and eventually, how gravity works, and how the physics of the world works. Like a little scientist! And I have predicted for decades that future scaled-up versions of such AI scientists will want to further expand their horizons, and eventually go where most of the physical resources are, to build more and bigger AIs. And of course, almost all of these resources are far away from earth out there in space, which is hostile to humans but friendly to appropriately designed AI-controlled robots and self-replicating robot factories. So here we are not talking any longer about our tiny biosphere; no, we are talking about the much bigger rest of the universe. Within a few tens of billions of years, curious self-improving AIs will colonize the visible cosmos in a way that’s infeasible for humans. Those who don’t won’t have an impact. Sounds like science fiction, but since the 1970s I have been unable to see a plausible alternative to this scenario, except for a global catastrophe such as an all-out nuclear war that stops this development before it takes off. Jones: How long have these AIs, which can set their own goals — how long have they existed? To what extent can they be independent of human interaction? Schmidhuber: Neural networks like that have existed for over 30 years. My first simple adversarial neural network system of this kind is the one from 1990 described above. You don’t need a teacher there; it's just a little agent running around in the world and trying to invent new experiments that surprise its own prediction machine. Once it has figured out certain parts of the world, the agent will become bored and will move on to more exciting experiments. The simple 1990 systems I mentioned have certain limitations, but in the past three decades, we have also built more sophisticated systems that are setting their own goals and such systems I think will be essential for achieving true intelligence. If you are only imitating humans, you will never go beyond them. So, you really must give AIs the freedom to explore previously unexplored regions of the world in a way that no human is really predefining. Jones: Where is this being done today? Schmidhuber: Variants of neural network-based artificial curiosity are used today for agents that learn to play video games in a human-competitive way. We have also started to use them for automatic design of experiments in fields such as materials science. I bet many other fields will be affected by it: chemistry, biology, drug design, you name it. However, at least for now, these artificial scientists, as I like to call them, cannot yet compete with human scientists. I don’t think it’s going to stay this way but, at the moment, it’s still the case. Sure, AI has made a lot of progress. Since 1997, there have been superhuman chess players, and since 2011, through the DanNet of my team, there have been superhuman visual pattern recognizers. But there are other things where humans, at the moment at least, are much better, in particular, science itself. In the lab we have many first examples of self-directed artificial scientists, but they are not yet convincing enough to appear on the radar screen of the public space, which is currently much more fascinated with simpler systems that just imitate humans and write texts based on previously seen human-written documents. Jones: You speak of these numerous instances dating back 30 years of these lab experiments where these self-driven agents are deciding and learning and moving on once they’ve learned. And I assume that that rate of learning becomes even faster over time. What kind of timeframe are we talking about when this eventually is taken outside of the lab and embedded into society? Schmidhuber: This could still take months or even years :-) Anyway, in the not-too-distant future, we will probably see artificial scientists who are good at devising experiments that allow them to discover new, previously unknown physical laws. As always, we are going to profit from the old trend that has held at least since 1941: every decade compute is getting 100 times cheaper. Jones: How does this trend affect modern AI such as ChatGPT? Schmidhuber: Perhaps you know that all the recent famous AI applications such as ChatGPT and similar models are largely based on principles of artificial neural networks invented in the previous millennium. The main reason why they works so well now is the incredible acceleration of compute per dollar. ChatGPT is driven by a neural network called “Transformer” described in 2017 by Google. I am happy about that because a quarter century earlier in 1991 I had a particular Transformer variant which is now called the “Transformer with linearized self-attention”. Back then, not much could be done with it, because the compute cost was a million times higher than today. But today, one can train such models on half the internet and achieve much more interesting results. Jones: And for how long will this acceleration continue? Schmidhuber: There's no reason to believe that in the next 30 years, we won't have another factor of 1 million and that's going to be really significant. In the near future, for the first time we will have many not-so expensive devices that can compute as much as a human brain. The physical limits of computation, however, are much further out so even if the trend of a factor of 100 every decade continues, the physical limits (of 1051 elementary instructions per second and kilogram of matter) won’t be hit until, say, the mid-next century. Even in our current century, however, we’ll probably have many machines that compute more than all 10 billion human brains collectively and you can imagine, everything will change then! Jones: That is the big question. Is everything going to change? If so, what do you say to the next generation of leaders, currently coming out of college and university. So much of this change is already impacting how they study, how they will work, or how the future of work and livelihood is defined. What is their purpose and how do we change our systems so they will adapt to this new version of intelligence? Schmidhuber: For decades, people have asked me questions like that, because you know what I'm saying now, I have basically said since the 1970s, it’s just that today, people are paying more attention because, back then, they thought this was science fiction. They didn't think that I would ever come close to achieving my crazy life goal of building a machine that learns to become smarter than myself such that I can retire. But now many have changed their minds and think it's conceivable. And now I have two daughters, 23 and 25. People ask me: what do I tell them? They know that Daddy always said, “It seems likely that within your lifetimes, you will have new types of intelligence that are probably going to be superior in many ways, and probably all kinds of interesting ways.” How should they prepare for that? And I kept telling them the obvious: Learn how to learn new things! It's not like in the previous millennium where within 20 years someone learned to be a useful member of society, and then took a job for 40 years and performed in this job until she received her pension. Now things are changing much faster and we must learn continuously just to keep up. I also told my girls that no matter how smart AIs are going to get, learn at least the basics of math and physics, because that’s the essence of our universe, and anybody who understands this will have an advantage, and learn all kinds of new things more easily. I also told them that social skills will remain important, because most future jobs for humans will continue to involve interactions with other humans, but I couldn’t teach them anything about that; they know much more about social skills than I do. You touched on the big philosophical question about people’s purpose. Can this be answered without answering the even grander question: What’s the purpose of the entire universe? We don’t know. But what’s happening right now might be connected to the unknown answer. Don’t think of humans as the crown of creation. Instead view human civilization as part of a much grander scheme, an important step (but not the last one) on the path of the universe from very simple initial conditions towards more and more unfathomable complexity. Now it seems ready to take its next step, a step comparable to the invention of life itself over 3.5 billion years ago. Alas, don’t worry, in the end, all will be good! Jones: Let’s get back to this transformation happening right now with OpenAI. There are many questioning the efficacy and accuracy of ChatGPT, and are concerned its release has been premature. In light of the rampant adoption, educators have banned its use over concerns of plagiarism and how it stifles individual development. Should large language models like ChatGPT be used in school? Schmidhuber: When the calculator was first introduced, instructors forbade students from using it in school. Today, the consensus is that kids should learn the basic methods of arithmetic, but they should also learn to use the “artificial multipliers” aka calculators, even in exams, because laziness and efficiency is a hallmark of intelligence. Any intelligent being wants to minimize its efforts to achieve things. And that's the reason why we have tools, and why our kids are learning to use these tools. The first stone tools were invented maybe 3.5 million years ago; tools just have become more sophisticated over time. In fact, humans have changed in response to the properties of their tools. Our anatomical evolution was shaped by tools such as spears and fire. So, it's going to continue this way. And there is no permanent way of preventing large language models from being used in school. Jones: And when our children, your children graduate, what does their future work look like? Schmidhuber: A single human trying to predict details of how 10 billion people and their machines will evolve in the future is like a single neuron in my brain trying to predict what the entire brain and its tens of billions of neurons will do next year. 40 years ago, before the WWW was created at CERN in Switzerland, who would have predicted all those young people making money as YouTube video bloggers? Nevertheless, let’s make a few limited job-related observations. For a long time, people have thought that desktop jobs may require more intelligence than skills trade or handicraft professions. But now, it turns out that it's much easier to replace certain aspects of desktop jobs than replacing a carpenter, for example. Because everything that works well in AI is happening behind the screen currently, but not so much in the physical world. There are now artificial systems that can read lots of documents and then make really nice summaries of these documents. That is a desktop job. Or you give them a description of an illustration that you want to have for your article and pretty good illustrations are being generated that may need some minimal fine-tuning. But you know, all these desktop jobs are much easier to facilitate than the real tough jobs in the physical world. And it's interesting that the things people thought required intelligence, like playing chess, or writing or summarizing documents, are much easier for machines than they thought. But for things like playing football or soccer, there is no physical robot that can remotely compete with the abilities of a little boy with these skills. So, AI in the physical world, interestingly, is much harder than AI behind the screen in virtual worlds. And it's really exciting, in my opinion, to see that jobs such as plumbers are much more challenging than playing chess or writing another tabloid story. Jones: The way data has been collected in these large language models does not guarantee personal information has not been excluded. Current consent laws already are outdated when it comes to these large language models (LLM). The concern, rightly so, is increasing surveillance and loss of privacy. What is your view on this? Schmidhuber: As I have indicated earlier: are surveillance and loss of privacy inevitable consequences of increasingly complex societies? Super-organisms such as cities and states and companies consist of numerous people, just like people consist of numerous cells. These cells enjoy little privacy. They are constantly monitored by specialized "police cells" and "border guard cells": Are you a cancer cell? Are you an external intruder, a pathogen? Individual cells sacrifice their freedom for the benefits of being part of a multicellular organism. Similarly, for super-organisms such as nations. Over 5000 years ago, writing enabled recorded history and thus became its inaugural and most important invention. Its initial purpose, however, was to facilitate surveillance, to track citizens and their tax payments. The more complex a super-organism, the more comprehensive its collection of information about its constituents. 200 years ago, at least, the parish priest in each village knew everything about all the village people, even about those who did not confess, because they appeared in the confessions of others. Also, everyone soon knew about the stranger who had entered the village, because some occasionally peered out of the window, and what they saw got around. Such control mechanisms were temporarily lost through anonymization in rapidly growing cities but are now returning with the help of new surveillance devices such as smartphones as part of digital nervous systems that tell companies and governments a lot about billions of users. Cameras and drones etc. are becoming increasingly tinier and more ubiquitous. More effective recognition of faces and other detection technology are becoming cheaper and cheaper, and many will use it to identify others anywhere on earth; the big wide world will not offer any more privacy than the local village. Is this good or bad? Some nations may find it easier than others to justify more complex kinds of super-organisms at the expense of the privacy rights of their constituents. Jones: So, there is no way to stop or change this process of collection, or how it continuously informs decisions over time? How do you see governance and rules responding to this, especially amid Italy’s ban on ChatGPT following suspected user data breach and the more recent news about the Meta’s record $1.3billion fine in the company’s handling of user information? Schmidhuber: Data collection has benefits and drawbacks, such as the loss of privacy. How to balance those? I have argued for addressing this through data ownership in data markets. If it is true that data is the new oil, then it should have a price, just like oil. At the moment, the major surveillance platforms such as Meta do not offer users any money for their data and the transitive loss of privacy. In the future, however, we will likely see attempts at creating efficient data markets to figure out the data's true financial value through the interplay between supply and demand. Even some of the sensitive medical data should not be priced by governmental regulators but by patients (and healthy persons) who own it and who may sell or license parts thereof as micro-entrepreneurs in a healthcare data market. Following a previous interview, I gave for one of the largest re-insurance companies , let's look at the different participants in such a data market: patients, hospitals, data companies. (1) Patients with a rare form of cancer can offer more valuable data than patients with a very common form of cancer. (2) Hospitals and their machines are needed to extract the data, e.g., through magnet spin tomography, radiology, evaluations through human doctors, and so on. (3) Companies such as Siemens, Google or IBM would like to buy annotated data to make better artificial neural networks that learn to predict pathologies and diseases and the consequences of therapies. Now the market’s invisible hand will decide about the data’s price through the interplay between demand and supply. On the demand side, you will have several companies offering something for the data, maybe through an app on the smartphone (a bit like a stock market app). On the supply side, each patient in this market should be able to profit from high prices for rare valuable types of data. Likewise, competing data extractors such as hospitals will profit from gaining recognition and trust for extracting data well at a reasonable price. The market will make the whole system efficient through incentives for all who are doing a good job. Soon there will be a flourishing ecosystem of commercial data market advisors and what not, just like the ecosystem surrounding the traditional stock market. The value of the data won’t be determined by governments or ethics committees, but by those who own the data and decide by themselves which parts thereof they want to license to others under certain conditions. At first glance, a market-based system seems to be detrimental to the interest of certain monopolistic companies, as they would have to pay for the data - some would prefer free data and keep their monopoly. However, since every healthy and sick person in the market would suddenly have an incentive to collect and share their data under self-chosen anonymity conditions, there will soon be many more useful data to evaluate all kinds of treatments. On average, people will live longer and healthier, and many companies and the entire healthcare system will benefit. Jones: Finally, what is your view on open source versus the private companies like Google and OpenAI? Is there a danger to supporting these private companies’ large language models versus trying to keep these models open source and transparent, very much like what LAION is doing? Schmidhuber: I signed this open letter by LAION because I strongly favor the open-source movement. And I think it's also something that is going to challenge whatever big tech dominance there might be at the moment. Sure, the best models today are run by big companies with huge budgets for computers, but the exciting fact is that open-source models are not so far behind, some people say maybe six to eight months only. Of course, the private company models are all based on stuff that was created in academia, often in little labs without so much funding, which publish without patenting their results and open source their code and others take it and improved it. Big tech has profited tremendously from academia; their main achievement being that they have scaled up everything greatly, sometimes even failing to credit the original inventors. So, it's very interesting to see that as soon as some big company comes up with a new scaled-up model, lots of students out there are competing, or collaborating, with each other, trying to come up with equal or better performance on smaller networks and smaller machines. And since they are open sourcing, the next guy can have another great idea to improve it, so now there’s tremendous competition also for the big companies. Because of that, and since AI is still getting exponentially cheaper all the time, I don't believe that big tech companies will dominate in the long run. They find it very hard to compete with the enormous open-source movement. As long as you can encourage the open-source community, I think you shouldn't worry too much. Now, of course, you might say if everything is open source, then the bad actors also will more easily have access to these AI tools. And there's truth to that. But as always since the invention of controlled fire, it was good that knowledge about how technology works quickly became public such that everybody could use it. And then, against any bad actor, there's almost immediately a counter actor trying to nullify his efforts. You see, I still believe in our old motto "AI∀" or "AI For All." Jones: Thank you, Juergen for sharing your perspective on this amazing time in history. It’s clear that with new technology, the enormous potential can be matched by disparate and troubling risks which we’ve yet to solve, and even those we have yet to identify. If we are to dispel the fear of a sentient system for which we have no control, humans, alone need to take steps for more responsible development and collaboration to ensure AI technology is used to ultimately benefit society. Humanity will be judged by what we do next.

[N] TheSequence Scope: When it comes to machine learning, size matters: Microsoft's DeepSpeed framework, which can train a model with up to a trillion parameters
reddit
LLM Vibe Score0
Human Vibe Score1
KseniaseThis week

[N] TheSequence Scope: When it comes to machine learning, size matters: Microsoft's DeepSpeed framework, which can train a model with up to a trillion parameters

Hi there! Offering to your attention the latest edition of a weekly ML-newsletter that focusing on three things: impactful ML research papers, cool ML tech solutions, and ML use cases supported by investors. Please, see it below. Reddit is a new thing for me, and I've been struggling a bit with it, so please don't judge me too harsh for this promotion. This weekly digest is free and I hope you'd find the format convenient for you. Your feedback is very appreciated, and please feel free to sign up if you like it. 📝 Editorial  The recent emergence of pre-trained language models and transformer architectures pushed the creation of larger and larger machine learning models. Google’s BERT presented attention mechanism and transformer architecture possibilities as the “next big thing” in ML, and the numbers seem surreal. OpenAI’s GPT-2 set a record by processing 1.5 billion parameters, followed by Microsoft’s Turing-NLG, which processed 17 billion parameters just to see the new GPT-3 processing an astonishing 175 billion parameters. To not feel complacent, just this week Microsoft announced a new release of its DeepSpeed framework (which powers Turing-NLG), which can train a model with up to a trillion parameters. That sounds insane but it really isn’t.   What we are seeing is a consequence of several factors. First, computation power and parallelization techniques have evolved to a point where it is relatively easy to train machine learning models in large clusters of machines. Second and most importantly, in the current state of machine learning, larger models have regularly outperformed smaller and more specialized models. Knowledge reusability methods like transfer learning are still in very nascent stages. As a result, it’s really hard to build small models that can operate in uncertain environments. Furthermore, as models like GPT-3 and Turing-NLG have shown, there is some unexplainable magic that happens after models go past a certain size. Many of the immediate machine learning problems might be solved by scaling the current generation of neural network architectures. Plain and simple, when it comes to machine learning, size matters.   We would love to hear your opinions about the debate between broader-larger vs. smaller and more specialized models.   Leave a comment Now, to the most important developments in the AI industry this week 🔎 ML Research GPT-3 Falls Short in Machine Comprehension Proposed by researchers from a few major American universities, a 57-task test to measure models’ ability to reason poses challenges even for sophisticated models like GPT-3 ->read more in the original paper Better Text Summarization OpenAI published a paper showing a reinforcement learning with human feedback technique that can surpass supervised models ->read more on OpenAI blog Reinforcement Learning with Offline Datasets Researchers from the Berkeley AI Research (BAIR) Lab published a paper unveiling a method that uses offline datasets to improve reinforcement learning models->read more on BAIR blog 🤖 Cool AI Tech Releases New Version of DeepSpeed Microsoft open-sourced a new version of DeepSpeed, an open-source library for parallelizable training that can scale up to models with 1 trillion parameters->read more on Microsoft Research blog 💸 Money in AI AI-powered customer experience management platform Sprinklr has raised $200 million (kudos to our subscribers from Sprinklr 👏). Sprinklr's “AI listening processing” solution allows companies to get structured and meaningful sentiments and insights from unstructured customer data that comes from public conversations on different websites and social platforms. Xometry, an on-demand industrial parts marketplace, raises $75 million in Series E funding. The company provides a digital way of creating the right combination of buyers and manufacturers. Another example of AI implementation into matching two sides for a deal. Real estate tech company Orchard raises $69 million in its recent funding round. Orchard aims to digitize the whole real estate market, by developing a solution that combines machine learning and rapid human assistance to smooth the search, match the right deal, and simplify buying and selling relationships. Cybersecurity startup Pcysys raised $25 million in its funding round. Pcysys’ platform, which doesn’t require installation or network reconfiguration, uses algorithms to scan and “ethically” attack enterprise networks. Robotics farming company Iron Ox raised $20 million in a funding round. The system of farming robots is still semi-autonomous, the company’s goal is to become fully autonomous.  Insurtech company Descartes Underwriting raised $18.5 million. The company applies AI and machine learning technologies to climate risk predicting and insurance underwriting. Legaltech startup ThoughtRiver raised $10 million in its Series A round. Its AI solution applied to contract pre-screening aims to boost operational efficiency. Medtech startup Skin Analytics raised $5.1 million in Series A funding. Skin Analytics has developed a clinically validated AI system that can identify not only the important skin cancers but also precancerous lesions that can be treated, as well as a range of lesions that are benign. Amazon, along with several government organizations and three other industry partners, helped fund the National Science Foundation, a high-priority AI research initiative. The amount of funding is not disclosed. The content of TheSequence is written by Jesus Rodriguez, one of the most-read contributors to KDNuggets and TDS. You can check his Medium here.

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!
reddit
LLM Vibe Score0
Human Vibe Score-0.5
regalalgorithmThis week

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!

Hi there, we at Skynet Today produce a weekly newsletter summarizing each week's major AI news, which seems like it'd be of interest to this subreddit. Here's what's in our latest one: Facebook’s AI for detecting hate speech is facing its biggest challenge yet Facebook has made significant progress recently to proactively take down content that violate its community standards. For example, in the second quarter of 2020, Facebook took down 104.6 million pieces of content. While reviews are typically performed by a vast workforce of human moderators, AI-powered tools have enabled Facebook to do this work at a greater scale for textual content. However, there’s a long way to go for these systems to match or exceed the capabilities of human moderators. This is because a large proportion of hate speech and misinformation is in the form of images and memes, and reasoning about the context and language-image interplay is an extremely difficult challenge for AI. Given Facebook’s scale and the speed at which some use it to spread hate, incite violence, and share lies with millions, Facebook will have to keep running to catch up. AI Slays Top F-16 Pilot In DARPA Dogfight Simulation The Defense Advanced Research Project Agency (DARPA) recently hosted a simulated F16 dogfight competition, with different AI bots competing with each other as well as with human pilots. The top AI bot was able to beat a human pilot 5-0 in the simulated contest. DARPA started this program “as a risk-reduction effort \[…\] to flesh out how human and machine pilots share operational control of a fighter jet to maximize its chances of mission success.” Competition runners are broadly optimistic about the demonstration of AI capabilities, even if they are not close to being deployed on a real aircraft. Of concern, the program had little discussion on the ethics of AI military applications, especially with the lethal autonomous weapon systems being considered. News Advances & Business Microsoft, Energy Dept. to Develop Disaster-Response AI Tools \- The U.S. Department of Energy and Microsoft Corp. on Tuesday announced a partnership to develop artificial-intelligence tools aimed at helping first-responders better react to fast-changing natural events, such as floods and wildfires. Coronavirus: Robot CERi is a bilingual Covid-19 expert \- Ceri is bilingual, clued-up on coronavirus and can tell what mood you are in. Ceri also happens to be a robot. Moscow DOH uses AI platform to detect lung cancer symptoms \- Moscow’s department of health is using an artificial intelligence (AI) platform to detect symptoms of lung cancer in CT scans, as part of a project to implement AI technology for radiology. Scientists develop artificial intelligence system for high precision recognition of hand gestures \- The recognition of human hand gestures by AI systems has been a valuable development over the last decade and has been adopted in high-precision surgical robots, health monitoring equipment and in gaming systems. Forget credit cards - now you can pay with your face. Creepy or cool? \- A new way to pay has arrived in Los Angeles: your face. Concerns & Hype The dystopian tech that companies are selling to help schools reopen sooner \- This fall, AI could be watching students social distance and checking their masks. Thousands of schools nationwide will not be reopening this fall. NYPD Used Facial Recognition Technology In Siege Of Black Lives Matter Activist’s Apartment \- The NYPD deployed facial recognition technology in its hunt for a prominent Black Lives Matter activist, whose home was besieged by dozens of officers and police dogs last week, a spokesperson confirmed to Gothamist. Machines can spot mental health issues - if you hand over your personal data \- Digital diagnosis could transform psychiatry by mining your most intimate data for clues. But is the privacy cost worth it? Supporting Black Artists Who Are Examining AI \- Technology has a complicated relationship with racial justice. Smartphones, internet platforms, and other digital tools can be used to document and expose racism. But digital tools can also fuel racism: smart doorbells surveil Black individuals. A-level and GCSE results in England to be based on teacher assessments in U-turn \- All A-level and GCSE results in England will be based on grades assesed by teachers instead of algorithms. Analysis & Policy GPT-3 and The Question of Automation \- Automation is not an all or nothing proposition. An AI model’s automation capability is highly conjoined with the task and application it is used in. An A.I. Movie Service Could One Day Serve You a New Custom Film Every Time \- How long will it be until an A.I. can make an actual feature film on demand? Fairness, evidence, and predictive equality \- How the causal fairness principle relates to predictive equality How robotics and automation could create new jobs in the new normal \- Depending on who you ask, AI and automation will either destroy jobs or create new ones. In reality, a greater push toward automation will probably both kill and create jobs - human workers will become redundant in certain spheres, sure, but many new roles will likely crop up. Expert Opinions & Discussion within the field Too many AI researchers think real-world problems are not relevant \- The community’s hyperfocus on novel methods ignores what’s really important.

[P]MMML | Deploy HuggingFace training model rapidly based on MetaSpore
reddit
LLM Vibe Score0
Human Vibe Score1
qazmkoppThis week

[P]MMML | Deploy HuggingFace training model rapidly based on MetaSpore

A few days ago, HuggingFace announced a $100 million Series C funding round, which was big news in open source machine learning and could be a sign of where the industry is headed. Two days before the HuggingFace funding announcement, open-source machine learning platform MetaSpore released a demo based on the HuggingFace Rapid deployment pre-training model. As deep learning technology makes innovative breakthroughs in computer vision, natural language processing, speech understanding, and other fields, more and more unstructured data are perceived, understood, and processed by machines. These advances are mainly due to the powerful learning ability of deep learning. Through pre-training of deep models on massive data, the models can capture the internal data patterns, thus helping many downstream tasks. With the industry and academia investing more and more energy in the research of pre-training technology, the distribution warehouses of pre-training models such as HuggingFace and Timm have emerged one after another. The open-source community release pre-training significant model dividends at an unprecedented speed. In recent years, the data form of machine modeling and understanding has gradually evolved from single-mode to multi-mode, and the semantic gap between different modes is being eliminated, making it possible to retrieve data across modes. Take CLIP, OpenAI’s open-source work, as an example, to pre-train the twin towers of images and texts on a dataset of 400 million pictures and texts and connect the semantics between pictures and texts. Many researchers in the academic world have been solving multimodal problems such as image generation and retrieval based on this technology. Although the frontier technology through the semantic gap between modal data, there is still a heavy and complicated model tuning, offline data processing, high performance online reasoning architecture design, heterogeneous computing, and online algorithm be born multiple processes and challenges, hindering the frontier multimodal retrieval technologies fall to the ground and pratt &whitney. DMetaSoul aims at the above technical pain points, abstracting and uniting many links such as model training optimization, online reasoning, and algorithm experiment, forming a set of solutions that can quickly apply offline pre-training model to online. This paper will introduce how to use the HuggingFace community pre-training model to conduct online reasoning and algorithm experiments based on MetaSpore technology ecology so that the benefits of the pre-training model can be fully released to the specific business or industry and small and medium-sized enterprises. And we will give the text search text and text search graph two multimodal retrieval demonstration examples for your reference. Multimodal semantic retrieval The sample architecture of multimodal retrieval is as follows: Our multimodal retrieval system supports both text search and text search application scenarios, including offline processing, model reasoning, online services, and other core modules: ​ https://preview.redd.it/w4v4c7vcez291.png?width=1834&format=png&auto=webp&s=0687efb1fddb26e8e30cb844d398ec712b947f31 Offline processing, including offline data processing processes for different application scenarios of text search and text search, including model tuning, model export, data index database construction, data push, etc. Model inference. After the offline model training, we deployed our NLP and CV large models based on the MetaSpore Serving framework. MetaSpore Serving helps us conveniently perform online inference, elastic scheduling, load balancing, and resource scheduling in heterogeneous environments. Online services. Based on MetaSpore’s online algorithm application framework, MetaSpore has a complete set of reusable online search services, including Front-end retrieval UI, multimodal data preprocessing, vector recall and sorting algorithm, AB experimental framework, etc. MetaSpore also supports text search by text and image scene search by text and can be migrated to other application scenarios at a low cost. The HuggingFace open source community has provided several excellent baseline models for similar multimodal retrieval problems, which are often the starting point for actual optimization in the industry. MetaSpore also uses the pre-training model of the HuggingFace community in its online services of searching words by words and images by words. Searching words by words is based on the semantic similarity model of the question and answer field optimized by MetaSpore, and searching images by words is based on the community pre-training model. These community open source pre-training models are exported to the general ONNX format and loaded into MetaSpore Serving for online reasoning. The following sections will provide a detailed description of the model export and online retrieval algorithm services. The reasoning part of the model is standardized SAAS services with low coupling with the business. Interested readers can refer to my previous post: The design concept of MetaSpore, a new generation of the one-stop machine learning platform. 1.1 Offline Processing Offline processing mainly involves the export and loading of online models and index building and pushing of the document library. You can follow the step-by-step instructions below to complete the offline processing of text search and image search and see how the offline pre-training model achieves reasoning at MetaSpore. 1.1.1 Search text by text Traditional text retrieval systems are based on literal matching algorithms such as BM25. Due to users’ diverse query words, a semantic gap between query words and documents is often encountered. For example, users misspell “iPhone” as “Phone,” and search terms are incredibly long, such as “1 \~ 3 months old baby autumn small size bag pants”. Traditional text retrieval systems will use spelling correction, synonym expansion, search terms rewriting, and other means to alleviate the semantic gap but fundamentally fail to solve this problem. Only when the retrieval system fully understands users’ query terms and documents can it meet users’ retrieval demands at the semantic level. With the continuous progress of pre-training and representational learning technology, some commercial search engines continue to integrate semantic vector retrieval methods based on symbolic learning into the retrieval ecology. Semantic retrieval model This paper introduces a set of semantic vector retrieval applications. MetaSpore built a set of semantic retrieval systems based on encyclopedia question and answer data. MetaSpore adopted the Sentence-Bert model as the semantic vector representation model, which fine-tunes the twin tower BERT in supervised or unsupervised ways to make the model more suitable for retrieval tasks. The model structure is as follows: The query-Doc symmetric two-tower model is used in text search and question and answer retrieval. The vector representation of online Query and offline DOC share the same vector representation model, so it is necessary to ensure the consistency of the offline DOC library building model and online Query inference model. The case uses MetaSpore’s text representation model Sbert-Chinese-QMC-domain-V1, optimized in the open-source semantically similar data set. This model will express the question and answer data as a vector in offline database construction. The user query will be expressed as a vector by this model in online retrieval, ensuring that query-doc in the same semantic space, users’ semantic retrieval demands can be guaranteed by vector similarity metric calculation. Since the text presentation model does vector encoding for Query online, we need to export the model for use by the online service. Go to the q&A data library code directory and export the model concerning the documentation. In the script, Pytorch Tracing is used to export the model. The models are exported to the “./export “directory. The exported models are mainly ONNX models used for wired reasoning, Tokenizer, and related configuration files. The exported models are loaded into MetaSpore Serving by the online Serving system described below for model reasoning. Since the exported model will be copied to the cloud storage, you need to configure related variables in env.sh. \Build library based on text search \ The retrieval database is built on the million-level encyclopedia question and answer data set. According to the description document, you need to download the data and complete the database construction. The question and answer data will be coded as a vector by the offline model, and then the database construction data will be pushed to the service component. The whole process of database construction is described as follows: Preprocessing, converting the original data into a more general JSonline format for database construction; Build index, use the same model as online “sbert-Chinese-qmc-domain-v1” to index documents (one document object per line); Push inverted (vector) and forward (document field) data to each component server. The following is an example of the database data format. After offline database construction is completed, various data are pushed to corresponding service components, such as Milvus storing vector representation of documents and MongoDB storing summary information of documents. Online retrieval algorithm services will use these service components to obtain relevant data. 1.1.2 Search by text Text and images are easy for humans to relate semantically but difficult for machines. First of all, from the perspective of data form, the text is the discrete ID type of one-dimensional data based on words and words. At the same time, images are continuous two-dimensional or three-dimensional data. Secondly, the text is a subjective creation of human beings, and its expressive ability is vibrant, including various turning points, metaphors, and other expressions, while images are machine representations of the objective world. In short, bridging the semantic gap between text and image data is much more complex than searching text by text. The traditional text search image retrieval technology generally relies on the external text description data of the image or the nearest neighbor retrieval technology and carries out the retrieval through the image associated text, which in essence degrades the problem to text search. However, it will also face many issues, such as obtaining the associated text of pictures and whether the accuracy of text search by text is high enough. The depth model has gradually evolved from single-mode to multi-mode in recent years. Taking the open-source project of OpenAI, CLIP, as an example, train the model through the massive image and text data of the Internet and map the text and image data into the same semantic space, making it possible to implement the text and image search technology based on semantic vector. CLIP graphic model The text search pictures introduced in this paper are implemented based on semantic vector retrieval, and the CLIP pre-training model is used as the two-tower retrieval architecture. Because the CLIP model has trained the semantic alignment of the twin towers’ text and image side models on the massive graphic and text data, it is particularly suitable for the text search graph scene. Due to the different image and text data forms, the Query-Doc asymmetric twin towers model is used for text search image retrieval. The image-side model of the twin towers is used for offline database construction, and the text-side model is used for the online return. In the final online retrieval, the database data of the image side model will be searched after the text side model encodes Query, and the CLIP pre-training model guarantees the semantic correlation between images and texts. The model can draw the graphic pairs closer in vector space by pre-training on a large amount of visual data. Here we need to export the text-side model for online MetaSpore Serving inference. Since the retrieval scene is based on Chinese, the CLIP model supporting Chinese understanding is selected. The exported content includes the ONNX model used for online reasoning and Tokenizer, similar to the text search. MetaSpore Serving can load model reasoning through the exported content. Build library on Image search You need to download the Unsplash Lite library data and complete the construction according to the instructions. The whole process of database construction is described as follows: Preprocessing, specify the image directory, and then generate a more general JSOnline file for library construction; Build index, use OpenAI/Clip-Vit-BASE-Patch32 pre-training model to index the gallery, and output one document object for each line of index data; Push inverted (vector) and forward (document field) data to each component server. Like text search, after offline database construction, relevant data will be pushed to service components, called by online retrieval algorithm services to obtain relevant data. 1.2 Online Services The overall online service architecture diagram is as follows: https://preview.redd.it/jfsl8hdfez291.png?width=1280&format=png&auto=webp&s=a858e2304a0c93e78ba5429612ca08cbee69b35a Multi-mode search online service system supports application scenarios such as text search and text search. The whole online service consists of the following parts: Query preprocessing service: encapsulate preprocessing logic (including text/image, etc.) of pre-training model, and provide services through gRPC interface; Retrieval algorithm service: the whole algorithm processing link includes AB experiment tangent flow configuration, MetaSpore Serving call, vector recall, sorting, document summary, etc.; User entry service: provides a Web UI interface for users to debug and track down problems in the retrieval service. From a user request perspective, these services form invocation dependencies from back to front, so to build up a multimodal sample, you need to run each service from front to back first. Before doing this, remember to export the offline model, put it online and build the library first. This article will introduce the various parts of the online service system and make the whole service system step by step according to the following guidance. See the ReadME at the end of this article for more details. 1.2.1 Query preprocessing service Deep learning models tend to be based on tensors, but NLP/CV models often have a preprocessing part that translates raw text and images into tensors that deep learning models can accept. For example, NLP class models often have a pre-tokenizer to transform text data of string type into discrete tensor data. CV class models also have similar processing logic to complete the cropping, scaling, transformation, and other processing of input images through preprocessing. On the one hand, considering that this part of preprocessing logic is decoupled from tensor reasoning of the depth model, on the other hand, the reason of the depth model has an independent technical system based on ONNX, so MetaSpore disassembled this part of preprocessing logic. NLP pretreatment Tokenizer has been integrated into the Query pretreatment service. MetaSpore dismantlement with a relatively general convention. Users only need to provide preprocessing logic files to realize the loading and prediction interface and export the necessary data and configuration files loaded into the preprocessing service. Subsequent CV preprocessing logic will also be integrated in this manner. The preprocessing service currently provides the gRPC interface invocation externally and is dependent on the Query preprocessing (QP) module in the retrieval algorithm service. After the user request reaches the retrieval algorithm service, it will be forwarded to the service to complete the data preprocessing and continue the subsequent processing. The ReadMe provides details on how the preprocessing service is started, how the preprocessing model exported offline to cloud storage enters the service, and how to debug the service. To further improve the efficiency and stability of model reasoning, MetaSpore Serving implements a Python preprocessing submodule. So MetaSpore can provide gRPC services through user-specified preprocessor.py, complete Tokenizer or CV-related preprocessing in NLP, and translate requests into a Tensor that deep models can handle. Finally, the model inference is carried out by MetaSpore, Serving subsequent sub-modules. Presented here on the lot code: https://github.com/meta-soul/MetaSpore/compare/add\python\preprocessor 1.2.2 Retrieval algorithm services Retrieval algorithm service is the core of the whole online service system, which is responsible for the triage of experiments, the assembly of algorithm chains such as preprocessing, recall, sorting, and the invocation of dependent component services. The whole retrieval algorithm service is developed based on the Java Spring framework and supports multi-mode retrieval scenarios of text search and text search graph. Due to good internal abstraction and modular design, it has high flexibility and can be migrated to similar application scenarios at a low cost. Here’s a quick guide to configuring the environment to set up the retrieval algorithm service. See ReadME for more details: Install dependent components. Use Maven to install the online-Serving component Search for service configurations. Copy the template configuration file and replace the MongoDB, Milvus, and other configurations based on the development/production environment. Install and configure Consul. Consul allows you to synchronize the search service configuration in real-time, including cutting the flow of experiments, recall parameters, and sorting parameters. The project’s configuration file shows the current configuration parameters of text search and text search. The parameter modelName in the stage of pretreatment and recall is the corresponding model exported in offline processing. Start the service. Once the above configuration is complete, the retrieval service can be started from the entry script. Once the service is started, you can test it! For example, for a user with userId=10 who wants to query “How to renew ID card,” access the text search service. 1.2.3 User Entry Service Considering that the retrieval algorithm service is in the form of the API interface, it is difficult to locate and trace the problem, especially for the text search image scene can intuitively display the retrieval results to facilitate the iterative optimization of the retrieval algorithm. This paper provides a lightweight Web UI interface for text search and image search, a search input box, and results in a display page for users. Developed by Flask, the service can be easily integrated with other retrieval applications. The service calls the retrieval algorithm service and displays the returned results on the page. It’s also easy to install and start the service. Once you’re done, go to http://127.0.0.1:8090 to see if the search UI service is working correctly. See the ReadME at the end of this article for details. Multimodal system demonstration The multimodal retrieval service can be started when offline processing and online service environment configuration have been completed following the above instructions. Examples of textual searches are shown below. Enter the entry of the text search map application, enter “cat” first, and you can see that the first three digits of the returned result are cats: https://preview.redd.it/0n5nuyvhez291.png?width=1280&format=png&auto=webp&s=1e9c054f541d53381674b8d6001b4bf524506bd2 If you add a color constraint to “cat” to retrieve “black cat,” you can see that it does return a black cat: https://preview.redd.it/rzc0qjyjez291.png?width=1280&format=png&auto=webp&s=d5bcc503ef0fb3360c7740e60e295cf372dcad47 Further, strengthen the constraint on the search term, change it to “black cat on the bed,” and return results containing pictures of a black cat climbing on the bed: ​ https://preview.redd.it/c4b2q8olez291.png?width=1280&format=png&auto=webp&s=4f3817b0b9f07e1e68d1d4a8281702ba3834a00a The cat can still be found through the text search system after the color and scene modification in the above example. Conclusion The cutting-edge pre-training technology can bridge the semantic gap between different modes, and the HuggingFace community can greatly reduce the cost for developers to use the pre-training model. Combined with the technological ecology of MetaSpore online reasoning and online microservices provided by DMetaSpore, the pre-training model is no longer mere offline dabbling. Instead, it can truly achieve end-to-end implementation from cutting-edge technology to industrial scenarios, fully releasing the dividends of the pre-training large model. In the future, DMetaSoul will continue to improve and optimize the MetaSpore technology ecosystem: More automated and wider access to HuggingFace community ecology. MetaSpore will soon release a common model rollout mechanism to make HuggingFace ecologically accessible and will later integrate preprocessing services into online services. Multi-mode retrieval offline algorithm optimization. For multimodal retrieval scenarios, MetaSpore will continuously iteratively optimize offline algorithm components, including text recall/sort model, graphic recall/sort model, etc., to improve the accuracy and efficiency of the retrieval algorithm. For related code and reference documentation in this article, please visit: https://github.com/meta-soul/MetaSpore/tree/main/demo/multimodal/online Some images source: https://github.com/openai/CLIP/raw/main/CLIP.png https://www.sbert.net/examples/training/sts/README.html

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!
reddit
LLM Vibe Score0
Human Vibe Score-0.5
regalalgorithmThis week

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!

Hi there, we at Skynet Today produce a weekly newsletter summarizing each week's major AI news, which seems like it'd be of interest to this subreddit. Here's what's in our latest one: Facebook’s AI for detecting hate speech is facing its biggest challenge yet Facebook has made significant progress recently to proactively take down content that violate its community standards. For example, in the second quarter of 2020, Facebook took down 104.6 million pieces of content. While reviews are typically performed by a vast workforce of human moderators, AI-powered tools have enabled Facebook to do this work at a greater scale for textual content. However, there’s a long way to go for these systems to match or exceed the capabilities of human moderators. This is because a large proportion of hate speech and misinformation is in the form of images and memes, and reasoning about the context and language-image interplay is an extremely difficult challenge for AI. Given Facebook’s scale and the speed at which some use it to spread hate, incite violence, and share lies with millions, Facebook will have to keep running to catch up. AI Slays Top F-16 Pilot In DARPA Dogfight Simulation The Defense Advanced Research Project Agency (DARPA) recently hosted a simulated F16 dogfight competition, with different AI bots competing with each other as well as with human pilots. The top AI bot was able to beat a human pilot 5-0 in the simulated contest. DARPA started this program “as a risk-reduction effort \[…\] to flesh out how human and machine pilots share operational control of a fighter jet to maximize its chances of mission success.” Competition runners are broadly optimistic about the demonstration of AI capabilities, even if they are not close to being deployed on a real aircraft. Of concern, the program had little discussion on the ethics of AI military applications, especially with the lethal autonomous weapon systems being considered. News Advances & Business Microsoft, Energy Dept. to Develop Disaster-Response AI Tools \- The U.S. Department of Energy and Microsoft Corp. on Tuesday announced a partnership to develop artificial-intelligence tools aimed at helping first-responders better react to fast-changing natural events, such as floods and wildfires. Coronavirus: Robot CERi is a bilingual Covid-19 expert \- Ceri is bilingual, clued-up on coronavirus and can tell what mood you are in. Ceri also happens to be a robot. Moscow DOH uses AI platform to detect lung cancer symptoms \- Moscow’s department of health is using an artificial intelligence (AI) platform to detect symptoms of lung cancer in CT scans, as part of a project to implement AI technology for radiology. Scientists develop artificial intelligence system for high precision recognition of hand gestures \- The recognition of human hand gestures by AI systems has been a valuable development over the last decade and has been adopted in high-precision surgical robots, health monitoring equipment and in gaming systems. Forget credit cards - now you can pay with your face. Creepy or cool? \- A new way to pay has arrived in Los Angeles: your face. Concerns & Hype The dystopian tech that companies are selling to help schools reopen sooner \- This fall, AI could be watching students social distance and checking their masks. Thousands of schools nationwide will not be reopening this fall. NYPD Used Facial Recognition Technology In Siege Of Black Lives Matter Activist’s Apartment \- The NYPD deployed facial recognition technology in its hunt for a prominent Black Lives Matter activist, whose home was besieged by dozens of officers and police dogs last week, a spokesperson confirmed to Gothamist. Machines can spot mental health issues - if you hand over your personal data \- Digital diagnosis could transform psychiatry by mining your most intimate data for clues. But is the privacy cost worth it? Supporting Black Artists Who Are Examining AI \- Technology has a complicated relationship with racial justice. Smartphones, internet platforms, and other digital tools can be used to document and expose racism. But digital tools can also fuel racism: smart doorbells surveil Black individuals. A-level and GCSE results in England to be based on teacher assessments in U-turn \- All A-level and GCSE results in England will be based on grades assesed by teachers instead of algorithms. Analysis & Policy GPT-3 and The Question of Automation \- Automation is not an all or nothing proposition. An AI model’s automation capability is highly conjoined with the task and application it is used in. An A.I. Movie Service Could One Day Serve You a New Custom Film Every Time \- How long will it be until an A.I. can make an actual feature film on demand? Fairness, evidence, and predictive equality \- How the causal fairness principle relates to predictive equality How robotics and automation could create new jobs in the new normal \- Depending on who you ask, AI and automation will either destroy jobs or create new ones. In reality, a greater push toward automation will probably both kill and create jobs - human workers will become redundant in certain spheres, sure, but many new roles will likely crop up. Expert Opinions & Discussion within the field Too many AI researchers think real-world problems are not relevant \- The community’s hyperfocus on novel methods ignores what’s really important.

[N] New Trends to Power Faster Artificial Intelligence and Machine Learning Adoption?
reddit
LLM Vibe Score0
Human Vibe Score1
EsotericaCapitalThis week

[N] New Trends to Power Faster Artificial Intelligence and Machine Learning Adoption?

In 2012, Google X lab created a neural network that can identify cats. Since then, technology companies have been increasingly adopting AI/ML on a large scale to build better applications and services for consumers (ToC). On the other hand, AI/ML's adoption on the enterprises' side (ToB) has yet to see the same growth trajectory due to the costs and complexities in both hardware and software. However, Since 2020, we started noticing three emerging tech trends that can help accelerate enterprises' adoption of AI/ML. Breakthrough in semiconductors: In 2020, Nvidia debut the concept of "Data Processing Unit," a new class of programmable processors that combine high-performance CPU with SmartNiC (network interface controller). Data centers can deploy DPUs to optimize computing offload and frees up CPUs to focus on intended tasks, such as machine learning. DPUs help resolve a significant bottleneck for ML training, where models, sometimes with billions of parameters, are way too big for traditional CPUs and GPUs to handle. Other leading semiconductor players, such as Marvell and Xilinx, follow suit with their in-house or partner-designed DPUs. Industry analysts have forecasted that the market size for DPUs in data centers alone could reach $50 billion by 2025. ​ https://preview.redd.it/l436muluhnn61.png?width=1430&format=png&auto=webp&s=ba8d1298056ea31bddd25f1596ff64b7e107580a Breakthrough in software: we also saw significant progress of "Conversational AI," a new form of AI that can understand and speak languages with human-like accuracy, in 2020. Conversational AI allows two-way interactions and provides a much better user experience than traditional AI-powered Chatbot, mostly a one-way response system. The secret of conversational AI is its ability to handle lots of human conversation variance. Developers have designed innovative algorithms such as "Switch transformers" and "Sparse training" to enable models to handle vast amounts of data. The size of conversational AI training models is enormous and keeps expanding. For example, in February 2021, Google Brain announced a model with 1.6 trillion parameters, nine times the size of the famous Open AI GPT-3 (175 billion parameters) unveiled in July 2020. GPT-3 is 100+ times bigger than GPT-2 introduced in 2019. ​ https://preview.redd.it/oajpi2yvhnn61.png?width=1430&format=png&auto=webp&s=1482913a98e17ddc1d62cc79864598d4012ad6f7 Cloud giants are expanding machine-learning platforms for developers. Andy Jassy famously said that "AI is shifting from a niche experiment inside technical departments to becoming more mainstream in business processes." in the 2020 AWS reInvent. During the conference, AWS rolled out many AI products across the technology stack, including AI chips (AWS Trainium), database (Aurora Machine Learning), and vertical solutions (Amazon Healthlake), etc. However, the most significant development is the drastic expansion of "Amazon SageMaker," one of the largest cloud machine-learning platforms. SageMaker has been offering new features to make it easier for developers to automate machine learning workflow. Microsoft Azure and Google Cloud are also growing their ML developer platforms. ​ https://preview.redd.it/z9wf2o8xhnn61.png?width=1430&format=png&auto=webp&s=9f607acfe8f0dbf36fb9b472f3cb40b80f13879e Witnessing these breakthroughs in semiconductor and software, coupled with cloud giants' effort to democratize AI, we see a coming inflection point of accelerated AI adoption in both ToC and ToB markets. So how do we benefit from this megatrend? In semiconductors, we like companies with DPUs exposure. In AI development and processing, we favor multi-cloud AI platforms such as Databricks. In enterprise software, we believe there will be a strong wave of new AI-based enterprise applications that can be creative and efficient in solving real-world problems.

[P] Improve AI 8.0: Free Contextual Multi-Armed Bandit Platform for Scoring, Ranking & Decisions
reddit
LLM Vibe Score0
Human Vibe Score1
gogogadgetlegzThis week

[P] Improve AI 8.0: Free Contextual Multi-Armed Bandit Platform for Scoring, Ranking & Decisions

Improve AI 8.0 - Contextual Multi-Armed Bandit Platform for Scoring, Ranking & Decisions Full announcement post at: https://improve.ai/2023/06/08/contextual-bandit.html We’re thrilled to introduce Improve AI 8.0, a modern, free, production-ready contextual multi-armed bandit platform that quickly scores and ranks items using intuitive reward-based training. Multi-armed bandits and contextual bandits are corner-stone machine learning algorithms that power a myriad of applications including recommendation systems, personalization, query re-ranking, automated decisions, and multi-variate optimization. With version 8, we’ve fully delivered on our original vision - providing a high performance, simple to use, low cost contextual multi-armed bandit platform. Key features of v8.0 include: Simplified APIs 90% more memory efficient XGBoost models The reward tracker & trainer is now free for most uses On-device scoring, ranking, and decisions for iOS and Android apps Native Swift SDK that can rank or score any Encodable Ranked Value Encoding* for accurate scoring of String properties Compact hash tables for reduced model sizes when encoding large numbers of string values Balanced exploration vs exploitation using Thompson Sampling Simple APIs With Swift, Python, or Java, create a list of JSON encodable items and simply call Ranker.rank(items). For instance, in an iOS bedtime story app, you may have a list of Story objects: struct Story: Codable { var title: String var author: String var pageCount: Int } To obtain a ranked list of stories, use just one line of code: let rankedStories = try Ranker(modelUrl).rank(stories) The expected best story will be the first element in the ranked list: let bestStory = rankedStories.first Simple Training Easily train your rankers using reinforcement learning. First, track when an item is used: let tracker = RewardTracker("stories", trackUrl) let rewardId = tracker.track(story, from: rankedStories) Later, if a positive outcome occurs, provide a reward: if (purchased) { tracker.addReward(profit, rewardId) } Reinforcement learning uses positive rewards for favorable outcomes (a “carrot”) and negative rewards for undesirable outcomes (a “stick”). By assigning rewards based on business metrics, such as revenue or conversions, the system optimizes these metrics over time. Contextual Ranking & Scoring Improve AI turns XGBoost into a contextual multi-armed bandit, meaning that context is considered when making ranking or scoring decisions. Often, the choice of the best variant depends on the context that the decision is made within. Let’s take the example of greetings for different times of the day: greetings = ["Good Morning", "Good Afternoon", "Good Evening", "Buenos Días", "Buenas Tardes", "Buenas Noches"] rank() also considers the context of each decision. The context can be any JSON-encodable data structure. ranked = ranker.rank(items=greetings, context={ "day_time": 12.0, "language": "en" }) greeting = ranked[0] Trained with appropriate rewards, Improve AI would learn from scratch which greeting is best for each time of day and language. XGBoost Model Improvements Improve AI v8.0 is 90%+ more memory efficient for most use cases. Feature hashing has been replaced with a feature encoding approach that only uses a single feature per item property, substantially improving both training performance as well as ranking / scoring. Ranked Value Encoding Ranked Value Encoding is our novel approach to encoding string values in a manner that is extremely space efficient, accurate, and helps approximate Thompson Sampling for balanced exploration vs exploitation. The concept of Ranked Value Encoding is similar to commonly used Target Value Encoding for encoding string or categorical features. With Target Value Encoding, each string or categorical feature is replaced with the mean of the target values for that string or category. Target Value Encoding tends to provide good results for regression. However, multi-armed bandits are less concerned with the absolute accuracy of the scores and more concerned with the relative scores between items. Since we don’t need the exact target value, we can simply store the relative ranking of the string values, which saves space in the resulting model, increasing performance and lowering distribution costs. Compact String Encoding In conjunction with Ranked Value Encoding, rather than store entire strings, which could be arbitrarily long, Improve AI v8 models only store compact string hashes, resulting in only \~4 bytes per string for typical models. Proven Performance Improve AI is a production ready implementation of a contextual multi-armed bandit algorithm, honed through years of iterative development. By merging Thompson Sampling with XGBoost, it provides a learning system that is both fast and flexible. Thompson Sampling maintains equilibrium between exploring novel possibilities and capitalizing on established options, while XGBoost ensures cost-effective, high-performance training for updated models. Get Started Today Improve AI is available now for Python, Swift, and Java. Check out the Quick-Start Guide for more information. Thank you for your efforts to improve the world a little bit today.

[P]MMML | Deploy HuggingFace training model rapidly based on MetaSpore
reddit
LLM Vibe Score0
Human Vibe Score1
qazmkoppThis week

[P]MMML | Deploy HuggingFace training model rapidly based on MetaSpore

A few days ago, HuggingFace announced a $100 million Series C funding round, which was big news in open source machine learning and could be a sign of where the industry is headed. Two days before the HuggingFace funding announcement, open-source machine learning platform MetaSpore released a demo based on the HuggingFace Rapid deployment pre-training model. As deep learning technology makes innovative breakthroughs in computer vision, natural language processing, speech understanding, and other fields, more and more unstructured data are perceived, understood, and processed by machines. These advances are mainly due to the powerful learning ability of deep learning. Through pre-training of deep models on massive data, the models can capture the internal data patterns, thus helping many downstream tasks. With the industry and academia investing more and more energy in the research of pre-training technology, the distribution warehouses of pre-training models such as HuggingFace and Timm have emerged one after another. The open-source community release pre-training significant model dividends at an unprecedented speed. In recent years, the data form of machine modeling and understanding has gradually evolved from single-mode to multi-mode, and the semantic gap between different modes is being eliminated, making it possible to retrieve data across modes. Take CLIP, OpenAI’s open-source work, as an example, to pre-train the twin towers of images and texts on a dataset of 400 million pictures and texts and connect the semantics between pictures and texts. Many researchers in the academic world have been solving multimodal problems such as image generation and retrieval based on this technology. Although the frontier technology through the semantic gap between modal data, there is still a heavy and complicated model tuning, offline data processing, high performance online reasoning architecture design, heterogeneous computing, and online algorithm be born multiple processes and challenges, hindering the frontier multimodal retrieval technologies fall to the ground and pratt &whitney. DMetaSoul aims at the above technical pain points, abstracting and uniting many links such as model training optimization, online reasoning, and algorithm experiment, forming a set of solutions that can quickly apply offline pre-training model to online. This paper will introduce how to use the HuggingFace community pre-training model to conduct online reasoning and algorithm experiments based on MetaSpore technology ecology so that the benefits of the pre-training model can be fully released to the specific business or industry and small and medium-sized enterprises. And we will give the text search text and text search graph two multimodal retrieval demonstration examples for your reference. Multimodal semantic retrieval The sample architecture of multimodal retrieval is as follows: Our multimodal retrieval system supports both text search and text search application scenarios, including offline processing, model reasoning, online services, and other core modules: ​ https://preview.redd.it/w4v4c7vcez291.png?width=1834&format=png&auto=webp&s=0687efb1fddb26e8e30cb844d398ec712b947f31 Offline processing, including offline data processing processes for different application scenarios of text search and text search, including model tuning, model export, data index database construction, data push, etc. Model inference. After the offline model training, we deployed our NLP and CV large models based on the MetaSpore Serving framework. MetaSpore Serving helps us conveniently perform online inference, elastic scheduling, load balancing, and resource scheduling in heterogeneous environments. Online services. Based on MetaSpore’s online algorithm application framework, MetaSpore has a complete set of reusable online search services, including Front-end retrieval UI, multimodal data preprocessing, vector recall and sorting algorithm, AB experimental framework, etc. MetaSpore also supports text search by text and image scene search by text and can be migrated to other application scenarios at a low cost. The HuggingFace open source community has provided several excellent baseline models for similar multimodal retrieval problems, which are often the starting point for actual optimization in the industry. MetaSpore also uses the pre-training model of the HuggingFace community in its online services of searching words by words and images by words. Searching words by words is based on the semantic similarity model of the question and answer field optimized by MetaSpore, and searching images by words is based on the community pre-training model. These community open source pre-training models are exported to the general ONNX format and loaded into MetaSpore Serving for online reasoning. The following sections will provide a detailed description of the model export and online retrieval algorithm services. The reasoning part of the model is standardized SAAS services with low coupling with the business. Interested readers can refer to my previous post: The design concept of MetaSpore, a new generation of the one-stop machine learning platform. 1.1 Offline Processing Offline processing mainly involves the export and loading of online models and index building and pushing of the document library. You can follow the step-by-step instructions below to complete the offline processing of text search and image search and see how the offline pre-training model achieves reasoning at MetaSpore. 1.1.1 Search text by text Traditional text retrieval systems are based on literal matching algorithms such as BM25. Due to users’ diverse query words, a semantic gap between query words and documents is often encountered. For example, users misspell “iPhone” as “Phone,” and search terms are incredibly long, such as “1 \~ 3 months old baby autumn small size bag pants”. Traditional text retrieval systems will use spelling correction, synonym expansion, search terms rewriting, and other means to alleviate the semantic gap but fundamentally fail to solve this problem. Only when the retrieval system fully understands users’ query terms and documents can it meet users’ retrieval demands at the semantic level. With the continuous progress of pre-training and representational learning technology, some commercial search engines continue to integrate semantic vector retrieval methods based on symbolic learning into the retrieval ecology. Semantic retrieval model This paper introduces a set of semantic vector retrieval applications. MetaSpore built a set of semantic retrieval systems based on encyclopedia question and answer data. MetaSpore adopted the Sentence-Bert model as the semantic vector representation model, which fine-tunes the twin tower BERT in supervised or unsupervised ways to make the model more suitable for retrieval tasks. The model structure is as follows: The query-Doc symmetric two-tower model is used in text search and question and answer retrieval. The vector representation of online Query and offline DOC share the same vector representation model, so it is necessary to ensure the consistency of the offline DOC library building model and online Query inference model. The case uses MetaSpore’s text representation model Sbert-Chinese-QMC-domain-V1, optimized in the open-source semantically similar data set. This model will express the question and answer data as a vector in offline database construction. The user query will be expressed as a vector by this model in online retrieval, ensuring that query-doc in the same semantic space, users’ semantic retrieval demands can be guaranteed by vector similarity metric calculation. Since the text presentation model does vector encoding for Query online, we need to export the model for use by the online service. Go to the q&A data library code directory and export the model concerning the documentation. In the script, Pytorch Tracing is used to export the model. The models are exported to the “./export “directory. The exported models are mainly ONNX models used for wired reasoning, Tokenizer, and related configuration files. The exported models are loaded into MetaSpore Serving by the online Serving system described below for model reasoning. Since the exported model will be copied to the cloud storage, you need to configure related variables in env.sh. \Build library based on text search \ The retrieval database is built on the million-level encyclopedia question and answer data set. According to the description document, you need to download the data and complete the database construction. The question and answer data will be coded as a vector by the offline model, and then the database construction data will be pushed to the service component. The whole process of database construction is described as follows: Preprocessing, converting the original data into a more general JSonline format for database construction; Build index, use the same model as online “sbert-Chinese-qmc-domain-v1” to index documents (one document object per line); Push inverted (vector) and forward (document field) data to each component server. The following is an example of the database data format. After offline database construction is completed, various data are pushed to corresponding service components, such as Milvus storing vector representation of documents and MongoDB storing summary information of documents. Online retrieval algorithm services will use these service components to obtain relevant data. 1.1.2 Search by text Text and images are easy for humans to relate semantically but difficult for machines. First of all, from the perspective of data form, the text is the discrete ID type of one-dimensional data based on words and words. At the same time, images are continuous two-dimensional or three-dimensional data. Secondly, the text is a subjective creation of human beings, and its expressive ability is vibrant, including various turning points, metaphors, and other expressions, while images are machine representations of the objective world. In short, bridging the semantic gap between text and image data is much more complex than searching text by text. The traditional text search image retrieval technology generally relies on the external text description data of the image or the nearest neighbor retrieval technology and carries out the retrieval through the image associated text, which in essence degrades the problem to text search. However, it will also face many issues, such as obtaining the associated text of pictures and whether the accuracy of text search by text is high enough. The depth model has gradually evolved from single-mode to multi-mode in recent years. Taking the open-source project of OpenAI, CLIP, as an example, train the model through the massive image and text data of the Internet and map the text and image data into the same semantic space, making it possible to implement the text and image search technology based on semantic vector. CLIP graphic model The text search pictures introduced in this paper are implemented based on semantic vector retrieval, and the CLIP pre-training model is used as the two-tower retrieval architecture. Because the CLIP model has trained the semantic alignment of the twin towers’ text and image side models on the massive graphic and text data, it is particularly suitable for the text search graph scene. Due to the different image and text data forms, the Query-Doc asymmetric twin towers model is used for text search image retrieval. The image-side model of the twin towers is used for offline database construction, and the text-side model is used for the online return. In the final online retrieval, the database data of the image side model will be searched after the text side model encodes Query, and the CLIP pre-training model guarantees the semantic correlation between images and texts. The model can draw the graphic pairs closer in vector space by pre-training on a large amount of visual data. Here we need to export the text-side model for online MetaSpore Serving inference. Since the retrieval scene is based on Chinese, the CLIP model supporting Chinese understanding is selected. The exported content includes the ONNX model used for online reasoning and Tokenizer, similar to the text search. MetaSpore Serving can load model reasoning through the exported content. Build library on Image search You need to download the Unsplash Lite library data and complete the construction according to the instructions. The whole process of database construction is described as follows: Preprocessing, specify the image directory, and then generate a more general JSOnline file for library construction; Build index, use OpenAI/Clip-Vit-BASE-Patch32 pre-training model to index the gallery, and output one document object for each line of index data; Push inverted (vector) and forward (document field) data to each component server. Like text search, after offline database construction, relevant data will be pushed to service components, called by online retrieval algorithm services to obtain relevant data. 1.2 Online Services The overall online service architecture diagram is as follows: https://preview.redd.it/jfsl8hdfez291.png?width=1280&format=png&auto=webp&s=a858e2304a0c93e78ba5429612ca08cbee69b35a Multi-mode search online service system supports application scenarios such as text search and text search. The whole online service consists of the following parts: Query preprocessing service: encapsulate preprocessing logic (including text/image, etc.) of pre-training model, and provide services through gRPC interface; Retrieval algorithm service: the whole algorithm processing link includes AB experiment tangent flow configuration, MetaSpore Serving call, vector recall, sorting, document summary, etc.; User entry service: provides a Web UI interface for users to debug and track down problems in the retrieval service. From a user request perspective, these services form invocation dependencies from back to front, so to build up a multimodal sample, you need to run each service from front to back first. Before doing this, remember to export the offline model, put it online and build the library first. This article will introduce the various parts of the online service system and make the whole service system step by step according to the following guidance. See the ReadME at the end of this article for more details. 1.2.1 Query preprocessing service Deep learning models tend to be based on tensors, but NLP/CV models often have a preprocessing part that translates raw text and images into tensors that deep learning models can accept. For example, NLP class models often have a pre-tokenizer to transform text data of string type into discrete tensor data. CV class models also have similar processing logic to complete the cropping, scaling, transformation, and other processing of input images through preprocessing. On the one hand, considering that this part of preprocessing logic is decoupled from tensor reasoning of the depth model, on the other hand, the reason of the depth model has an independent technical system based on ONNX, so MetaSpore disassembled this part of preprocessing logic. NLP pretreatment Tokenizer has been integrated into the Query pretreatment service. MetaSpore dismantlement with a relatively general convention. Users only need to provide preprocessing logic files to realize the loading and prediction interface and export the necessary data and configuration files loaded into the preprocessing service. Subsequent CV preprocessing logic will also be integrated in this manner. The preprocessing service currently provides the gRPC interface invocation externally and is dependent on the Query preprocessing (QP) module in the retrieval algorithm service. After the user request reaches the retrieval algorithm service, it will be forwarded to the service to complete the data preprocessing and continue the subsequent processing. The ReadMe provides details on how the preprocessing service is started, how the preprocessing model exported offline to cloud storage enters the service, and how to debug the service. To further improve the efficiency and stability of model reasoning, MetaSpore Serving implements a Python preprocessing submodule. So MetaSpore can provide gRPC services through user-specified preprocessor.py, complete Tokenizer or CV-related preprocessing in NLP, and translate requests into a Tensor that deep models can handle. Finally, the model inference is carried out by MetaSpore, Serving subsequent sub-modules. Presented here on the lot code: https://github.com/meta-soul/MetaSpore/compare/add\python\preprocessor 1.2.2 Retrieval algorithm services Retrieval algorithm service is the core of the whole online service system, which is responsible for the triage of experiments, the assembly of algorithm chains such as preprocessing, recall, sorting, and the invocation of dependent component services. The whole retrieval algorithm service is developed based on the Java Spring framework and supports multi-mode retrieval scenarios of text search and text search graph. Due to good internal abstraction and modular design, it has high flexibility and can be migrated to similar application scenarios at a low cost. Here’s a quick guide to configuring the environment to set up the retrieval algorithm service. See ReadME for more details: Install dependent components. Use Maven to install the online-Serving component Search for service configurations. Copy the template configuration file and replace the MongoDB, Milvus, and other configurations based on the development/production environment. Install and configure Consul. Consul allows you to synchronize the search service configuration in real-time, including cutting the flow of experiments, recall parameters, and sorting parameters. The project’s configuration file shows the current configuration parameters of text search and text search. The parameter modelName in the stage of pretreatment and recall is the corresponding model exported in offline processing. Start the service. Once the above configuration is complete, the retrieval service can be started from the entry script. Once the service is started, you can test it! For example, for a user with userId=10 who wants to query “How to renew ID card,” access the text search service. 1.2.3 User Entry Service Considering that the retrieval algorithm service is in the form of the API interface, it is difficult to locate and trace the problem, especially for the text search image scene can intuitively display the retrieval results to facilitate the iterative optimization of the retrieval algorithm. This paper provides a lightweight Web UI interface for text search and image search, a search input box, and results in a display page for users. Developed by Flask, the service can be easily integrated with other retrieval applications. The service calls the retrieval algorithm service and displays the returned results on the page. It’s also easy to install and start the service. Once you’re done, go to http://127.0.0.1:8090 to see if the search UI service is working correctly. See the ReadME at the end of this article for details. Multimodal system demonstration The multimodal retrieval service can be started when offline processing and online service environment configuration have been completed following the above instructions. Examples of textual searches are shown below. Enter the entry of the text search map application, enter “cat” first, and you can see that the first three digits of the returned result are cats: https://preview.redd.it/0n5nuyvhez291.png?width=1280&format=png&auto=webp&s=1e9c054f541d53381674b8d6001b4bf524506bd2 If you add a color constraint to “cat” to retrieve “black cat,” you can see that it does return a black cat: https://preview.redd.it/rzc0qjyjez291.png?width=1280&format=png&auto=webp&s=d5bcc503ef0fb3360c7740e60e295cf372dcad47 Further, strengthen the constraint on the search term, change it to “black cat on the bed,” and return results containing pictures of a black cat climbing on the bed: ​ https://preview.redd.it/c4b2q8olez291.png?width=1280&format=png&auto=webp&s=4f3817b0b9f07e1e68d1d4a8281702ba3834a00a The cat can still be found through the text search system after the color and scene modification in the above example. Conclusion The cutting-edge pre-training technology can bridge the semantic gap between different modes, and the HuggingFace community can greatly reduce the cost for developers to use the pre-training model. Combined with the technological ecology of MetaSpore online reasoning and online microservices provided by DMetaSpore, the pre-training model is no longer mere offline dabbling. Instead, it can truly achieve end-to-end implementation from cutting-edge technology to industrial scenarios, fully releasing the dividends of the pre-training large model. In the future, DMetaSoul will continue to improve and optimize the MetaSpore technology ecosystem: More automated and wider access to HuggingFace community ecology. MetaSpore will soon release a common model rollout mechanism to make HuggingFace ecologically accessible and will later integrate preprocessing services into online services. Multi-mode retrieval offline algorithm optimization. For multimodal retrieval scenarios, MetaSpore will continuously iteratively optimize offline algorithm components, including text recall/sort model, graphic recall/sort model, etc., to improve the accuracy and efficiency of the retrieval algorithm. For related code and reference documentation in this article, please visit: https://github.com/meta-soul/MetaSpore/tree/main/demo/multimodal/online Some images source: https://github.com/openai/CLIP/raw/main/CLIP.png https://www.sbert.net/examples/training/sts/README.html

[News] AAAI 2025 Workshop on AI for Music 🎶
reddit
LLM Vibe Score0
Human Vibe Score0
Saysike_rightnow69This week

[News] AAAI 2025 Workshop on AI for Music 🎶

Hi everyone! We’re hosting the first “AI for Music” workshop at AAAI on March 3, 2025. The workshop will explore how AI is transforming music creation, recognition, education, and more. Topics include AI-driven composition, sound design, legal and ethical challenges, and AI’s impact on musicians’ careers. Submissions (up to 6 pages) are welcome until November 22, 2024. Work in progress is encouraged! Workshop Summary This one-day workshop will explore the dynamic intersection of artificial intelligence and music. It explores how AI is transforming music creation, recognition, and education, ethical and legal implications, as well as business opportunities. We will investigate how AI is changing the music industry and education—from composition to performance, production, collaboration, and audience experience. Participants will gain insights into the technological challenges in music and how AI can enhance creativity, enabling musicians and producers to push the boundaries of their art. The workshop will cover topics such as AI-driven music composition, where algorithms generate melodies, harmonies, and even full orchestral arrangements. We will discuss how AI tools assist in sound design, remixing, and mastering, allowing for new sonic possibilities and efficiencies in music production. Additionally, we'll examine AI's impact on music education and the careers of musicians, exploring advanced learning tools and teaching methods. AI technologies are increasingly adopted in the music and entertainment industry. The workshop will also discuss the legal and ethical implications of AI in music, including questions of authorship, originality, and the evolving role of human artists in an increasingly automated world. This workshop is designed for AI researchers, musicians, producers, and educators interested in the current status and future of AI in music. Call for Papers Submissions should be a maximum of 6 pages. Work in progress is welcome. Authors are encouraged to include descriptions of their prototype implementations. Additionally, authors are encouraged to interact with workshop attendees by including posters or demonstrations at the end of the workshop. Conceptual designs without any evidence of practical implementation are discouraged. Topics of interest are (but not limited to) AI-Driven Music Composition and Generation AI in Music Practice and Performance AI-based Music Recognition and Transcription AI Applications in Sound Design AI-Generated Videos and Lyrics Based on Music Legal and Ethical Implications of AI in Music AI’s Impact on Musicians’ Careers and Education Business Opportunities of AI in Music Music Datasets and Data Analysis Important Dates Submission Deadline: November 22, 2024 Notification: December 9, 2024 Final Version Due: December 31, 2024 We hope to see you there! 🎶

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.
reddit
LLM Vibe Score0
Human Vibe Score0.765
hardmaruThis week

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.

Schmidhuber interview expressing his views on the future of AI and AGI. Original source. I think the interview is of interest to r/MachineLearning, and presents an alternate view, compared to other influential leaders in AI. Juergen Schmidhuber, Renowned 'Father Of Modern AI,' Says His Life’s Work Won't Lead To Dystopia May 23, 2023. Contributed by Hessie Jones. Amid the growing concern about the impact of more advanced artificial intelligence (AI) technologies on society, there are many in the technology community who fear the implications of the advancements in Generative AI if they go unchecked. Dr. Juergen Schmidhuber, a renowned scientist, artificial intelligence researcher and widely regarded as one of the pioneers in the field, is more optimistic. He declares that many of those who suddenly warn against the dangers of AI are just seeking publicity, exploiting the media’s obsession with killer robots which has attracted more attention than “good AI” for healthcare etc. The potential to revolutionize various industries and improve our lives is clear, as are the equal dangers if bad actors leverage the technology for personal gain. Are we headed towards a dystopian future, or is there reason to be optimistic? I had a chance to sit down with Dr. Juergen Schmidhuber to understand his perspective on this seemingly fast-moving AI-train that will leap us into the future. As a teenager in the 1970s, Juergen Schmidhuber became fascinated with the idea of creating intelligent machines that could learn and improve on their own, becoming smarter than himself within his lifetime. This would ultimately lead to his groundbreaking work in the field of deep learning. In the 1980s, he studied computer science at the Technical University of Munich (TUM), where he earned his diploma in 1987. His thesis was on the ultimate self-improving machines that, not only, learn through some pre-wired human-designed learning algorithm, but also learn and improve the learning algorithm itself. Decades later, this became a hot topic. He also received his Ph.D. at TUM in 1991 for work that laid some of the foundations of modern AI. Schmidhuber is best known for his contributions to the development of recurrent neural networks (RNNs), the most powerful type of artificial neural network that can process sequential data such as speech and natural language. With his students Sepp Hochreiter, Felix Gers, Alex Graves, Daan Wierstra, and others, he published architectures and training algorithms for the long short-term memory (LSTM), a type of RNN that is widely used in natural language processing, speech recognition, video games, robotics, and other applications. LSTM has become the most cited neural network of the 20th century, and Business Week called it "arguably the most commercial AI achievement." Throughout his career, Schmidhuber has received various awards and accolades for his groundbreaking work. In 2013, he was awarded the Helmholtz Prize, which recognizes significant contributions to the field of machine learning. In 2016, he was awarded the IEEE Neural Network Pioneer Award for "pioneering contributions to deep learning and neural networks." The media have often called him the “father of modern AI,” because the most cited neural networks all build on his lab’s work. He is quick to point out, however, that AI history goes back centuries. Despite his many accomplishments, at the age of 60, he feels mounting time pressure towards building an Artificial General Intelligence within his lifetime and remains committed to pushing the boundaries of AI research and development. He is currently director of the KAUST AI Initiative, scientific director of the Swiss AI Lab IDSIA, and co-founder and chief scientist of AI company NNAISENSE, whose motto is "AI∀" which is a math-inspired way of saying "AI For All." He continues to work on cutting-edge AI technologies and applications to improve human health and extend human lives and make lives easier for everyone. The following interview has been edited for clarity. Jones: Thank you Juergen for joining me. You have signed letters warning about AI weapons. But you didn't sign the recent publication, "Pause Gigantic AI Experiments: An Open Letter"? Is there a reason? Schmidhuber: Thank you Hessie. Glad to speak with you. I have realized that many of those who warn in public against the dangers of AI are just seeking publicity. I don't think the latest letter will have any significant impact because many AI researchers, companies, and governments will ignore it completely. The proposal frequently uses the word "we" and refers to "us," the humans. But as I have pointed out many times in the past, there is no "we" that everyone can identify with. Ask 10 different people, and you will hear 10 different opinions about what is "good." Some of those opinions will be completely incompatible with each other. Don't forget the enormous amount of conflict between the many people. The letter also says, "If such a pause cannot be quickly put in place, governments should intervene and impose a moratorium." The problem is that different governments have ALSO different opinions about what is good for them and for others. Great Power A will say, if we don't do it, Great Power B will, perhaps secretly, and gain an advantage over us. The same is true for Great Powers C and D. Jones: Everyone acknowledges this fear surrounding current generative AI technology. Moreover, the existential threat of this technology has been publicly acknowledged by Sam Altman, CEO of OpenAI himself, calling for AI regulation. From your perspective, is there an existential threat? Schmidhuber: It is true that AI can be weaponized, and I have no doubt that there will be all kinds of AI arms races, but AI does not introduce a new quality of existential threat. The threat coming from AI weapons seems to pale in comparison to the much older threat from nuclear hydrogen bombs that don’t need AI at all. We should be much more afraid of half-century-old tech in the form of H-bomb rockets. The Tsar Bomba of 1961 had almost 15 times more destructive power than all weapons of WW-II combined. Despite the dramatic nuclear disarmament since the 1980s, there are still more than enough nuclear warheads to wipe out human civilization within two hours, without any AI I’m much more worried about that old existential threat than the rather harmless AI weapons. Jones: I realize that while you compare AI to the threat of nuclear bombs, there is a current danger that a current technology can be put in the hands of humans and enable them to “eventually” exact further harms to individuals of group in a very precise way, like targeted drone attacks. You are giving people a toolset that they've never had before, enabling bad actors, as some have pointed out, to be able to do a lot more than previously because they didn't have this technology. Schmidhuber: Now, all that sounds horrible in principle, but our existing laws are sufficient to deal with these new types of weapons enabled by AI. If you kill someone with a gun, you will go to jail. Same if you kill someone with one of these drones. Law enforcement will get better at understanding new threats and new weapons and will respond with better technology to combat these threats. Enabling drones to target persons from a distance in a way that requires some tracking and some intelligence to perform, which has traditionally been performed by skilled humans, to me, it seems is just an improved version of a traditional weapon, like a gun, which is, you know, a little bit smarter than the old guns. But, in principle, all of that is not a new development. For many centuries, we have had the evolution of better weaponry and deadlier poisons and so on, and law enforcement has evolved their policies to react to these threats over time. So, it's not that we suddenly have a new quality of existential threat and it's much more worrisome than what we have had for about six decades. A large nuclear warhead doesn’t need fancy face recognition to kill an individual. No, it simply wipes out an entire city with ten million inhabitants. Jones: The existential threat that’s implied is the extent to which humans have control over this technology. We see some early cases of opportunism which, as you say, tends to get more media attention than positive breakthroughs. But you’re implying that this will all balance out? Schmidhuber: Historically, we have a long tradition of technological breakthroughs that led to advancements in weapons for the purpose of defense but also for protection. From sticks, to rocks, to axes to gunpowder to cannons to rockets… and now to drones… this has had a drastic influence on human history but what has been consistent throughout history is that those who are using technology to achieve their own ends are themselves, facing the same technology because the opposing side is learning to use it against them. And that's what has been repeated in thousands of years of human history and it will continue. I don't see the new AI arms race as something that is remotely as existential a threat as the good old nuclear warheads. You said something important, in that some people prefer to talk about the downsides rather than the benefits of this technology, but that's misleading, because 95% of all AI research and AI development is about making people happier and advancing human life and health. Jones: Let’s touch on some of those beneficial advances in AI research that have been able to radically change present day methods and achieve breakthroughs. Schmidhuber: All right! For example, eleven years ago, our team with my postdoc Dan Ciresan was the first to win a medical imaging competition through deep learning. We analyzed female breast cells with the objective to determine harmless cells vs. those in the pre-cancer stage. Typically, a trained oncologist needs a long time to make these determinations. Our team, who knew nothing about cancer, were able to train an artificial neural network, which was totally dumb in the beginning, on lots of this kind of data. It was able to outperform all the other methods. Today, this is being used not only for breast cancer, but also for radiology and detecting plaque in arteries, and many other things. Some of the neural networks that we have developed in the last 3 decades are now prevalent across thousands of healthcare applications, detecting Diabetes and Covid-19 and what not. This will eventually permeate across all healthcare. The good consequences of this type of AI are much more important than the click-bait new ways of conducting crimes with AI. Jones: Adoption is a product of reinforced outcomes. The massive scale of adoption either leads us to believe that people have been led astray, or conversely, technology is having a positive effect on people’s lives. Schmidhuber: The latter is the likely case. There's intense commercial pressure towards good AI rather than bad AI because companies want to sell you something, and you are going to buy only stuff you think is going to be good for you. So already just through this simple, commercial pressure, you have a tremendous bias towards good AI rather than bad AI. However, doomsday scenarios like in Schwarzenegger movies grab more attention than documentaries on AI that improve people’s lives. Jones: I would argue that people are drawn to good stories – narratives that contain an adversary and struggle, but in the end, have happy endings. And this is consistent with your comment on human nature and how history, despite its tendency for violence and destruction of humanity, somehow tends to correct itself. Let’s take the example of a technology, which you are aware – GANs – General Adversarial Networks, which today has been used in applications for fake news and disinformation. In actuality, the purpose in the invention of GANs was far from what it is used for today. Schmidhuber: Yes, the name GANs was created in 2014 but we had the basic principle already in the early 1990s. More than 30 years ago, I called it artificial curiosity. It's a very simple way of injecting creativity into a little two network system. This creative AI is not just trying to slavishly imitate humans. Rather, it’s inventing its own goals. Let me explain: You have two networks. One network is producing outputs that could be anything, any action. Then the second network is looking at these actions and it’s trying to predict the consequences of these actions. An action could move a robot, then something happens, and the other network is just trying to predict what will happen. Now we can implement artificial curiosity by reducing the prediction error of the second network, which, at the same time, is the reward of the first network. The first network wants to maximize its reward and so it will invent actions that will lead to situations that will surprise the second network, which it has not yet learned to predict well. In the case where the outputs are fake images, the first network will try to generate images that are good enough to fool the second network, which will attempt to predict the reaction of the environment: fake or real image, and it will try to become better at it. The first network will continue to also improve at generating images whose type the second network will not be able to predict. So, they fight each other. The 2nd network will continue to reduce its prediction error, while the 1st network will attempt to maximize it. Through this zero-sum game the first network gets better and better at producing these convincing fake outputs which look almost realistic. So, once you have an interesting set of images by Vincent Van Gogh, you can generate new images that leverage his style, without the original artist having ever produced the artwork himself. Jones: I see how the Van Gogh example can be applied in an education setting and there are countless examples of artists mimicking styles from famous painters but image generation from this instance that can happen within seconds is quite another feat. And you know this is how GANs has been used. What’s more prevalent today is a socialized enablement of generating images or information to intentionally fool people. It also surfaces new harms that deal with the threat to intellectual property and copyright, where laws have yet to account for. And from your perspective this was not the intention when the model was conceived. What was your motivation in your early conception of what is now GANs? Schmidhuber: My old motivation for GANs was actually very important and it was not to create deepfakes or fake news but to enable AIs to be curious and invent their own goals, to make them explore their environment and make them creative. Suppose you have a robot that executes one action, then something happens, then it executes another action, and so on, because it wants to achieve certain goals in the environment. For example, when the battery is low, this will trigger “pain” through hunger sensors, so it wants to go to the charging station, without running into obstacles, which will trigger other pain sensors. It will seek to minimize pain (encoded through numbers). Now the robot has a friend, the second network, which is a world model ––it’s a prediction machine that learns to predict the consequences of the robot’s actions. Once the robot has a good model of the world, it can use it for planning. It can be used as a simulation of the real world. And then it can determine what is a good action sequence. If the robot imagines this sequence of actions, the model will predict a lot of pain, which it wants to avoid. If it plays this alternative action sequence in its mental model of the world, then it will predict a rewarding situation where it’s going to sit on the charging station and its battery is going to load again. So, it'll prefer to execute the latter action sequence. In the beginning, however, the model of the world knows nothing, so how can we motivate the first network to generate experiments that lead to data that helps the world model learn something it didn’t already know? That’s what artificial curiosity is about. The dueling two network systems effectively explore uncharted environments by creating experiments so that over time the curious AI gets a better sense of how the environment works. This can be applied to all kinds of environments, and has medical applications. Jones: Let’s talk about the future. You have said, “Traditional humans won’t play a significant role in spreading intelligence across the universe.” Schmidhuber: Let’s first conceptually separate two types of AIs. The first type of AI are tools directed by humans. They are trained to do specific things like accurately detect diabetes or heart disease and prevent attacks before they happen. In these cases, the goal is coming from the human. More interesting AIs are setting their own goals. They are inventing their own experiments and learning from them. Their horizons expand and eventually they become more and more general problem solvers in the real world. They are not controlled by their parents, but much of what they learn is through self-invented experiments. A robot, for example, is rotating a toy, and as it is doing this, the video coming in through the camera eyes, changes over time and it begins to learn how this video changes and learns how the 3D nature of the toy generates certain videos if you rotate it a certain way, and eventually, how gravity works, and how the physics of the world works. Like a little scientist! And I have predicted for decades that future scaled-up versions of such AI scientists will want to further expand their horizons, and eventually go where most of the physical resources are, to build more and bigger AIs. And of course, almost all of these resources are far away from earth out there in space, which is hostile to humans but friendly to appropriately designed AI-controlled robots and self-replicating robot factories. So here we are not talking any longer about our tiny biosphere; no, we are talking about the much bigger rest of the universe. Within a few tens of billions of years, curious self-improving AIs will colonize the visible cosmos in a way that’s infeasible for humans. Those who don’t won’t have an impact. Sounds like science fiction, but since the 1970s I have been unable to see a plausible alternative to this scenario, except for a global catastrophe such as an all-out nuclear war that stops this development before it takes off. Jones: How long have these AIs, which can set their own goals — how long have they existed? To what extent can they be independent of human interaction? Schmidhuber: Neural networks like that have existed for over 30 years. My first simple adversarial neural network system of this kind is the one from 1990 described above. You don’t need a teacher there; it's just a little agent running around in the world and trying to invent new experiments that surprise its own prediction machine. Once it has figured out certain parts of the world, the agent will become bored and will move on to more exciting experiments. The simple 1990 systems I mentioned have certain limitations, but in the past three decades, we have also built more sophisticated systems that are setting their own goals and such systems I think will be essential for achieving true intelligence. If you are only imitating humans, you will never go beyond them. So, you really must give AIs the freedom to explore previously unexplored regions of the world in a way that no human is really predefining. Jones: Where is this being done today? Schmidhuber: Variants of neural network-based artificial curiosity are used today for agents that learn to play video games in a human-competitive way. We have also started to use them for automatic design of experiments in fields such as materials science. I bet many other fields will be affected by it: chemistry, biology, drug design, you name it. However, at least for now, these artificial scientists, as I like to call them, cannot yet compete with human scientists. I don’t think it’s going to stay this way but, at the moment, it’s still the case. Sure, AI has made a lot of progress. Since 1997, there have been superhuman chess players, and since 2011, through the DanNet of my team, there have been superhuman visual pattern recognizers. But there are other things where humans, at the moment at least, are much better, in particular, science itself. In the lab we have many first examples of self-directed artificial scientists, but they are not yet convincing enough to appear on the radar screen of the public space, which is currently much more fascinated with simpler systems that just imitate humans and write texts based on previously seen human-written documents. Jones: You speak of these numerous instances dating back 30 years of these lab experiments where these self-driven agents are deciding and learning and moving on once they’ve learned. And I assume that that rate of learning becomes even faster over time. What kind of timeframe are we talking about when this eventually is taken outside of the lab and embedded into society? Schmidhuber: This could still take months or even years :-) Anyway, in the not-too-distant future, we will probably see artificial scientists who are good at devising experiments that allow them to discover new, previously unknown physical laws. As always, we are going to profit from the old trend that has held at least since 1941: every decade compute is getting 100 times cheaper. Jones: How does this trend affect modern AI such as ChatGPT? Schmidhuber: Perhaps you know that all the recent famous AI applications such as ChatGPT and similar models are largely based on principles of artificial neural networks invented in the previous millennium. The main reason why they works so well now is the incredible acceleration of compute per dollar. ChatGPT is driven by a neural network called “Transformer” described in 2017 by Google. I am happy about that because a quarter century earlier in 1991 I had a particular Transformer variant which is now called the “Transformer with linearized self-attention”. Back then, not much could be done with it, because the compute cost was a million times higher than today. But today, one can train such models on half the internet and achieve much more interesting results. Jones: And for how long will this acceleration continue? Schmidhuber: There's no reason to believe that in the next 30 years, we won't have another factor of 1 million and that's going to be really significant. In the near future, for the first time we will have many not-so expensive devices that can compute as much as a human brain. The physical limits of computation, however, are much further out so even if the trend of a factor of 100 every decade continues, the physical limits (of 1051 elementary instructions per second and kilogram of matter) won’t be hit until, say, the mid-next century. Even in our current century, however, we’ll probably have many machines that compute more than all 10 billion human brains collectively and you can imagine, everything will change then! Jones: That is the big question. Is everything going to change? If so, what do you say to the next generation of leaders, currently coming out of college and university. So much of this change is already impacting how they study, how they will work, or how the future of work and livelihood is defined. What is their purpose and how do we change our systems so they will adapt to this new version of intelligence? Schmidhuber: For decades, people have asked me questions like that, because you know what I'm saying now, I have basically said since the 1970s, it’s just that today, people are paying more attention because, back then, they thought this was science fiction. They didn't think that I would ever come close to achieving my crazy life goal of building a machine that learns to become smarter than myself such that I can retire. But now many have changed their minds and think it's conceivable. And now I have two daughters, 23 and 25. People ask me: what do I tell them? They know that Daddy always said, “It seems likely that within your lifetimes, you will have new types of intelligence that are probably going to be superior in many ways, and probably all kinds of interesting ways.” How should they prepare for that? And I kept telling them the obvious: Learn how to learn new things! It's not like in the previous millennium where within 20 years someone learned to be a useful member of society, and then took a job for 40 years and performed in this job until she received her pension. Now things are changing much faster and we must learn continuously just to keep up. I also told my girls that no matter how smart AIs are going to get, learn at least the basics of math and physics, because that’s the essence of our universe, and anybody who understands this will have an advantage, and learn all kinds of new things more easily. I also told them that social skills will remain important, because most future jobs for humans will continue to involve interactions with other humans, but I couldn’t teach them anything about that; they know much more about social skills than I do. You touched on the big philosophical question about people’s purpose. Can this be answered without answering the even grander question: What’s the purpose of the entire universe? We don’t know. But what’s happening right now might be connected to the unknown answer. Don’t think of humans as the crown of creation. Instead view human civilization as part of a much grander scheme, an important step (but not the last one) on the path of the universe from very simple initial conditions towards more and more unfathomable complexity. Now it seems ready to take its next step, a step comparable to the invention of life itself over 3.5 billion years ago. Alas, don’t worry, in the end, all will be good! Jones: Let’s get back to this transformation happening right now with OpenAI. There are many questioning the efficacy and accuracy of ChatGPT, and are concerned its release has been premature. In light of the rampant adoption, educators have banned its use over concerns of plagiarism and how it stifles individual development. Should large language models like ChatGPT be used in school? Schmidhuber: When the calculator was first introduced, instructors forbade students from using it in school. Today, the consensus is that kids should learn the basic methods of arithmetic, but they should also learn to use the “artificial multipliers” aka calculators, even in exams, because laziness and efficiency is a hallmark of intelligence. Any intelligent being wants to minimize its efforts to achieve things. And that's the reason why we have tools, and why our kids are learning to use these tools. The first stone tools were invented maybe 3.5 million years ago; tools just have become more sophisticated over time. In fact, humans have changed in response to the properties of their tools. Our anatomical evolution was shaped by tools such as spears and fire. So, it's going to continue this way. And there is no permanent way of preventing large language models from being used in school. Jones: And when our children, your children graduate, what does their future work look like? Schmidhuber: A single human trying to predict details of how 10 billion people and their machines will evolve in the future is like a single neuron in my brain trying to predict what the entire brain and its tens of billions of neurons will do next year. 40 years ago, before the WWW was created at CERN in Switzerland, who would have predicted all those young people making money as YouTube video bloggers? Nevertheless, let’s make a few limited job-related observations. For a long time, people have thought that desktop jobs may require more intelligence than skills trade or handicraft professions. But now, it turns out that it's much easier to replace certain aspects of desktop jobs than replacing a carpenter, for example. Because everything that works well in AI is happening behind the screen currently, but not so much in the physical world. There are now artificial systems that can read lots of documents and then make really nice summaries of these documents. That is a desktop job. Or you give them a description of an illustration that you want to have for your article and pretty good illustrations are being generated that may need some minimal fine-tuning. But you know, all these desktop jobs are much easier to facilitate than the real tough jobs in the physical world. And it's interesting that the things people thought required intelligence, like playing chess, or writing or summarizing documents, are much easier for machines than they thought. But for things like playing football or soccer, there is no physical robot that can remotely compete with the abilities of a little boy with these skills. So, AI in the physical world, interestingly, is much harder than AI behind the screen in virtual worlds. And it's really exciting, in my opinion, to see that jobs such as plumbers are much more challenging than playing chess or writing another tabloid story. Jones: The way data has been collected in these large language models does not guarantee personal information has not been excluded. Current consent laws already are outdated when it comes to these large language models (LLM). The concern, rightly so, is increasing surveillance and loss of privacy. What is your view on this? Schmidhuber: As I have indicated earlier: are surveillance and loss of privacy inevitable consequences of increasingly complex societies? Super-organisms such as cities and states and companies consist of numerous people, just like people consist of numerous cells. These cells enjoy little privacy. They are constantly monitored by specialized "police cells" and "border guard cells": Are you a cancer cell? Are you an external intruder, a pathogen? Individual cells sacrifice their freedom for the benefits of being part of a multicellular organism. Similarly, for super-organisms such as nations. Over 5000 years ago, writing enabled recorded history and thus became its inaugural and most important invention. Its initial purpose, however, was to facilitate surveillance, to track citizens and their tax payments. The more complex a super-organism, the more comprehensive its collection of information about its constituents. 200 years ago, at least, the parish priest in each village knew everything about all the village people, even about those who did not confess, because they appeared in the confessions of others. Also, everyone soon knew about the stranger who had entered the village, because some occasionally peered out of the window, and what they saw got around. Such control mechanisms were temporarily lost through anonymization in rapidly growing cities but are now returning with the help of new surveillance devices such as smartphones as part of digital nervous systems that tell companies and governments a lot about billions of users. Cameras and drones etc. are becoming increasingly tinier and more ubiquitous. More effective recognition of faces and other detection technology are becoming cheaper and cheaper, and many will use it to identify others anywhere on earth; the big wide world will not offer any more privacy than the local village. Is this good or bad? Some nations may find it easier than others to justify more complex kinds of super-organisms at the expense of the privacy rights of their constituents. Jones: So, there is no way to stop or change this process of collection, or how it continuously informs decisions over time? How do you see governance and rules responding to this, especially amid Italy’s ban on ChatGPT following suspected user data breach and the more recent news about the Meta’s record $1.3billion fine in the company’s handling of user information? Schmidhuber: Data collection has benefits and drawbacks, such as the loss of privacy. How to balance those? I have argued for addressing this through data ownership in data markets. If it is true that data is the new oil, then it should have a price, just like oil. At the moment, the major surveillance platforms such as Meta do not offer users any money for their data and the transitive loss of privacy. In the future, however, we will likely see attempts at creating efficient data markets to figure out the data's true financial value through the interplay between supply and demand. Even some of the sensitive medical data should not be priced by governmental regulators but by patients (and healthy persons) who own it and who may sell or license parts thereof as micro-entrepreneurs in a healthcare data market. Following a previous interview, I gave for one of the largest re-insurance companies , let's look at the different participants in such a data market: patients, hospitals, data companies. (1) Patients with a rare form of cancer can offer more valuable data than patients with a very common form of cancer. (2) Hospitals and their machines are needed to extract the data, e.g., through magnet spin tomography, radiology, evaluations through human doctors, and so on. (3) Companies such as Siemens, Google or IBM would like to buy annotated data to make better artificial neural networks that learn to predict pathologies and diseases and the consequences of therapies. Now the market’s invisible hand will decide about the data’s price through the interplay between demand and supply. On the demand side, you will have several companies offering something for the data, maybe through an app on the smartphone (a bit like a stock market app). On the supply side, each patient in this market should be able to profit from high prices for rare valuable types of data. Likewise, competing data extractors such as hospitals will profit from gaining recognition and trust for extracting data well at a reasonable price. The market will make the whole system efficient through incentives for all who are doing a good job. Soon there will be a flourishing ecosystem of commercial data market advisors and what not, just like the ecosystem surrounding the traditional stock market. The value of the data won’t be determined by governments or ethics committees, but by those who own the data and decide by themselves which parts thereof they want to license to others under certain conditions. At first glance, a market-based system seems to be detrimental to the interest of certain monopolistic companies, as they would have to pay for the data - some would prefer free data and keep their monopoly. However, since every healthy and sick person in the market would suddenly have an incentive to collect and share their data under self-chosen anonymity conditions, there will soon be many more useful data to evaluate all kinds of treatments. On average, people will live longer and healthier, and many companies and the entire healthcare system will benefit. Jones: Finally, what is your view on open source versus the private companies like Google and OpenAI? Is there a danger to supporting these private companies’ large language models versus trying to keep these models open source and transparent, very much like what LAION is doing? Schmidhuber: I signed this open letter by LAION because I strongly favor the open-source movement. And I think it's also something that is going to challenge whatever big tech dominance there might be at the moment. Sure, the best models today are run by big companies with huge budgets for computers, but the exciting fact is that open-source models are not so far behind, some people say maybe six to eight months only. Of course, the private company models are all based on stuff that was created in academia, often in little labs without so much funding, which publish without patenting their results and open source their code and others take it and improved it. Big tech has profited tremendously from academia; their main achievement being that they have scaled up everything greatly, sometimes even failing to credit the original inventors. So, it's very interesting to see that as soon as some big company comes up with a new scaled-up model, lots of students out there are competing, or collaborating, with each other, trying to come up with equal or better performance on smaller networks and smaller machines. And since they are open sourcing, the next guy can have another great idea to improve it, so now there’s tremendous competition also for the big companies. Because of that, and since AI is still getting exponentially cheaper all the time, I don't believe that big tech companies will dominate in the long run. They find it very hard to compete with the enormous open-source movement. As long as you can encourage the open-source community, I think you shouldn't worry too much. Now, of course, you might say if everything is open source, then the bad actors also will more easily have access to these AI tools. And there's truth to that. But as always since the invention of controlled fire, it was good that knowledge about how technology works quickly became public such that everybody could use it. And then, against any bad actor, there's almost immediately a counter actor trying to nullify his efforts. You see, I still believe in our old motto "AI∀" or "AI For All." Jones: Thank you, Juergen for sharing your perspective on this amazing time in history. It’s clear that with new technology, the enormous potential can be matched by disparate and troubling risks which we’ve yet to solve, and even those we have yet to identify. If we are to dispel the fear of a sentient system for which we have no control, humans, alone need to take steps for more responsible development and collaboration to ensure AI technology is used to ultimately benefit society. Humanity will be judged by what we do next.

[D] A Jobless Rant - ML is a Fool's Gold
reddit
LLM Vibe Score0
Human Vibe Score1
good_riceThis week

[D] A Jobless Rant - ML is a Fool's Gold

Aside from the clickbait title, I am earnestly looking for some advice and discussion from people who are actually employed. That being said, here's my gripe: I have been relentlessly inundated by the words "AI, ML, Big Data" throughout my undergrad from other CS majors, business and sales oriented people, media, and .ai type startups. It seems like everyone was peddling ML as the go to solution, the big money earner, and the future of the field. I've heard college freshman ask stuff like, "if I want to do CS, am I going to need to learn ML to be relevant" - if you're on this sub, I probably do not need to continue to elaborate on just how ridiculous the ML craze is. Every single university has opened up ML departments or programs and are pumping out ML graduates at an unprecedented rate. Surely, there'd be a job market to meet the incredible supply of graduates and cultural interest? Swept up in a mixture of genuine interest and hype, I decided to pursue computer vision. I majored in Math-CS at a top-10 CS university (based on at least one arbitrary ranking). I had three computer vision internships, two at startups, one at NASA JPL, in each doing non-trivial CV work; I (re)implemented and integrated CV systems from mixtures of recently published papers. I have a bunch of projects showing both CV and CS fundamentals (OS, networking, data structures, algorithms, etc) knowledge. I have taken graduate level ML coursework. I was accepted to Carnegie Mellon for an MS in Computer Vision, but I deferred to 2021 - all in all, I worked my ass off to try to simultaneously get a solid background in math AND computer science AND computer vision. That brings me to where I am now, which is unemployed and looking for jobs. Almost every single position I have seen requires a PhD and/or 5+ years of experience, and whatever I have applied for has ghosted me so far. The notion that ML is a high paying in-demand field seems to only be true if your name is Andrej Karpathy - and I'm only sort of joking. It seems like unless you have a PhD from one of the big 4 in CS and multiple publications in top tier journals you're out of luck, or at least vying for one of the few remaining positions at small companies. This seems normalized in ML, but this is not the case for quite literally every other subfield or even generalized CS positions. Getting a high paying job at a Big N company is possible as a new grad with just a bachelors and general SWE knowledge, and there are a plethora of positions elsewhere. Getting the equivalent with basically every specialization, whether operating systems, distributed systems, security, networking, etc, is also possible, and doesn't require 5 CVPR publications. TL;DR From my personal perspective, if you want to do ML because of career prospects, salaries, or job security, pick almost any other CS specialization. In ML, you'll find yourself working 2x as hard through difficult theory and math to find yourself competing with more applicants for fewer positions. I am absolutely complaining and would love to hear a more positive perspective, but in the meanwhile I'll be applying to jobs, working on more post-grad projects, and contemplating switching fields.

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.
reddit
LLM Vibe Score0
Human Vibe Score0.6
AlexSnakeKingThis week

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.

TD;LR: At Company A, Team X does advanced analytics using on-prem ERP tools and older programming languages. Their tools work very well and are designed based on very deep business and domain expertise. Team Y is a new and ambitious Data Science team that thinks they can replace Team X's tools with a bunch of R scripts and a custom built ML platform. Their models are simplistic, but more "fashionable" compared to the econometric models used by Team X, and team Y benefits from the ML/DS moniker so leadership is allowing Team Y to start a large scale overhaul of the analytics platform in question. Team Y doesn't have the experience for such a larger scale transformation, and is refusing to collaborate with team X. This project is very likely going to fail, and cause serious harm to the company as a whole financially and from a people perspective. I argue that this is not just because of bad leadership, but also because of various trends and mindsets in the DS community at large. Update (Jump to below the line for the original story): Several people in the comments are pointing out that this just a management failure, not something due to ML/DS, and that you can replace DS with any buzz tech and the story will still be relevant. My response: Of course, any failure at an organization level is ultimately a management failure one way or the other. Moreover, it is also the case that ML/DS when done correctly, will always improve a company's bottom line. There is no scenario where the proper ML solution, delivered at a reasonable cost and in a timely fashion, will somehow hurt the company's bottom line. My point is that in this case management is failing because of certain trends and practices that are specific to the ML/DS community, namely: The idea that DS teams should operate independently of tech and business orgs -- too much autonomy for DS teams The disregard for domain knowledge that seems prevalent nowadays thanks to the ML hype, that DS can be generalists and someone with good enough ML chops can solve any business problem. That wasn't the case when I first left academia for the industry in 2009 (back then nobody would even bother with a phone screen if you didn't have the right domain knowledge). Over reliance on resources who check all the ML hype related boxes (knows Python, R, Tensorflow, Shiny, etc..., has the right Coursera certifications, has blogged on the topic, etc...), but are lacking in depth of experience. DS interviews nowadays all seem to be: Can you tell me what a p-value is? What is elastic net regression? Show me how to fit a model in sklearn? How do you impute NAs in an R dataframe? Any smart person can look those up on Stackoverflow or Cross-Validated,.....Instead teams should be asking stuff like: why does portfolio optimization use QP not LP? How does a forecast influence a customer service level? When should a recommendation engine be content based and when should it use collaborative filtering? etc... (This is a true story, happening to the company I currently work for. Names, domains, algorithms, and roles have been shuffled around to protect my anonymity)  Company A has been around for several decades. It is not the biggest name in its domain, but it is a well respected one. Risk analysis and portfolio optimization have been a core of Company A's business since the 90s. They have a large team of 30 or so analysts who perform those tasks on a daily basis. These analysts use ERP solutions implemented for them by one the big ERP companies (SAP, Teradata, Oracle, JD Edwards,...) or one of the major tech consulting companies (Deloitte, Accenture, PWC, Capgemini, etc...) in collaboration with their own in house engineering team. The tools used are embarrassingly old school: Classic RDBMS running on on-prem servers or maybe even on mainframes, code written in COBOL, Fortran, weird proprietary stuff like ABAP or SPSS.....you get the picture. But the models and analytic functions were pretty sophisticated, and surprisingly cutting edge compared to the published academic literature. Most of all, they fit well with the company's enterprise ecosystem, and were honed based on years of deep domain knowledge.  They have a tech team of several engineers (poached from the aforementioned software and consulting companies) and product managers (who came from the experienced pools of analysts and managers who use the software, or poached from business rivals) maintaining and running this software. Their technology might be old school, but collectively, they know the domain and the company's overall architecture very, very well. They've guided the company through several large scale upgrades and migrations and they have a track record of delivering on time, without too much overhead. The few times they've stumbled, they knew how to pick themselves up very quickly. In fact within their industry niche, they have a reputation for their expertise, and have very good relations with the various vendors they've had to deal with. They were the launching pad of several successful ERP consulting careers.  Interestingly, despite dealing on a daily basis with statistical modeling and optimization algorithms, none of the analysts, engineers, or product managers involved describe themselves as data scientists or machine learning experts. It is mostly a cultural thing: Their expertise predates the Data Science/ML hype that started circa 2010, and they got most of their chops using proprietary enterprise tools instead of the open source tools popular nowadays. A few of them have formal statistical training, but most of them came from engineering or domain backgrounds and learned stats on the fly while doing their job. Call this team "Team X".  Sometime around the mid 2010s, Company A started having some serious anxiety issues: Although still doing very well for a company its size, overall economic and demographic trends were shrinking its customer base, and a couple of so called disruptors came up with a new app and business model that started seriously eating into their revenue. A suitable reaction to appease shareholders and Wall Street was necessary. The company already had a decent website and a pretty snazzy app, what more could be done? Leadership decided that it was high time that AI and ML become a core part of the company's business. An ambitious Manager, with no science or engineering background, but who had very briefly toyed with a recommender system a couple of years back, was chosen to build a data science team, call it team "Y" (he had a bachelor's in history from the local state college and worked for several years in the company's marketing org). Team "Y" consists mostly of internal hires who decided they wanted to be data scientists and completed a Coursera certification or a Galvanize boot camp, before being brought on to the team, along with a few of fresh Ph.D or M.Sc holders who didn't like academia and wanted to try their hand at an industry role. All of them were very bright people, they could write great Medium blog posts and give inspiring TED talks, but collectively they had very little real world industry experience. As is the fashion nowadays, this group was made part of a data science org that reported directly to the CEO and Board, bypassing the CIO and any tech or business VPs, since Company A wanted to claim the monikers "data driven" and "AI powered" in their upcoming shareholder meetings. In 3 or 4 years of existence, team Y produced a few Python and R scripts. Their architectural experience  consisted almost entirely in connecting Flask to S3 buckets or Redshift tables, with a couple of the more resourceful ones learning how to plug their models into Tableau or how to spin up a Kuberneties pod.  But they needn't worry: The aforementioned manager, who was now a director (and was also doing an online Masters to make up for his qualifications gap and bolster his chances of becoming VP soon - at least he now understands what L1 regularization is), was a master at playing corporate politics and self-promotion. No matter how few actionable insights team Y produced or how little code they deployed to production, he always had their back and made sure they had ample funding. In fact he now had grandiose plans for setting up an all-purpose machine learning platform that can be used to solve all of the company's data problems.  A couple of sharp minded members of team Y, upon googling their industry name along with the word "data science", realized that risk analysis was a prime candidate for being solved with Bayesian models, and there was already a nifty R package for doing just that, whose tutorial they went through on R-Bloggers.com. One of them had even submitted a Bayesian classifier Kernel for a competition on Kaggle (he was 203rd on the leaderboard), and was eager to put his new-found expertise to use on a real world problem. They pitched the idea to their director, who saw a perfect use case for his upcoming ML platform. They started work on it immediately, without bothering to check whether anybody at Company A was already doing risk analysis. Since their org was independent, they didn't really need to check with anybody else before they got funding for their initiative. Although it was basically a Naive Bayes classifier, the term ML was added to the project tile, to impress the board.  As they progressed with their work however, tensions started to build. They had asked the data warehousing and CA analytics teams to build pipelines for them, and word eventually got out to team X about their project. Team X was initially thrilled: They offered to collaborate whole heartedly, and would have loved to add an ML based feather to their already impressive cap. The product owners and analysts were totally onboard as well: They saw a chance to get in on the whole Data Science hype that they kept hearing about. But through some weird mix of arrogance and insecurity, team Y refused to collaborate with them or share any of their long term goals with them, even as they went to other parts of the company giving brown bag presentations and tutorials on the new model they created.  Team X got resentful: from what they saw of team Y's model, their approach was hopelessly naive and had little chances of scaling or being sustainable in production, and they knew exactly how to help with that. Deploying the model to production would have taken them a few days, given how comfortable they were with DevOps and continuous delivery (team Y had taken several months to figure out how to deploy a simple R script to production). And despite how old school their own tech was, team X were crafty enough to be able to plug it in to their existing architecture. Moreover, the output of the model was such that it didn't take into account how the business will consume it or how it was going to be fed to downstream systems, and the product owners could have gone a long way in making the model more amenable to adoption by the business stakeholders. But team Y wouldn't listen, and their leads brushed off any attempts at communication, let alone collaboration. The vibe that team Y was giving off was "We are the cutting edge ML team, you guys are the legacy server grunts. We don't need your opinion.", and they seemed to have a complete disregard for domain knowledge, or worse, they thought that all that domain knowledge consisted of was being able to grasp the definitions of a few business metrics.  Team X got frustrated and tried to express their concerns to leadership. But despite owning a vital link in Company A's business process, they were only \~50 people in a large 1000 strong technology and operations org, and they were several layers removed from the C-suite, so it was impossible for them to get their voices heard.  Meanwhile, the unstoppable director was doing what he did best: Playing corporate politics. Despite how little his team had actually delivered, he had convinced the board that all analysis and optimization tasks should now be migrated to his yet to be delivered ML platform. Since most leaders now knew that there was overlap between team Y and team X's objectives, his pitch was no longer that team Y was going to create a new insight, but that they were going to replace (or modernize) the legacy statistics based on-prem tools with more accurate cloud based ML tools. Never mind that there was no support in the academic literature for the idea that Naive Bayes works better than the Econometric approaches used by team X, let alone the additional wacky idea that Bayesian Optimization would definitely outperform the QP solvers that were running in production.  Unbeknownst to team X, the original Bayesian risk analysis project has now grown into a multimillion dollar major overhaul initiative, which included the eventual replacement of all of the tools and functions supported by team X along with the necessary migration to the cloud. The CIO and a couple of business VPs are on now board, and tech leadership is treating it as a done deal. An outside vendor, a startup who nobody had heard of, was contracted to help build the platform, since team Y has no engineering skills. The choice was deliberate, as calling on any of the established consulting or software companies would have eventually led leadership to the conclusion that team X was better suited for a transformation on this scale than team Y.  Team Y has no experience with any major ERP deployments, and no domain knowledge, yet they are being tasked with fundamentally changing the business process that is at the core of Company A's business. Their models actually perform worse than those deployed by team X, and their architecture is hopelessly simplistic, compared to what is necessary for running such a solution in production.  Ironically, using Bayesian thinking and based on all the evidence, the likelihood that team Y succeeds is close to 0%. At best, the project is going to end up being a write off of 50 million dollars or more. Once the !@#$!@hits the fan, a couple of executive heads are going to role, and dozens of people will get laid off. At worst, given how vital risk analysis and portfolio optimization is to Company A's revenue stream, the failure will eventually sink the whole company. It probably won't go bankrupt, but it will lose a significant portion of its business and work force. Failed ERP implementations can and do sink large companies: Just see what happened to National Grid US, SuperValu or Target Canada.  One might argue that this is more about corporate disfunction and bad leadership than about data science and AI. But I disagree. I think the core driver of this debacle is indeed the blind faith in Data Scientists, ML models and the promise of AI, and the overall culture of hype and self promotion that is very common among the ML crowd.  We haven't seen the end of this story: I sincerely hope that this ends well for the sake of my colleagues and all involved. Company A is a good company, and both its customers and its employees deserver better. But the chances of that happening are negligible given all the information available, and this failure will hit my company hard.

[News] AAAI 2025 Workshop on AI for Music 🎶
reddit
LLM Vibe Score0
Human Vibe Score0
Saysike_rightnow69This week

[News] AAAI 2025 Workshop on AI for Music 🎶

Hi everyone! We’re hosting the first “AI for Music” workshop at AAAI on March 3, 2025. The workshop will explore how AI is transforming music creation, recognition, education, and more. Topics include AI-driven composition, sound design, legal and ethical challenges, and AI’s impact on musicians’ careers. Submissions (up to 6 pages) are welcome until November 22, 2024. Work in progress is encouraged! Workshop Summary This one-day workshop will explore the dynamic intersection of artificial intelligence and music. It explores how AI is transforming music creation, recognition, and education, ethical and legal implications, as well as business opportunities. We will investigate how AI is changing the music industry and education—from composition to performance, production, collaboration, and audience experience. Participants will gain insights into the technological challenges in music and how AI can enhance creativity, enabling musicians and producers to push the boundaries of their art. The workshop will cover topics such as AI-driven music composition, where algorithms generate melodies, harmonies, and even full orchestral arrangements. We will discuss how AI tools assist in sound design, remixing, and mastering, allowing for new sonic possibilities and efficiencies in music production. Additionally, we'll examine AI's impact on music education and the careers of musicians, exploring advanced learning tools and teaching methods. AI technologies are increasingly adopted in the music and entertainment industry. The workshop will also discuss the legal and ethical implications of AI in music, including questions of authorship, originality, and the evolving role of human artists in an increasingly automated world. This workshop is designed for AI researchers, musicians, producers, and educators interested in the current status and future of AI in music. Call for Papers Submissions should be a maximum of 6 pages. Work in progress is welcome. Authors are encouraged to include descriptions of their prototype implementations. Additionally, authors are encouraged to interact with workshop attendees by including posters or demonstrations at the end of the workshop. Conceptual designs without any evidence of practical implementation are discouraged. Topics of interest are (but not limited to) AI-Driven Music Composition and Generation AI in Music Practice and Performance AI-based Music Recognition and Transcription AI Applications in Sound Design AI-Generated Videos and Lyrics Based on Music Legal and Ethical Implications of AI in Music AI’s Impact on Musicians’ Careers and Education Business Opportunities of AI in Music Music Datasets and Data Analysis Important Dates Submission Deadline: November 22, 2024 Notification: December 9, 2024 Final Version Due: December 31, 2024 We hope to see you there! 🎶

[D] What is your honest experience with reinforcement learning?
reddit
LLM Vibe Score0
Human Vibe Score1
Starks-TechnologyThis week

[D] What is your honest experience with reinforcement learning?

In my personal experience, SOTA RL algorithms simply don't work. I've tried working with reinforcement learning for over 5 years. I remember when Alpha Go defeated the world famous Go player, Lee Sedol, and everybody thought RL would take the ML community by storm. Yet, outside of toy problems, I've personally never found a practical use-case of RL. What is your experience with it? Aside from Ad recommendation systems and RLHF, are there legitimate use-cases of RL? Or, was it all hype? Edit: I know a lot about AI. I built NexusTrade, an AI-Powered automated investing tool that lets non-technical users create, update, and deploy their trading strategies. I’m not an idiot nor a noob; RL is just ridiculously hard. Edit 2: Since my comments are being downvoted, here is a link to my article that better describes my position. It's not that I don't understand RL. I released my open-source code and wrote a paper on it. It's the fact that it's EXTREMELY difficult to understand. Other deep learning algorithms like CNNs (including ResNets), RNNs (including GRUs and LSTMs), Transformers, and GANs are not hard to understand. These algorithms work and have practical use-cases outside of the lab. Traditional SOTA RL algorithms like PPO, DDPG, and TD3 are just very hard. You need to do a bunch of research to even implement a toy problem. In contrast, the decision transformer is something anybody can implement, and it seems to match or surpass the SOTA. You don't need two networks battling each other. You don't have to go through hell to debug your network. It just naturally learns the best set of actions in an auto-regressive manner. I also didn't mean to come off as arrogant or imply that RL is not worth learning. I just haven't seen any real-world, practical use-cases of it. I simply wanted to start a discussion, not claim that I know everything. Edit 3: There's a shockingly number of people calling me an idiot for not fully understanding RL. You guys are wayyy too comfortable calling people you disagree with names. News-flash, not everybody has a PhD in ML. My undergraduate degree is in biology. I self-taught myself the high-level maths to understand ML. I'm very passionate about the field; I just have VERY disappointing experiences with RL. Funny enough, there are very few people refuting my actual points. To summarize: Lack of real-world applications Extremely complex and inaccessible to 99% of the population Much harder than traditional DL algorithms like CNNs, RNNs, and GANs Sample inefficiency and instability Difficult to debug Better alternatives, such as the Decision Transformer Are these not legitimate criticisms? Is the purpose of this sub not to have discussions related to Machine Learning? To the few commenters that aren't calling me an idiot...thank you! Remember, it costs you nothing to be nice! Edit 4: Lots of people seem to agree that RL is over-hyped. Unfortunately those comments are downvoted. To clear up some things: We've invested HEAVILY into reinforcement learning. All we got from this investment is a robot that can be super-human at (some) video games. AlphaFold did not use any reinforcement learning. SpaceX doesn't either. I concede that it can be useful for robotics, but still argue that it's use-cases outside the lab are extremely limited. If you're stumbling on this thread and curious about an RL alternative, check out the Decision Transformer. It can be used in any situation that a traditional RL algorithm can be used. Final Edit: To those who contributed more recently, thank you for the thoughtful discussion! From what I learned, model-based models like Dreamer and IRIS MIGHT have a future. But everybody who has actually used model-free models like DDPG unanimously agree that they suck and don’t work.

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company
reddit
LLM Vibe Score0
Human Vibe Score0.778
wutangsamThis week

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company

I’ve learned so much over the years from this subreddit. I thought I’d return the favour and share some of my own learnings. In November 2020 my best friend and I had an idea. “What if we could find out which stocks the Internet is talking about?” This formed the origins of Ticker Nerd. 9 months later we sold Ticker Nerd to Finder (an Australian fintech company valued at around $500m). In this post, I am going to lay out how we got there. How we came up with the idea First off, like other posts have covered - you don’t NEED a revolutionary or original idea to build a business. There are tonnes of “boring” businesses making over 7 figures a year e.g. law firms, marketing agencies, real estate companies etc. If you’re looking for an exact formula to come up with a great business idea I’m sorry, but it doesn’t exist. Finding new business opportunities is more of an art than a science. Although, there are ways you can make it easier to find inspiration. Below are the same resources I use for inspiration. I rarely ever come up with ideas without first searching one of the resources below for inspiration: Starter Story Twitter Startup Ideas My First Million Trends by the Hustle Trends VC To show how you how messy, random and unpredictable it can be to find an idea - let me explain how my co-founder and I came up with the idea for Ticker Nerd: We discovered a new product on Twitter called Exploding Topics. It was a newsletter that uses a bunch of software and algorithms to find trends that are growing quickly before they hit the mainstream. I had recently listened to a podcast episode from My First Million where they spoke about Motley Fool making hundreds of millions from their investment newsletters. We asked ourselves what if we could build a SaaS platform similar to Exploding Topics but it focused on stocks? We built a quick landing page using Carrd + Gumroad that explained what our new idea will do and included a payment option to get early access for $49. We called it Exploding Stock (lol). We shared it around a bunch of Facebook groups and subreddits. We made $1,000 in pre-sales within a couple days. My co-founder and I can’t code so we had to find a developer to build our idea. We interviewed a bunch of potential candidates. Meanwhile, I was trawling through Wall Street Bets and found a bunch of free tools that did roughly what we wanted to build. Instead of building another SaaS tool that did the same thing as these free tools we decided to pivot from our original idea. Our new idea = a paid newsletter that sends a weekly report that summarises 2 of the best stocks that are growing in interest on the Internet. We emailed everyone who pre-ordered access, telling them about the change and offered a full refund if they wanted. tl;dr: We essentially combined two existing businesses (Exploding Topics and Motley Fool) and made it way better. We validated the idea by finding out if people will actually pay money for it BEFORE we decided to build it. The idea we started out with changed over time. How to work out if your idea will actually make money It’s easy to get hung up on designing the logo or choosing the perfect domain name for your new idea. At this stage none of that matters. The most important thing is working out if people will pay money for it. This is where validation comes in. We usually validate ideas using Carrd. It lets you build a simple one page site without having to code. The Ticker Nerd site was actually built using a Carrd template. Here’s how you can do it yourself (at a high level): Create a Carrd pro account (yes it's a $49 one off payment but you’ll get way more value out of it). Buy a cheap template and send it to your Carrd account. You can build your own template but this will save you a lot of time. Once the template reaches your Carrd account, duplicate it. Leave the original so it can be duplicated for other ideas. Jump onto Canva (free) and create a logo using the free logos provided. Import your logo. Add copy to the page that explains your idea. Use the AIDA formula. Sign up to Gumroad (free) and create a pre-sale campaign. Create a discounted lifetime subscription or version of the product. This will be used pre-sales. Add the copy from the site into the pre-sale campaign on Gumroad. Add a ‘widget’ to Carrd and connect it to Gumroad using the existing easy integration feature. Purchase a domain name. Connect it to Carrd. Test the site works. Share your website Now the site is ready you can start promoting it in various places to see how the market reacts. An easy method is to find relevant subreddits using Anvaka (Github tool) or Subreddit Stats. The Anvaka tool provides a spider map of all the connected subreddits that users are active in. The highlighted ones are most relevant. You can post a thread in these subreddits that offer value or can generate discussion. For example: ‘I’m creating a tool that can write all your copy, would anyone actually use this?’ ‘What does everything think of using AI to get our copy written faster?’ ‘It’s time to scratch my own itch, I’m creating a tool that writes marketing copy using GPT-3. What are the biggest problems you face writing marketing copy? I’ll build a solution for it’ Reddit is pretty brutal these days so make sure the post is genuine and only drop your link in the comments or in the post if it seems natural. If people are interested they’ll ask for the link. Another great place to post is r/entrepreuerridealong and r/business_ideas. These subreddits expect people to share their ideas and you’ll likely make some sales straight off the bat. I also suggest posting in some Facebook groups (related to your idea) as well just for good measure. Assess the results If people are paying you for early access you can assume that it’s worth building your idea. The beauty of posting your idea on Reddit or in Facebook groups is you’ll quickly learn why people love/hate your idea. This can help you decide how to tweak the idea or if you should drop it and move on to the next one. How we got our first 100 customers (for free) By validating Ticker Nerd using subreddits and Facebook groups this gave us our first paying customers. But we knew this wouldn’t be sustainable. We sat down and brainstormed every organic strategy we could use to get traction as quickly as possible. The winner: a Product Hunt launch. A successful Product Hunt launch isn’t easy. You need: Someone that has a solid reputation and audience to “hunt” your product (essentially an endorsement). An aged Product Hunt account - you can’t post any products if your account is less than a week old. To be following relevant Product Hunt members - since they get notified when you launch a new product if they’re following you. Relationships with other builders and makers on Product Hunt that also have a solid reputation and following. Although, if you can pull it off you can get your idea in front of tens of thousands of people actively looking for new products. Over the next few weeks, I worked with my co-founder on connecting with different founders, indie hackers and entrepreneurs mainly via Twitter. We explained to them our plans for the Product Hunt launch and managed to get a small army of people ready to upvote our product on launch day. We were both nervous on the day of the launch. We told ourselves to have zero expectations. The worst that could happen was no one signed up and we were in the same position as we’re in now. Luckily, within a couple of hours Ticker Nerd was on the homepage of Product Hunt and in the top 10. The results were instant. After 24 hours we had around 200 people enter their payment details to sign up for our free trial. These signups were equal to around $5,800 in monthly recurring revenue. \-- I hope this post was useful! Drop any questions you have below and I’ll do my best to respond :)

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security

Uzair Javaid, a Ph.D. with a passion for data privacy, co-founded Betterdata to tackle one of AI's most pressing challenges: protecting privacy while enabling innovation. Recently, Betterdata secured a lucrative contract with the US Department of Homeland Security, 1 of only 4 companies worldwide to do so and the only one in Asia. Here's how he did it: The Story So what's your story? I grew up in Peshawar, Pakistan, excelling in coding despite studying electrical engineering. Inspired by my professors, I set my sights on studying abroad and eventually earned a Ph.D. scholarship at NUS Singapore, specializing in data security and privacy. During my research, I ethically hacked Ethereum and published 15 papers—three times the requirement. While wrapping up my Ph.D., I explored startup ideas and joined Entrepreneur First, where I met Kevin Yee. With his expertise in generative models and mine in privacy, we founded Betterdata. Now, nearly three years in, we’ve secured a major contract with the U.S. Department of Homeland Security—one of only four companies globally and the only one from Asia. The Startup In a nutshell, what does your startup do? Betterdata is a startup that uses AI and synthetic data generation to address two major challenges: data privacy and the scarcity of high-quality data for training AI models. By leveraging generative models and privacy-enhancing technologies, Betterdata enables businesses, such as banks, to use customer data without breaching privacy regulations. The platform trains AI on real data, learns its patterns, and generates synthetic data that mimics the real thing without containing any personal or sensitive information. This allows companies to innovate and develop AI solutions safely and ethically, all while tackling the growing need for diverse, high-quality data in AI development. How did you conduct ideation and validation for your startup? The initial idea for Betterdata came from personal experience. During my Ph.D., I ethically hacked Ethereum’s blockchain, exposing flaws in encryption-based data sharing. This led me to explore AI-driven deep synthesis technology—similar to deepfakes but for structured data privacy. With GDPR impacting 28M+ businesses, I saw a massive opportunity to help enterprises securely share data while staying compliant. To validate the idea, I spoke to 50 potential customers—a number that strikes the right balance. Some say 100, but that’s impractical for early-stage founders. At 50, patterns emerge: if 3 out of 10 mention the same problem, and this repeats across 50, you have 10–15 strong signals, making it a solid foundation for an MVP. Instead of outbound sales, which I dislike, we used three key methods: Account-Based Marketing (ABM)—targeting technically savvy users with solutions for niche problems, like scaling synthetic data for banks. Targeted Content Marketing—regular customer conversations shaped our thought leadership and outreach. Raising Awareness Through Partnerships—collaborating with NUS, Singapore’s PDPC, and Plug and Play to build credibility and educate the market. These strategies attracted serious customers willing to pay, guiding Betterdata’s product development and market fit. How did you approach the initial building and ongoing product development? In the early stages, we built synthetic data generation algorithms and a basic UI for proof-of-concept, using open-source datasets to engage with banks. We quickly learned that banks wouldn't share actual customer data due to privacy concerns, so we had to conduct on-site installations and gather feedback to refine our MVP. Through continuous consultation with customers, we discovered real enterprise data posed challenges, such as missing values, which led us to adapt our prototype accordingly. This iterative approach of listening to customer feedback and observing their usage allowed us to improve our product, enhance UX, and address unmet needs while building trust and loyalty. Working closely with our customers also gives us a data advantage. Our solution’s effectiveness depends on customer data, which we can't fully access, but bridging this knowledge gap gives us a competitive edge. The more customers we test on, the more our algorithms adapt to diverse use cases, making it harder for competitors to replicate our insights. My approach to iteration is simple: focus solely on customer feedback and ignore external noise like trends or advice. The key question for the team is: which customer is asking for this feature or solution? As long as there's a clear answer, we move forward. External influences, such as AI hype, often bring more confusion than clarity. True long-term success comes from solving real customer problems, not chasing trends. Customers may not always know exactly what they want, but they understand their problems. Our job is to identify these problems and solve them in innovative ways. While customers may suggest specific features, we stay focused on solving the core issue rather than just fulfilling their exact requests. The idea aligns with the quote often attributed to Henry Ford: "If I asked people what they wanted, they would have said faster horses." The key is understanding their problems, not just taking requests at face value. How do you assess product-market fit? To assess product-market fit, we track two key metrics: Customers' Willingness to Pay: We measure both the quantity and quality of meetings with potential customers. A high number of meetings with key decision-makers signals genuine interest. At Betterdata, we focused on getting meetings with people in banks and large enterprises to gauge our product's resonance with the target market. How Much Customers Are Willing to Pay: We monitor the price customers are willing to pay, especially in the early stages. For us, large enterprises, like banks, were willing to pay a premium for our synthetic data platform due to the growing need for privacy tech. This feedback guided our product refinement and scaling strategy. By focusing on these metrics, we refined our product and positioned it for scaling. What is your business model? We employ a structured, phase-driven approach for out business model, as a B2B startup. I initially struggled with focusing on the core value proposition in sales, often becoming overly educational. Eventually, we developed a product roadmap with models that allowed us to match customer needs to specific offerings and justify our pricing. Our pricing structure includes project-based pilots and annual contracts for successful deployments. At Betterdata, our customer engagement unfolds across three phases: Phase 1: Trial and Benchmarking \- We start with outreach and use open-source datasets to showcase results, offering customers a trial period to evaluate the solution. Phase 2: Pilot or PoC \- After positive trial results, we conduct a PoC or pilot using the customer’s private data, with the understanding that successful pilots lead to an annual contract. Phase 3: Multi-Year Contracts \- Following a successful pilot, we transition to long-term commercial contracts, focusing on multi-year agreements to ensure stability and ongoing partnerships. How do you do marketing for your brand? We take a non-conventional approach to marketing, focusing on answering one key question: Which customers are willing to pay, and how much? This drives our messaging to show how our solution meets their needs. Our strategy centers around two main components: Building a network of lead magnets \- These are influential figures like senior advisors, thought leaders, and strategic partners. Engaging with institutions like IMDA, SUTD, and investors like Plug and Play helps us gain access to the right people and foster warm introductions, which shorten our sales cycle and ensure we’re reaching the right audience. Thought leadership \- We build our brand through customer traction, technology evidence, and regulatory guidelines. This helps us establish credibility in the market and position ourselves as trusted leaders in our field. This holistic approach has enabled us to navigate diverse market conditions in Asia and grow our B2B relationships. By focusing on these areas, we drive business growth and establish strong trust with stakeholders. What's your advice for fundraising? Here are my key takeaways for other founders when it comes to fundraising: Fundraise When You Don’t Need To We closed our seed round in April 2023, a time when we weren't actively raising. Founders should always be in fundraising mode, even when they're not immediately in need of capital. Don’t wait until you have only a few months of runway left. Keep the pipeline open and build relationships. When the timing is right, execution becomes much easier. For us, our investment came through a combination of referrals and inbound interest. Even our lead investor initially rejected us, but after re-engaging, things eventually fell into place. It’s crucial to stay humble, treat everyone with respect, and maintain those relationships for when the time is right. Be Mindful of How You Present Information When fundraising, how you present information matters a lot. We created a comprehensive, easily digestible investment memo, hosted on Notion, which included everything an investor might need—problem, solution, market, team, risks, opportunities, and data. The goal was for investors to be able to get the full picture within 30 minutes without chasing down extra details. We also focused on making our financial model clear and meaningful, even though a 5-year forecast might be overkill at the seed stage. The key was clarity and conciseness, and making it as easy as possible for investors to understand the opportunity. I learned that brevity and simplicity are often the best ways to make a memorable impact. For the pitch itself, keep it simple and focus on 4 things: problem, solution, team, and market. If you can summarize each of these clearly and concisely, you’ll have a compelling pitch. Later on, you can expand into market segments, traction, and other metrics, but for seed-stage, focus on those four areas, and make sure you’re strong in at least three of them. If you do, you'll have a compelling case. How do you run things day-to-day? i.e what's your operational workflow and team structure? Here's an overview of our team structure and process: Internally: Our team is divided into two main areas: backend (internal team) and frontend (market-facing team). There's no formal hierarchy within the backend team. We all operate as equals, defining our goals based on what needs to be developed, assigning tasks, and meeting weekly to share updates and review progress. The focus is on full ownership of tasks and accountability for getting things done. I also contribute to product development, identifying challenges and clearing obstacles to help the team move forward. Backend Team: We approach tasks based on the scope defined by customers, with no blame or hierarchy. It's like a sports team—sometimes someone excels, and other times they struggle, but we support each other and move forward together. Everyone has the creative freedom to work in the way that suits them best, but we establish regular meetings and check-ins to ensure alignment and progress. Frontend Team: For the market-facing side, we implement a hierarchy because the market expects this structure. If I present myself as "CEO," it signals authority and credibility. This distinction affects how we communicate with the market and how we build our brand. The frontend team is split into four main areas: Business Product (Software Engineering) Machine Learning Engineering R&D The C-suite sits at the top, followed by team leads, and then the executors. We distill market expectations into actionable tasks, ensuring that everyone is clear on their role and responsibilities. Process: We start by receiving market expectations and defining tasks based on them. Tasks are assigned to relevant teams, and execution happens with no communication barriers between team members. This ensures seamless collaboration and focused execution. The main goal is always effectiveness—getting things done efficiently while maintaining flexibility in how individuals approach their work. In both teams, there's an emphasis on accountability, collaboration, and clear communication, but the structure varies according to the nature of the work and external expectations.

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company
reddit
LLM Vibe Score0
Human Vibe Score0.778
wutangsamThis week

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company

I’ve learned so much over the years from this subreddit. I thought I’d return the favour and share some of my own learnings. In November 2020 my best friend and I had an idea. “What if we could find out which stocks the Internet is talking about?” This formed the origins of Ticker Nerd. 9 months later we sold Ticker Nerd to Finder (an Australian fintech company valued at around $500m). In this post, I am going to lay out how we got there. How we came up with the idea First off, like other posts have covered - you don’t NEED a revolutionary or original idea to build a business. There are tonnes of “boring” businesses making over 7 figures a year e.g. law firms, marketing agencies, real estate companies etc. If you’re looking for an exact formula to come up with a great business idea I’m sorry, but it doesn’t exist. Finding new business opportunities is more of an art than a science. Although, there are ways you can make it easier to find inspiration. Below are the same resources I use for inspiration. I rarely ever come up with ideas without first searching one of the resources below for inspiration: Starter Story Twitter Startup Ideas My First Million Trends by the Hustle Trends VC To show how you how messy, random and unpredictable it can be to find an idea - let me explain how my co-founder and I came up with the idea for Ticker Nerd: We discovered a new product on Twitter called Exploding Topics. It was a newsletter that uses a bunch of software and algorithms to find trends that are growing quickly before they hit the mainstream. I had recently listened to a podcast episode from My First Million where they spoke about Motley Fool making hundreds of millions from their investment newsletters. We asked ourselves what if we could build a SaaS platform similar to Exploding Topics but it focused on stocks? We built a quick landing page using Carrd + Gumroad that explained what our new idea will do and included a payment option to get early access for $49. We called it Exploding Stock (lol). We shared it around a bunch of Facebook groups and subreddits. We made $1,000 in pre-sales within a couple days. My co-founder and I can’t code so we had to find a developer to build our idea. We interviewed a bunch of potential candidates. Meanwhile, I was trawling through Wall Street Bets and found a bunch of free tools that did roughly what we wanted to build. Instead of building another SaaS tool that did the same thing as these free tools we decided to pivot from our original idea. Our new idea = a paid newsletter that sends a weekly report that summarises 2 of the best stocks that are growing in interest on the Internet. We emailed everyone who pre-ordered access, telling them about the change and offered a full refund if they wanted. tl;dr: We essentially combined two existing businesses (Exploding Topics and Motley Fool) and made it way better. We validated the idea by finding out if people will actually pay money for it BEFORE we decided to build it. The idea we started out with changed over time. How to work out if your idea will actually make money It’s easy to get hung up on designing the logo or choosing the perfect domain name for your new idea. At this stage none of that matters. The most important thing is working out if people will pay money for it. This is where validation comes in. We usually validate ideas using Carrd. It lets you build a simple one page site without having to code. The Ticker Nerd site was actually built using a Carrd template. Here’s how you can do it yourself (at a high level): Create a Carrd pro account (yes it's a $49 one off payment but you’ll get way more value out of it). Buy a cheap template and send it to your Carrd account. You can build your own template but this will save you a lot of time. Once the template reaches your Carrd account, duplicate it. Leave the original so it can be duplicated for other ideas. Jump onto Canva (free) and create a logo using the free logos provided. Import your logo. Add copy to the page that explains your idea. Use the AIDA formula. Sign up to Gumroad (free) and create a pre-sale campaign. Create a discounted lifetime subscription or version of the product. This will be used pre-sales. Add the copy from the site into the pre-sale campaign on Gumroad. Add a ‘widget’ to Carrd and connect it to Gumroad using the existing easy integration feature. Purchase a domain name. Connect it to Carrd. Test the site works. Share your website Now the site is ready you can start promoting it in various places to see how the market reacts. An easy method is to find relevant subreddits using Anvaka (Github tool) or Subreddit Stats. The Anvaka tool provides a spider map of all the connected subreddits that users are active in. The highlighted ones are most relevant. You can post a thread in these subreddits that offer value or can generate discussion. For example: ‘I’m creating a tool that can write all your copy, would anyone actually use this?’ ‘What does everything think of using AI to get our copy written faster?’ ‘It’s time to scratch my own itch, I’m creating a tool that writes marketing copy using GPT-3. What are the biggest problems you face writing marketing copy? I’ll build a solution for it’ Reddit is pretty brutal these days so make sure the post is genuine and only drop your link in the comments or in the post if it seems natural. If people are interested they’ll ask for the link. Another great place to post is r/entrepreuerridealong and r/business_ideas. These subreddits expect people to share their ideas and you’ll likely make some sales straight off the bat. I also suggest posting in some Facebook groups (related to your idea) as well just for good measure. Assess the results If people are paying you for early access you can assume that it’s worth building your idea. The beauty of posting your idea on Reddit or in Facebook groups is you’ll quickly learn why people love/hate your idea. This can help you decide how to tweak the idea or if you should drop it and move on to the next one. How we got our first 100 customers (for free) By validating Ticker Nerd using subreddits and Facebook groups this gave us our first paying customers. But we knew this wouldn’t be sustainable. We sat down and brainstormed every organic strategy we could use to get traction as quickly as possible. The winner: a Product Hunt launch. A successful Product Hunt launch isn’t easy. You need: Someone that has a solid reputation and audience to “hunt” your product (essentially an endorsement). An aged Product Hunt account - you can’t post any products if your account is less than a week old. To be following relevant Product Hunt members - since they get notified when you launch a new product if they’re following you. Relationships with other builders and makers on Product Hunt that also have a solid reputation and following. Although, if you can pull it off you can get your idea in front of tens of thousands of people actively looking for new products. Over the next few weeks, I worked with my co-founder on connecting with different founders, indie hackers and entrepreneurs mainly via Twitter. We explained to them our plans for the Product Hunt launch and managed to get a small army of people ready to upvote our product on launch day. We were both nervous on the day of the launch. We told ourselves to have zero expectations. The worst that could happen was no one signed up and we were in the same position as we’re in now. Luckily, within a couple of hours Ticker Nerd was on the homepage of Product Hunt and in the top 10. The results were instant. After 24 hours we had around 200 people enter their payment details to sign up for our free trial. These signups were equal to around $5,800 in monthly recurring revenue. \-- I hope this post was useful! Drop any questions you have below and I’ll do my best to respond :)

Why Ignoring AI Agents in 2025 Will Kill Your Marketing Strategy
reddit
LLM Vibe Score0
Human Vibe Score1
frankiemuiruriThis week

Why Ignoring AI Agents in 2025 Will Kill Your Marketing Strategy

If you're still focusing solely on grabbing the attention of human beings with your marketing efforts, you're already behind. In 2025, the game will change. Good marketing will demand an in-depth understanding of the AI space, especially the AI Agent space. Why? Your ads and content won’t just be seen by humans anymore. They’ll be analyzed, indexed, and often acted upon by AI agents—automated systems that will be working on behalf of companies and consumers alike. Your New Audience: Humans + AI Agents It’s not just about appealing to people. Companies are employing AI robots to research, negotiate, and make purchasing decisions. These AI agents are fast, thorough, and unrelenting. Unlike humans, they can analyze millions of options in seconds. And if your marketing isn’t optimized for them, you’ll get filtered out before you even reach the human decision-maker. How to Prepare Your Marketing for AI Agents The companies that dominate marketing in 2025 will be the ones that master the art of capturing AI attention. To do this, marketers will need to: Understand the AI agents shaping their industry. Research how AI agents function in your niche. What are they prioritizing? How do they rank options? Create AI-friendly content. Design ads and messaging that are easily understandable and accessible to AI agents. This means clear metadata, structured data, and AI-readable formats. Invest in AI analytics. AI agents leave behind footprints. Tracking and analyzing their behavior is critical. Stay ahead of AI trends. The AI agent space is evolving rapidly. What works today might be obsolete tomorrow. How My Agency Adapted and Thrived in the AI Space At my digital agency, we saw this shift coming and decided to act early. In 2023, we started integrating AI optimization into our marketing strategies. One of our clients—a B2B SaaS company—struggled to get traction because their competitors were drowning them out in Google search rankings and ad platforms. By analyzing the algorithms and behaviors of AI agents in their space, we: Rewrote their website copy with structured data and optimized metadata that was more AI-agent friendly. Created ad campaigns with clear, concise messaging and technical attributes that AI agents could quickly process and index. Implemented predictive analytics to understand what AI agents would prioritize based on past behaviors. The results? Their website traffic doubled in three months, and their lead conversion rate skyrocketed by 40%. Over half of the traffic increase was traced back to AI agents recommending their platform to human users. The Takeaway In 2025, marketing won’t just be about human attention. It’ll be about AI attention—and that requires a completely different mindset. AI agents are not your enemy; they’re your new gatekeepers. Learn to speak their language, and you’ll dominate the marketing game.

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security

Uzair Javaid, a Ph.D. with a passion for data privacy, co-founded Betterdata to tackle one of AI's most pressing challenges: protecting privacy while enabling innovation. Recently, Betterdata secured a lucrative contract with the US Department of Homeland Security, 1 of only 4 companies worldwide to do so and the only one in Asia. Here's how he did it: The Story So what's your story? I grew up in Peshawar, Pakistan, excelling in coding despite studying electrical engineering. Inspired by my professors, I set my sights on studying abroad and eventually earned a Ph.D. scholarship at NUS Singapore, specializing in data security and privacy. During my research, I ethically hacked Ethereum and published 15 papers—three times the requirement. While wrapping up my Ph.D., I explored startup ideas and joined Entrepreneur First, where I met Kevin Yee. With his expertise in generative models and mine in privacy, we founded Betterdata. Now, nearly three years in, we’ve secured a major contract with the U.S. Department of Homeland Security—one of only four companies globally and the only one from Asia. The Startup In a nutshell, what does your startup do? Betterdata is a startup that uses AI and synthetic data generation to address two major challenges: data privacy and the scarcity of high-quality data for training AI models. By leveraging generative models and privacy-enhancing technologies, Betterdata enables businesses, such as banks, to use customer data without breaching privacy regulations. The platform trains AI on real data, learns its patterns, and generates synthetic data that mimics the real thing without containing any personal or sensitive information. This allows companies to innovate and develop AI solutions safely and ethically, all while tackling the growing need for diverse, high-quality data in AI development. How did you conduct ideation and validation for your startup? The initial idea for Betterdata came from personal experience. During my Ph.D., I ethically hacked Ethereum’s blockchain, exposing flaws in encryption-based data sharing. This led me to explore AI-driven deep synthesis technology—similar to deepfakes but for structured data privacy. With GDPR impacting 28M+ businesses, I saw a massive opportunity to help enterprises securely share data while staying compliant. To validate the idea, I spoke to 50 potential customers—a number that strikes the right balance. Some say 100, but that’s impractical for early-stage founders. At 50, patterns emerge: if 3 out of 10 mention the same problem, and this repeats across 50, you have 10–15 strong signals, making it a solid foundation for an MVP. Instead of outbound sales, which I dislike, we used three key methods: Account-Based Marketing (ABM)—targeting technically savvy users with solutions for niche problems, like scaling synthetic data for banks. Targeted Content Marketing—regular customer conversations shaped our thought leadership and outreach. Raising Awareness Through Partnerships—collaborating with NUS, Singapore’s PDPC, and Plug and Play to build credibility and educate the market. These strategies attracted serious customers willing to pay, guiding Betterdata’s product development and market fit. How did you approach the initial building and ongoing product development? In the early stages, we built synthetic data generation algorithms and a basic UI for proof-of-concept, using open-source datasets to engage with banks. We quickly learned that banks wouldn't share actual customer data due to privacy concerns, so we had to conduct on-site installations and gather feedback to refine our MVP. Through continuous consultation with customers, we discovered real enterprise data posed challenges, such as missing values, which led us to adapt our prototype accordingly. This iterative approach of listening to customer feedback and observing their usage allowed us to improve our product, enhance UX, and address unmet needs while building trust and loyalty. Working closely with our customers also gives us a data advantage. Our solution’s effectiveness depends on customer data, which we can't fully access, but bridging this knowledge gap gives us a competitive edge. The more customers we test on, the more our algorithms adapt to diverse use cases, making it harder for competitors to replicate our insights. My approach to iteration is simple: focus solely on customer feedback and ignore external noise like trends or advice. The key question for the team is: which customer is asking for this feature or solution? As long as there's a clear answer, we move forward. External influences, such as AI hype, often bring more confusion than clarity. True long-term success comes from solving real customer problems, not chasing trends. Customers may not always know exactly what they want, but they understand their problems. Our job is to identify these problems and solve them in innovative ways. While customers may suggest specific features, we stay focused on solving the core issue rather than just fulfilling their exact requests. The idea aligns with the quote often attributed to Henry Ford: "If I asked people what they wanted, they would have said faster horses." The key is understanding their problems, not just taking requests at face value. How do you assess product-market fit? To assess product-market fit, we track two key metrics: Customers' Willingness to Pay: We measure both the quantity and quality of meetings with potential customers. A high number of meetings with key decision-makers signals genuine interest. At Betterdata, we focused on getting meetings with people in banks and large enterprises to gauge our product's resonance with the target market. How Much Customers Are Willing to Pay: We monitor the price customers are willing to pay, especially in the early stages. For us, large enterprises, like banks, were willing to pay a premium for our synthetic data platform due to the growing need for privacy tech. This feedback guided our product refinement and scaling strategy. By focusing on these metrics, we refined our product and positioned it for scaling. What is your business model? We employ a structured, phase-driven approach for out business model, as a B2B startup. I initially struggled with focusing on the core value proposition in sales, often becoming overly educational. Eventually, we developed a product roadmap with models that allowed us to match customer needs to specific offerings and justify our pricing. Our pricing structure includes project-based pilots and annual contracts for successful deployments. At Betterdata, our customer engagement unfolds across three phases: Phase 1: Trial and Benchmarking \- We start with outreach and use open-source datasets to showcase results, offering customers a trial period to evaluate the solution. Phase 2: Pilot or PoC \- After positive trial results, we conduct a PoC or pilot using the customer’s private data, with the understanding that successful pilots lead to an annual contract. Phase 3: Multi-Year Contracts \- Following a successful pilot, we transition to long-term commercial contracts, focusing on multi-year agreements to ensure stability and ongoing partnerships. How do you do marketing for your brand? We take a non-conventional approach to marketing, focusing on answering one key question: Which customers are willing to pay, and how much? This drives our messaging to show how our solution meets their needs. Our strategy centers around two main components: Building a network of lead magnets \- These are influential figures like senior advisors, thought leaders, and strategic partners. Engaging with institutions like IMDA, SUTD, and investors like Plug and Play helps us gain access to the right people and foster warm introductions, which shorten our sales cycle and ensure we’re reaching the right audience. Thought leadership \- We build our brand through customer traction, technology evidence, and regulatory guidelines. This helps us establish credibility in the market and position ourselves as trusted leaders in our field. This holistic approach has enabled us to navigate diverse market conditions in Asia and grow our B2B relationships. By focusing on these areas, we drive business growth and establish strong trust with stakeholders. What's your advice for fundraising? Here are my key takeaways for other founders when it comes to fundraising: Fundraise When You Don’t Need To We closed our seed round in April 2023, a time when we weren't actively raising. Founders should always be in fundraising mode, even when they're not immediately in need of capital. Don’t wait until you have only a few months of runway left. Keep the pipeline open and build relationships. When the timing is right, execution becomes much easier. For us, our investment came through a combination of referrals and inbound interest. Even our lead investor initially rejected us, but after re-engaging, things eventually fell into place. It’s crucial to stay humble, treat everyone with respect, and maintain those relationships for when the time is right. Be Mindful of How You Present Information When fundraising, how you present information matters a lot. We created a comprehensive, easily digestible investment memo, hosted on Notion, which included everything an investor might need—problem, solution, market, team, risks, opportunities, and data. The goal was for investors to be able to get the full picture within 30 minutes without chasing down extra details. We also focused on making our financial model clear and meaningful, even though a 5-year forecast might be overkill at the seed stage. The key was clarity and conciseness, and making it as easy as possible for investors to understand the opportunity. I learned that brevity and simplicity are often the best ways to make a memorable impact. For the pitch itself, keep it simple and focus on 4 things: problem, solution, team, and market. If you can summarize each of these clearly and concisely, you’ll have a compelling pitch. Later on, you can expand into market segments, traction, and other metrics, but for seed-stage, focus on those four areas, and make sure you’re strong in at least three of them. If you do, you'll have a compelling case. How do you run things day-to-day? i.e what's your operational workflow and team structure? Here's an overview of our team structure and process: Internally: Our team is divided into two main areas: backend (internal team) and frontend (market-facing team). There's no formal hierarchy within the backend team. We all operate as equals, defining our goals based on what needs to be developed, assigning tasks, and meeting weekly to share updates and review progress. The focus is on full ownership of tasks and accountability for getting things done. I also contribute to product development, identifying challenges and clearing obstacles to help the team move forward. Backend Team: We approach tasks based on the scope defined by customers, with no blame or hierarchy. It's like a sports team—sometimes someone excels, and other times they struggle, but we support each other and move forward together. Everyone has the creative freedom to work in the way that suits them best, but we establish regular meetings and check-ins to ensure alignment and progress. Frontend Team: For the market-facing side, we implement a hierarchy because the market expects this structure. If I present myself as "CEO," it signals authority and credibility. This distinction affects how we communicate with the market and how we build our brand. The frontend team is split into four main areas: Business Product (Software Engineering) Machine Learning Engineering R&D The C-suite sits at the top, followed by team leads, and then the executors. We distill market expectations into actionable tasks, ensuring that everyone is clear on their role and responsibilities. Process: We start by receiving market expectations and defining tasks based on them. Tasks are assigned to relevant teams, and execution happens with no communication barriers between team members. This ensures seamless collaboration and focused execution. The main goal is always effectiveness—getting things done efficiently while maintaining flexibility in how individuals approach their work. In both teams, there's an emphasis on accountability, collaboration, and clear communication, but the structure varies according to the nature of the work and external expectations.

Why Ignoring AI Agents in 2025 Will Kill Your Marketing Strategy
reddit
LLM Vibe Score0
Human Vibe Score1
frankiemuiruriThis week

Why Ignoring AI Agents in 2025 Will Kill Your Marketing Strategy

If you're still focusing solely on grabbing the attention of human beings with your marketing efforts, you're already behind. In 2025, the game will change. Good marketing will demand an in-depth understanding of the AI space, especially the AI Agent space. Why? Your ads and content won’t just be seen by humans anymore. They’ll be analyzed, indexed, and often acted upon by AI agents—automated systems that will be working on behalf of companies and consumers alike. Your New Audience: Humans + AI Agents It’s not just about appealing to people. Companies are employing AI robots to research, negotiate, and make purchasing decisions. These AI agents are fast, thorough, and unrelenting. Unlike humans, they can analyze millions of options in seconds. And if your marketing isn’t optimized for them, you’ll get filtered out before you even reach the human decision-maker. How to Prepare Your Marketing for AI Agents The companies that dominate marketing in 2025 will be the ones that master the art of capturing AI attention. To do this, marketers will need to: Understand the AI agents shaping their industry. Research how AI agents function in your niche. What are they prioritizing? How do they rank options? Create AI-friendly content. Design ads and messaging that are easily understandable and accessible to AI agents. This means clear metadata, structured data, and AI-readable formats. Invest in AI analytics. AI agents leave behind footprints. Tracking and analyzing their behavior is critical. Stay ahead of AI trends. The AI agent space is evolving rapidly. What works today might be obsolete tomorrow. How My Agency Adapted and Thrived in the AI Space At my digital agency, we saw this shift coming and decided to act early. In 2023, we started integrating AI optimization into our marketing strategies. One of our clients—a B2B SaaS company—struggled to get traction because their competitors were drowning them out in Google search rankings and ad platforms. By analyzing the algorithms and behaviors of AI agents in their space, we: Rewrote their website copy with structured data and optimized metadata that was more AI-agent friendly. Created ad campaigns with clear, concise messaging and technical attributes that AI agents could quickly process and index. Implemented predictive analytics to understand what AI agents would prioritize based on past behaviors. The results? Their website traffic doubled in three months, and their lead conversion rate skyrocketed by 40%. Over half of the traffic increase was traced back to AI agents recommending their platform to human users. The Takeaway In 2025, marketing won’t just be about human attention. It’ll be about AI attention—and that requires a completely different mindset. AI agents are not your enemy; they’re your new gatekeepers. Learn to speak their language, and you’ll dominate the marketing game.

We create AI software and provide AI automation for companies. Here is a list of the best AI tools for sales IMHO
reddit
LLM Vibe Score0
Human Vibe Score1
IntellectualAINCThis week

We create AI software and provide AI automation for companies. Here is a list of the best AI tools for sales IMHO

Here are some AI tools that are useful for sales. I tried to touch as many different parts of the sales process so the tools are all quite different but all useful for sales. I tried to include some of the best and underrated AI tools. Most of them are free so check them out if you want. I did not include ChatGPT as it can basically be used for anything with the right prompts. So these tools will be more research-oriented. A quick disclaimer – I work for the company Idealink where we create custom ChatGPT for businesses and other AI products. Apollo AI Seamless AI CoPilot AI Lavender AI Regie AI Gemini Plusdocs Make Midjourney Fireflies AI Apollo AI - Find potential customers Apollo is a platform for sales and business development. It offers a range of tools to find and engage with ideal customers. The platform has an extensive B2B database and features that streamline the sales process from prospecting to closing deals. Key Features: Extensive B2B Database: Apollo boasts a large, accurate database of over 275 million contacts, providing a wealth of potential leads and opportunities for sales teams. Data Enrichment and Lead Insights: The platform offers data enrichment capabilities, ensuring CRM systems are continuously updated with detailed and actionable lead information. AI-Driven Sales Engagement: Apollo's AI technology assists in crafting effective communication and prioritizing high-value leads, enhancing the overall sales engagement process. Comprehensive Sales Tools: The platform provides an integrated suite of tools for email, call, and social media engagement, combined with analytics and automation features to streamline the sales cycle. Tailored Solutions for Teams: Apollo offers customized solutions for different team types, including sales and business development, founders, and marketing teams, addressing specific needs and goals. Seamless AI - Sale process made easier Seamless.AI is an innovative B2B sales lead generation solution that allows sales teams to efficiently connect with their ideal customers. The platform's features provide accurate and up-to-date contact information and integrate easily with existing sales and marketing tools. Key Features: Real-Time Search Engine: Seamless.AI uses AI to scour the web in real time, ensuring the contact information for sales leads is current and accurate. Comprehensive Integration: Easily integrates with popular CRMs and sales tools like Salesforce, HubSpot, and LinkedIn Sales Navigator, enhancing productivity and eliminating manual data entry. Chrome Extension: Enhances web browsing experience for sales teams, allowing them to build lead lists directly from their browser. Pitch Intelligence and Writer: Tools for crafting effective sales messages and marketing content, personalized for each potential customer. Data Enrichment and Autopilot: Keeps customer data current and automates lead-building, supporting consistent lead generation. Buyer Intent Data and Job Changes: Offers insights into potential customers' buying intentions and keeps track of significant job changes within key accounts. CoPilot AI - Helps sales reps manage leads CoPilot AI is an advanced AI-powered sales support platform designed for B2B sales teams and agencies to drive consistent revenue growth. The tool focuses on using LinkedIn for sales prospecting, engagement, and conversion. Key Features: LinkedIn Lead Generation: Targets and automates outreach to high-intent LinkedIn leads, enhancing efficiency and scalability in lead generation. Personalized Messaging Automation: Facilitates sending of personalized, one-click messages at scale, maintaining a human touch in digital interactions. Sales Conversion Insights: Offers tools to understand and adapt to prospects' communication styles, improving the likelihood of conversion. Sales Process Optimization: Provides analytics to evaluate and refine sales strategies, identifying opportunities for improvement in the sales funnel. Industry Versatility: Adapts to diverse industries, offering tailored solutions for B2B sales, marketing, HR, and financial services sectors. Collaborative Team Tools: Enables team synchronization and collaboration, boosting productivity and synergy in sales teams Lavender AI - Email AI assistant Lavender AI is an AI-powered email tool that helps users write better emails. It provides real-time feedback and personalized suggestions to optimize email communication efficiency. Key Features: Email Coaching and Scoring: Lavender evaluates emails using AI and a vast database of email interactions, offering a score and tips for improvement. It identifies factors that might reduce the likelihood of receiving a reply, helping users refine their email content. Personalization Assistant: This feature integrates prospect data directly into the user's email platform, suggesting personalization strategies based on recipient data and personality insights to foster deeper connections. Adaptive Improvement: Lavender's scoring and recommendations evolve in real-time with changing email behaviors and practices, thanks to its generative AI and extensive data analysis, ensuring users always follow the best practices. Data-Driven Managerial Insights: The platform provides managers with valuable insights derived from actual email interactions, aiding them in coaching their teams more effectively based on real performance and communication trends. Broad Integration Capability: Lavender integrates with various email and sales platforms including Gmail, Outlook, and others, making it versatile for different user preferences and workflows. Regie AI - Great for business intelligence Regie.ai simplifies the sales prospecting process for businesses, using GenAI and automation to improve interactions with prospects. The platform offers tools like Auto-Pilot for automatic prospecting and meeting scheduling, Co-Pilot for sales rep support, and integrations with various CRM and sales engagement platforms. It also includes a Chrome Extension and CMS for content management and customization. Key Features: Automated Prospecting with Auto-Pilot: Regie.ai's Auto-Pilot feature autonomously prospects and schedules meetings, using Generative AI for Sales Agents to enhance outbound sales efforts. Audience Discovery and Content Generation: The platform identifies target accounts not in the CRM, generating relevant, on-brand content for each message, thus ensuring efficiency in list building and message personalization. Outbound Prioritization and Dynamic Engagement: It utilizes engagement and intent data to prioritize outreach to in-market prospects and adjust engagement strategies based on buyer responsiveness. Full Funnel Brand Protection and Analytics: Regie.ai ensures consistent use of marketing-approved language in all sales outreach and provides insights into campaign and document performance, thereby safeguarding brand integrity throughout the sales funnel. Gemini - AI powered conversational platform Gemini is a large language model chatbot developed by Google AI. It can generate text, translate languages, write different creative text formats, and answer your questions in an informative way. It is still under development but has learned to perform many kinds of tasks. Key features: Generate different creative text formats of text content (poems, code, scripts, musical pieces, email, letters, etc.) Answer your questions in an informative way, even if they are open ended, challenging, or strange. Translate languages Follow your instructions and complete your requests thoughtfully. Plusdocs (Plus AI) - AI tool for presentations Plus AI is a versatile tool that helps improve presentations and integrates with Slides in a simple and intuitive way. It simplifies slide creation and customization by converting text into slides and utilizing AI for various languages. Key Features: Text-to-Slide Conversion: Plus AI excels in transforming textual content into visually appealing slides, streamlining the presentation creation process. Multilingual AI Support: The tool is equipped to handle various languages, making it adaptable for a global user base. Professional Design Options: Users have access to professionally designed slide layouts, enabling the creation of polished presentations with ease. Customization and AI Design: Plus AI allows for extensive customization, including the use of AI for designing and editing slides, ensuring unique and personalized presentations. Live Snapshots and Templates: The tool offers live snapshots for real-time updates and a wide range of templates for quick and effective slide creation. Make - AI automation Make is a powerful visual platform that allows users to build and automate tasks, workflows, apps, and systems. It offers an intuitive, no-code interface that empowers users across various business functions to design and implement complex processes without the need for developer resources. Key Features: No-Code Visual Workflow Builder: Make's core feature is its user-friendly interface that allows for the creation of intricate workflows without coding expertise, making it accessible to a wide range of users. Extensive App Integration: The platform boasts compatibility with over 1000 apps, facilitating seamless connections and data sharing across diverse tools and systems. Custom Automation Solutions: Make enables personalized automation strategies, fitting various business needs from marketing automation to IT workflow control. Template Library: Users can jumpstart their automation projects with a vast collection of pre-built templates, which are customizable to fit specific workflow requirements. Enterprise-Level Solutions: Make offers advanced options for larger organizations, including enhanced security, single sign-on, custom functions, and dedicated support. Midjourney - Making sales content Midjourney is an AI-based image generation tool that changes the way we visualise and create digital art. It offers a lot of artistic possibilities, allowing users to create stunning images from text prompts. This innovative service caters to artists, designers, and anyone seeking to bring their creative visions to life. Key Features: Advanced AI Image Generation: Midjourney's core strength lies in its powerful AI algorithms, which interpret text prompts to generate detailed, high-quality images. This feature allows users to explore an endless array of visual concepts and styles. User-driven Customization: The tool offers significant control over the image creation process, enabling users to guide the AI with specific instructions, ensuring that the final output aligns closely with their vision. Diverse Artistic Styles: Midjourney can mimic various artistic styles, from classical to contemporary, providing users with a wide range of aesthetic options for their creations. Collaboration and Community Features: The platform fosters a community of users who can share, critique, and collaborate on artistic projects, enriching the creative experience. Fireflies AI - Sales meeting assistant Fireflies.ai is a powerful tool for improving team productivity and efficiency in managing meetings and voice conversations. It offers a range of features to simplify the process of capturing, organizing, and analyzing meeting content. Key Features: Automatic Meeting Transcription: Fireflies.ai can transcribe meetings held on various video-conferencing platforms and dialers. The tool captures both video and audio, providing transcripts quickly and efficiently. AI-Powered Search and Summarization: It allows users to review long meetings in a fraction of the time, highlighting key action items, tasks, and questions. Users can filter and focus on specific topics discussed in meetings. Improved Collaboration: The tool enables adding comments, pins, and reactions to specific conversation parts. Users can create and share soundbites and integrate meeting notes with popular collaboration apps such as Slack, Notion, and Asana. Conversation Intelligence: Fireflies.ai offers insights into meetings by tracking metrics like speaker talk time and sentiment. It helps in coaching team members and improving performance in sales, recruiting, and other internal processes. Workflow Automation: The AI assistant from Fireflies.ai can log call notes and activities in CRMs, create tasks through voice commands, and share meeting recaps instantly across various platforms. Comprehensive Knowledge Base: It compiles all voice conversations into an easily accessible and updatable knowledge base, with features to organize meetings into channels and set custom privacy controls. I’ll keep updating this little guide, so add your comments and I’ll try to add more tools. This is all just a personal opinion, so it’s completely cool if you disagree with it. Btw here is the link to the full blog post about all the AI tools in a bit more depth.

IVAN.ed: The platform for Social Learning ( SOMEONE CAN USE THIS IDEA BECAUSE I CURRENTLY DON'T HAVE THE TECH KNOWLEDEGE TO MAKE IT COME TRUE )
reddit
LLM Vibe Score0
Human Vibe Score1
Different_Tip8185This week

IVAN.ed: The platform for Social Learning ( SOMEONE CAN USE THIS IDEA BECAUSE I CURRENTLY DON'T HAVE THE TECH KNOWLEDEGE TO MAKE IT COME TRUE )

Overview: IVAN.ed is an innovative educational platform designed to transform the way students and educators interact and share knowledge. By combining the best elements of social media with a focus on learning, IVAN.ed aims to create a dynamic, engaging, and user-friendly environment for educational content. Key Features: Social Learning Network: A platform where students, educators, and experts can create and share educational content, similar to a social media experience but dedicated to learning. AI-Driven Content Moderation: Implementing advanced AI algorithms to ensure high-quality and relevant content, maintaining the platform’s integrity and usefulness. User Profiles and Content Creation: Users can build profiles, upload videos, create posts, and engage with content through comments, (instead of like there is the knowledge meter , based on what as taught in the videos), notes will be provided down of each video using ai. Enhanced Discovery: Advanced search and recommendation systems to help users find content that matches their interests and educational needs. Minimal Distractions: user interface designed to minimize distractions and enhance focus, making the learning experience more efficient. Goals: Accessibility: Provide a free or low-cost platform where knowledge is accessible to all. Community Engagement: Foster a vibrant learning community with meaningful interactions. Innovation: Leverage AI to maintain high standards of content and user experience. Conclusion: IVAN.ed aims to bridge the gap between traditional education and modern social media, creating an interactive and engaging space for learning. By prioritizing user experience and content quality, IVAN.ed will empower educators and learners alike, making education more accessible and impactful. THIS MESSAGE WAS GENERATED USING GPT , SINCE I AM NOT VERY GOOD AN CONVEING MY IDEAS , BUT NOW I NEED PEOPLE TO SEE THIS IDEA AND CRITIZE IT OR EVEN GIVE ME SOME IDEAS TO MAKE IT BETTER , BUT THIS IS JUST THE BLUEPRINT AND I HAVEN'T EVEN BEGUN THE ACTUAL DEVELOPMENT PHASE, BUT I AM OPEN FOR SOME HELP ! -thank you if you read it this far

ChatGPT Full Course For 2025 | ChatGPT Tutorial For Beginnners | ChatGPT Course | Simplilearn
youtube
LLM Vibe Score0.369
Human Vibe Score0.26
SimplilearnMar 28, 2025

ChatGPT Full Course For 2025 | ChatGPT Tutorial For Beginnners | ChatGPT Course | Simplilearn

🔥Purdue - Applied Generative AI Specialization - https://www.simplilearn.com/applied-ai-course?utmcampaign=C4lBsBlloL0&utmmedium=Lives&utm_source=Youtube 🔥Professional Certificate Program in Generative AI and Machine Learning - IITG (India Only) - https://www.simplilearn.com/iitg-generative-ai-machine-learning-program?utmcampaign=C4lBsBlloL0&utmmedium=Lives&utm_source=Youtube 🔥Advanced Executive Program In Applied Generative AI - https://www.simplilearn.com/applied-generative-ai-course?utmcampaign=C4lBsBlloL0&utmmedium=Lives&utm_source=Youtube This ChatGPT Full Course 2025 by Simplilearn provides a comprehensive learning journey, starting with an introduction to ChatGPT and Generative AI, followed by insights into AI job opportunities and a comparison between ChatGPT 4.0 and 4.0 Turbo. The tutorial covers prompt engineering techniques, machine learning fundamentals, and running Llama models privately. Learners will explore ChatGPT-powered application development, its role in programming, and Excel automation. The course also dives into blogging, PowerPoint automation, customer support, and finance applications. Advanced topics like RAG vs. Prompt Tuning, prompt injection, and LangChain are included, along with discussions on OpenAI's latest innovations, including Sora and Strawberry. By the end, participants will gain a strong understanding of ChatGPT’s capabilities and monetization strategies. 🚀 Following are the topics covered in the ChatGPT Full Course 2025: 00:00:00 - Introduction to ChatGPT Full Course 2025 00:09:26 - What is ChatGPT 00:10:11 - What is Gen AI 00:26:29 - How to get Job in AI 00:27:06 - ChatGPT 40 vs ChatGPT 4 01:03:14 - Chatgpt analyse 02:13:57 - Prompt Engineering Tutorial 03:10:34 - What is Machine Learning 04:07:06 - Machine Learning Tutorial 04:08:13 - Run Lama Privately 04:23:50 - Search GPT 04:25:31 - Build App Using ChatGPT 06:31:11 - ChatGPT for Programming 06:46:08 - Prompt Formulae Chatgpt 07:58:38 - Automate Excel using Chatgpt 08:00:06 - Blogging with ChatGpt 08:27:25 - Powerpoint using Chatgpt 08:28:31 - Rag Vs Prompt Tuning 09:37:43 - Chatgpt for Customer Support 11:11:06 - ChatGPT for finance 11:17:38 - Prompt injection 11:18:38 - How to Earn Money using ChatGPT 11:41:46 - Open AI Strawberry 11:52:42 - Openai sora 11:54:57 - Langchain 12:22:19 - Open ai chatgpt o1 model ✅ Subscribe to our Channel to learn more about the top Technologies: https://bit.ly/2VT4WtH ⏩ Check out the Artificial Intelligence training videos: https://youtube.com/playlist?list=PLEiEAq2VkUULa5aOQmO_al2VVmhC-eqeI #gpt #chatgpt #chatgptforbeginners #chatgptcourse #genai #generativeai #artificialintelligence #ai #machinelearning #llm #simplilearn #2025 ➡️ About Professional Certificate Program in Generative AI and Machine Learning Dive into the future of AI with our Generative AI & Machine Learning course, in collaboration with E&ICT Academy, IIT Guwahati. Learn tools like ChatGPT, OpenAI, Hugging Face, Python, and more. Join masterclasses led by IITG faculty, engage in hands-on projects, and earn Executive Alumni Status. Key Features: ✅ Program completion certificate from E&ICT Academy, IIT Guwahati ✅ Curriculum delivered in live virtual classes by seasoned industry experts ✅ Exposure to the latest AI advancements, such as generative AI, LLMs, and prompt engineering ✅ Interactive live-virtual masterclasses delivered by esteemed IIT Guwahati faculty ✅ Opportunity to earn an 'Executive Alumni Status' from E&ICT Academy, IIT Guwahati ✅ Eligibility for a campus immersion program organized at IIT Guwahati ✅ Exclusive hackathons and “ask-me-anything” sessions by IBM ✅ Certificates for IBM courses and industry masterclasses by IBM experts ✅ Practical learning through 25+ hands-on projects and 3 industry-oriented capstone projects ✅ Access to a wide array of AI tools such as ChatGPT, Hugging Face, DALL-E 2, Midjourney and more ✅ Simplilearn's JobAssist helps you get noticed by top hiring companies Skills Covered: ✅ Generative AI ✅ Prompt Engineering ✅ Chatbot Development ✅ Supervised and Unsupervised Learning ✅ Model Training and Optimization ✅ Model Evaluation and Validation ✅ Ensemble Methods ✅ Deep Learning ✅ Natural Language Processing ✅ Computer Vision ✅ Reinforcement Learning ✅ Machine Learning Algorithms ✅ Speech Recognition ✅ Statistics Learning Path: ✅ Program Induction ✅ Programming Fundamentals ✅ Python for Data Science (IBM) ✅ Applied Data Science with Python ✅ Machine Learning ✅ Deep Learning with TensorFlow (IBM) ✅ Deep Learning Specialization ✅ Essentials of Generative AI, Prompt Engineering & ChatGPT ✅ Advanced Generative AI ✅ Capstone Electives: ✅ ADL & Computer Vision ✅ NLP and Speech Recognition ✅ Reinforcement Learning ✅ Academic Masterclass ✅ Industry Masterclass 👉 Learn More At: https://www.simplilearn.com/iitg-generative-ai-machine-learning-program?utmcampaign=C4lBsBlloL0&utmmedium=Lives&utm_source=Youtube

mentals-ai
github
LLM Vibe Score0.476
Human Vibe Score0.004852164397547106
turing-machinesMar 28, 2025

mentals-ai

Mentals AI is a tool designed for creating and operating agents that feature loops, memory, and various tools, all through straightforward markdown files with a .gen extension. Think of an agent file as an executable file. You focus entirely on the logic of the agent, eliminating the necessity to write scaffolding code in Python or any other language. Essentially, it redefines the foundational frameworks for future AI applications 🍓 [!NOTE] [work in progress] A local vector database to store your chats with the agents as well as your private information. See memory branch. [work in progress] Web UI with agents, tools, and vector storage Getting Started Differences from Other Frameworks Key Concepts Instruction (prompt) Working Memory (context) Short-Term Memory (experimental) Control flow: From strings to algorithms Roadmap The Idea 📌 Examples Word chain game in a self-loop controlled by LLM: !Word Chain game in a loop NLOP — Natural Language Operation Or more complex use cases: | 🔄 Any multi-agent interactions | 👾 Space Invaders generator agent | 🍄 2D platformer generator agent | |--------------------|-----------|--------------| |!react | !spaceinvaders.gen | !mario.gen | Or help with the content: Collect YouTube videos on a given topic and save them to a .csv file with the videos, views, channel name, and link; Get the transcription from the video and create a table of contents; Take top news from Hacker News, choose a topic and write an article on the topic with the participation of the critic, and save to a file. All of the above examples are located in the agents folder. [!NOTE] Llama3 support is available for providers using a compatible OpenAI API. 🚀 Getting Started Begin by securing an OpenAI API key through the creation of an OpenAI account. If you already have an API key, skip this step. 🏗️ Build and Run Prerequisites Before building the project, ensure the following dependencies are installed: libcurl: Used for making HTTP requests libfmt: Provides an API for formatting pgvector: Vector operations with PostgreSQL poppler: Required for PDF processing Depending on your operating system, you can install these using the following commands: Linux macOS Windows For Windows, it's recommended to use vcpkg or a similar package manager: pgvector installation [!NOTE] In the main branch you can skip this step Build from sources Docker, Homebrew, PGXN, APT, etc. Clone the repository Configuration Place your API key in the config.toml file: Build the project Run 🆚 Differences from Other Frameworks Mentals AI distinguishes itself from other frameworks in three significant ways: The Agent Executor 🧠 operates through a recursive loop. The LLM determines the next steps: selecting instructions (prompts) and managing data based on previous loops. This recursive decision-making process is integral to our system, outlined in mentalssystem.prompt Agents of any complexity can be created using Markdown, eliminating the need for traditional programming languages. However, Python can be integrated directly into the agent's Markdown script if necessary. Unlike platforms that include preset reasoning frameworks, Mentals AI serves as a blank canvas. It enables the creation and integration of your own reasoning frameworks, including existing ones: Tree of Thoughts, ReAct, Self-Discovery, Auto-CoT, and others. One can also link these frameworks together into more complex sequences, even creating a network of various reasoning frameworks. 🗝️ Key Concepts The agent file is a textual description of the agent instructions with a .gen extension. 📖 Instruction (prompt) Instruction is the basic component of an agent in Mentals. An agent can consist of one or more instructions, which can refer to each other. Instructions can be written in free form, but they always have a name that starts with the # symbol. The use: directive is used to specify a reference to other instructions. Multiple references are listed separated by commas. Below is an example with two instructions root and meme_explain with a reference: In this example, the root instruction calls the memeexplain instruction. The response from memeexplain is then returned to the instruction from which it was called, namely the root. An instruction can take an input parameter, which is automatically generated based on the context when the instruction is called. To specify the input data more precisely, you can use a free-form prompt in the input: directive, such as a JSON object or null. Using a document for input: Using a JSON object as input: [!NOTE] Instruction calls are implemented independently from function or tool calls at OpenAI, enabling the operation of agents with models like Llama3. The implementation of instruction calls is transparent and included in the mentals_system.prompt file. 🛠️ Tool Tool is a kind of instruction. Mentals has a set of native tools to handle message output, user input, file handling, Python interpreter, Bash commands, and Short-term memory. Ask user example: File handling example: The full list of native tools is listed in the file native_tools.toml. 🧠 Working Memory (context) Each instruction has its own working memory — context. When exiting an instruction and re-entering it, the context is kept by default. To clear the context when exiting an instruction, you can use the keep_context: false directive: By default, the size of the instruction context is not limited. To limit the context, there is a directive max_context: number which specifies that only the number of the most recent messages should be stored. Older messages will be pushed out of the context. This feature is useful when you want to keep the most recent data in context so that older data does not affect the chain of reasoning. ⏳ Short-Term Memory (experimental) Short-term memory allows for the storage of intermediate results from an agent's activities, which can then be used for further reasoning. The contents of this memory are accessible across all instruction contexts. The memory tool is used to store data. When data is stored, a keyword and a description of the content are generated. In the example below, the meme_recall instruction is aware of the meme because it was previously stored in memory. ⚙️ Control flow: From strings to algorithms The control flow, which includes conditions, instruction calls, and loops (such as ReAct, Auto-CoT, etc.), is fully expressed in natural language. This method enables the creation of semantic conditions that direct data stream branching. For instance, you can request an agent to autonomously play a word chain game in a loop or establish an ambiguous exit condition: exit the loop if you are satisfied with the result. Here, the language model and its context determine whether to continue or stop. All this is achieved without needing to define flow logic in Python or any other programming language. ⚖️ Reason Action (ReAct) example 🌳 Tree of Thoughts (ToT) example The idea behind ToT is to generate multiple ideas to solve a problem and then evaluate their value. Valuable ideas are kept and developed, other ideas are discarded. Let's take the example of the 24 game. The 24 puzzle is an arithmetical puzzle in which the objective is to find a way to manipulate four integers so that the end result is 24. First, we define the instruction that creates and manipulates the tree data structure. The model knows what a tree is and can represent it in any format, from plain text to XML/JSON or any custom format. In this example, we will use the plain text format: Next, we need to initialize the tree with initial data, let's start with the root instruction: Calling the root instruction will suggest 8 possible next steps to calculate with the first 2 numbers and store these steps as tree nodes. Further work by the agent results in the construction of a tree that is convenient for the model to understand and infer the final answer. A complete example is contained in the agents/treestructure.gen 🗺️ Roadmap [ ] Web UI -- WIP [ ] Vector database tools -- WIP [ ] Agent's experience (experimental) [ ] Tools: Image generation, Browser ✨ The Idea The concept originated from studies on psychoanalysis Executive functions, Exploring Central Executive, Alan Baddeley, 1996. He described a system that orchestrates cognitive processes and working memory, facilitating retrievals from long-term memory. The LLM functions as System 1, processing queries and executing instructions without inherent motivation or goal-setting. So, what then is System 2? Drawing from historical insights now reconsidered through a scientific lens: The central executive, or executive functions, is crucial for controlled processing in working memory. It manages tasks including directing attention, maintaining task objectives, decision-making, and memory retrieval. This sparks an intriguing possibility: constructing more sophisticated agents by integrating System 1 and System 2. The LLM, as the cognitive executor System 1, works in tandem with the Central Executive System 2, which governs and controls the LLM. This partnership forms the dual relationship foundational to Mentals AI.

AITreasureBox
github
LLM Vibe Score0.447
Human Vibe Score0.1014145151561518
superiorluMar 28, 2025

AITreasureBox

AI TreasureBox English | 中文 Collect practical AI repos, tools, websites, papers and tutorials on AI. Translated from ChatGPT, picture from Midjourney. Catalog Repos Tools Websites Report&Paper Tutorials Repos updated repos and stars every 2 hours and re-ranking automatically. | No. | Repos | Description | | ----:|:-----------------------------------------|:------------------------------------------------------------------------------------------------------| | 1|🔥codecrafters-io/build-your-own-x !2025-03-28364681428|Master programming by recreating your favorite technologies from scratch.| | 2|sindresorhus/awesome !2025-03-28353614145|😎 Awesome lists about all kinds of interesting topics| | 3|public-apis/public-apis !2025-03-28334299125|A collective list of free APIs| | 4|kamranahmedse/developer-roadmap !2025-03-2831269540|Interactive roadmaps, guides and other educational content to help developers grow in their careers.| | 5|vinta/awesome-python !2025-03-28238581114|A curated list of awesome Python frameworks, libraries, software and resources| | 6|practical-tutorials/project-based-learning !2025-03-28222661124|Curated list of project-based tutorials| | 7|tensorflow/tensorflow !2025-03-281888714|An Open Source Machine Learning Framework for Everyone| | 8|Significant-Gravitas/AutoGPT !2025-03-2817391338|An experimental open-source attempt to make GPT-4 fully autonomous.| | 9|jackfrued/Python-100-Days !2025-03-2816305141|Python - 100天从新手到大师| | 10|AUTOMATIC1111/stable-diffusion-webui !2025-03-2815011553|Stable Diffusion web UI| | 11|huggingface/transformers !2025-03-2814207850|🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.| | 12|ollama/ollama !2025-03-28135166151|Get up and running with Llama 2, Mistral, Gemma, and other large language models.| | 13|f/awesome-chatgpt-prompts !2025-03-2812212738 |This repo includes ChatGPT prompt curation to use ChatGPT better.| | 14|justjavac/free-programming-books-zhCN !2025-03-2811316119|📚 免费的计算机编程类中文书籍,欢迎投稿| | 15|krahets/hello-algo !2025-03-2811107930|《Hello 算法》:动画图解、一键运行的数据结构与算法教程。支持 Python, Java, C++, C, C#, JS, Go, Swift, Rust, Ruby, Kotlin, TS, Dart 代码。简体版和繁体版同步更新,English version ongoing| | 16|yt-dlp/yt-dlp !2025-03-28105801114|A feature-rich command-line audio/video downloader| | 17|langchain-ai/langchain !2025-03-2810449479|⚡ Building applications with LLMs through composability ⚡| | 18|goldbergyoni/nodebestpractices !2025-03-281021629|✅ The Node.js best practices list (July 2024)| | 19|puppeteer/puppeteer !2025-03-289018212|JavaScript API for Chrome and Firefox| | 20|pytorch/pytorch !2025-03-288833938|Tensors and Dynamic neural networks in Python with strong GPU acceleration| | 21|neovim/neovim !2025-03-288781482|Vim-fork focused on extensibility and usability| | 22|🔥🔥langgenius/dify !2025-03-2887342639 |One API for plugins and datasets, one interface for prompt engineering and visual operation, all for creating powerful AI applications.| | 23|mtdvio/every-programmer-should-know !2025-03-28867069|A collection of (mostly) technical things every software developer should know about| | 24|open-webui/open-webui !2025-03-2886025159|User-friendly WebUI for LLMs (Formerly Ollama WebUI)| | 25|ChatGPTNextWeb/NextChat !2025-03-288231521|✨ Light and Fast AI Assistant. Support: Web | | 26|supabase/supabase !2025-03-287990956|The open source Firebase alternative.| | 27|openai/whisper !2025-03-287905542|Robust Speech Recognition via Large-Scale Weak Supervision| | 28|home-assistant/core !2025-03-287773219|🏡 Open source home automation that puts local control and privacy first.| | 29|tensorflow/models !2025-03-28774694|Models and examples built with TensorFlow| | 30| ggerganov/llama.cpp !2025-03-287731836 | Port of Facebook's LLaMA model in C/C++ | | 31|3b1b/manim !2025-03-287641918|Animation engine for explanatory math videos| | 32|microsoft/generative-ai-for-beginners !2025-03-287623860|12 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/| | 33|nomic-ai/gpt4all !2025-03-28729285 |gpt4all: an ecosystem of open-source chatbots trained on a massive collection of clean assistant data including code, stories and dialogue| | 34|comfyanonymous/ComfyUI !2025-03-2872635111|The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.| | 35|bregman-arie/devops-exercises !2025-03-2872225209|Linux, Jenkins, AWS, SRE, Prometheus, Docker, Python, Ansible, Git, Kubernetes, Terraform, OpenStack, SQL, NoSQL, Azure, GCP, DNS, Elastic, Network, Virtualization. DevOps Interview Questions| | 36|elastic/elasticsearch !2025-03-28721419|Free and Open, Distributed, RESTful Search Engine| | 37|🔥n8n-io/n8n !2025-03-2872093495|Free and source-available fair-code licensed workflow automation tool. Easily automate tasks across different services.| | 38|fighting41love/funNLP !2025-03-287200422|The Most Powerful NLP-Weapon Arsenal| | 39|hoppscotch/hoppscotch !2025-03-287060134|Open source API development ecosystem - https://hoppscotch.io (open-source alternative to Postman, Insomnia)| | 40|abi/screenshot-to-code !2025-03-286932817|Drop in a screenshot and convert it to clean HTML/Tailwind/JS code| | 41|binary-husky/gptacademic !2025-03-28680374|Academic Optimization of GPT| | 42|d2l-ai/d2l-zh !2025-03-286774142|Targeting Chinese readers, functional and open for discussion. The Chinese and English versions are used for teaching in over 400 universities across more than 60 countries| | 43|josephmisiti/awesome-machine-learning !2025-03-286739215|A curated list of awesome Machine Learning frameworks, libraries and software.| | 44|grafana/grafana !2025-03-286725414|The open and composable observability and data visualization platform. Visualize metrics, logs, and traces from multiple sources like Prometheus, Loki, Elasticsearch, InfluxDB, Postgres and many more.| | 45|python/cpython !2025-03-286602218|The Python programming language| | 46|apache/superset !2025-03-286519020|Apache Superset is a Data Visualization and Data Exploration Platform| | 47|xtekky/gpt4free !2025-03-28639391 |decentralizing the Ai Industry, free gpt-4/3.5 scripts through several reverse engineered API's ( poe.com, phind.com, chat.openai.com etc...)| | 48|sherlock-project/sherlock !2025-03-286332536|Hunt down social media accounts by username across social networks| | 49|twitter/the-algorithm !2025-03-28630586 |Source code for Twitter's Recommendation Algorithm| | 50|keras-team/keras !2025-03-28627835|Deep Learning for humans| | 51|openai/openai-cookbook !2025-03-28625136 |Examples and guides for using the OpenAI API| | 52|immich-app/immich !2025-03-286238670|High performance self-hosted photo and video management solution.| | 53|AppFlowy-IO/AppFlowy !2025-03-286173528|Bring projects, wikis, and teams together with AI. AppFlowy is an AI collaborative workspace where you achieve more without losing control of your data. The best open source alternative to Notion.| | 54|scikit-learn/scikit-learn !2025-03-286158212|scikit-learn: machine learning in Python| | 55|binhnguyennus/awesome-scalability !2025-03-286117021|The Patterns of Scalable, Reliable, and Performant Large-Scale Systems| | 56|labmlai/annotateddeeplearningpaperimplementations !2025-03-285951726|🧑‍🏫 59 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠| | 57|OpenInterpreter/open-interpreter !2025-03-285894710|A natural language interface for computers| | 58|lobehub/lobe-chat !2025-03-285832054|🤖 Lobe Chat - an open-source, extensible (Function Calling), high-performance chatbot framework. It supports one-click free deployment of your private ChatGPT/LLM web application.| | 59|meta-llama/llama !2025-03-28579536|Inference code for Llama models| | 60|nuxt/nuxt !2025-03-28566437|The Intuitive Vue Framework.| | 61|imartinez/privateGPT !2025-03-28555192|Interact with your documents using the power of GPT, 100% privately, no data leaks| | 62|Stirling-Tools/Stirling-PDF !2025-03-285500846|#1 Locally hosted web application that allows you to perform various operations on PDF files| | 63|PlexPt/awesome-chatgpt-prompts-zh !2025-03-285459720|ChatGPT Chinese Training Guide. Guidelines for various scenarios. Learn how to make it listen to you| | 64|dair-ai/Prompt-Engineering-Guide !2025-03-285451025 |🐙 Guides, papers, lecture, notebooks and resources for prompt engineering| | 65|ageitgey/facerecognition !2025-03-28544382|The world's simplest facial recognition api for Python and the command line| | 66|CorentinJ/Real-Time-Voice-Cloning !2025-03-285384814|Clone a voice in 5 seconds to generate arbitrary speech in real-time| | 67|geekan/MetaGPT !2025-03-285375376|The Multi-Agent Meta Programming Framework: Given one line Requirement, return PRD, Design, Tasks, Repo | | 68|gpt-engineer-org/gpt-engineer !2025-03-285367419|Specify what you want it to build, the AI asks for clarification, and then builds it.| | 69|lencx/ChatGPT !2025-03-2853653-3|🔮 ChatGPT Desktop Application (Mac, Windows and Linux)| | 70|deepfakes/faceswap !2025-03-28535672|Deepfakes Software For All| | 71|langflow-ai/langflow !2025-03-285319584|Langflow is a low-code app builder for RAG and multi-agent AI applications. It’s Python-based and agnostic to any model, API, or database.| | 72|commaai/openpilot !2025-03-28529759|openpilot is an operating system for robotics. Currently, it upgrades the driver assistance system on 275+ supported cars.| | 73|clash-verge-rev/clash-verge-rev !2025-03-2852848124|Continuation of Clash Verge - A Clash Meta GUI based on Tauri (Windows, MacOS, Linux)| | 74|All-Hands-AI/OpenHands !2025-03-285150675|🙌 OpenHands: Code Less, Make More| | 75|xai-org/grok-1 !2025-03-28502504|Grok open release| | 76|meilisearch/meilisearch !2025-03-284999122|A lightning-fast search API that fits effortlessly into your apps, websites, and workflow| | 77|🔥browser-use/browser-use !2025-03-2849910294|Make websites accessible for AI agents| | 78|jgthms/bulma !2025-03-28496783|Modern CSS framework based on Flexbox| | 79|facebookresearch/segment-anything !2025-03-284947116|The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.| |!green-up-arrow.svg 80|hacksider/Deep-Live-Cam !2025-03-2848612146|real time face swap and one-click video deepfake with only a single image (uncensored)| |!red-down-arrow 81|mlabonne/llm-course !2025-03-284860934|Course with a roadmap and notebooks to get into Large Language Models (LLMs).| | 82|PaddlePaddle/PaddleOCR !2025-03-284785530|Awesome multilingual OCR toolkits based on PaddlePaddle (practical ultra lightweight OCR system, support 80+ languages recognition, provide data annotation and synthesis tools, support training and deployment among server, mobile, embedded and IoT devices)| | 83|alist-org/alist !2025-03-284732618|🗂️A file list/WebDAV program that supports multiple storages, powered by Gin and Solidjs. / 一个支持多存储的文件列表/WebDAV程序,使用 Gin 和 Solidjs。| | 84|infiniflow/ragflow !2025-03-2847027129|RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.| | 85|Avik-Jain/100-Days-Of-ML-Code !2025-03-284679312|100 Days of ML Coding| | 86|v2ray/v2ray-core !2025-03-28458706|A platform for building proxies to bypass network restrictions.| | 87|hiyouga/LLaMA-Factory !2025-03-284555881|Easy-to-use LLM fine-tuning framework (LLaMA, BLOOM, Mistral, Baichuan, Qwen, ChatGLM)| | 88|Asabeneh/30-Days-Of-Python !2025-03-284544930|30 days of Python programming challenge is a step-by-step guide to learn the Python programming language in 30 days. This challenge may take more than100 days, follow your own pace. These videos may help too: https://www.youtube.com/channel/UC7PNRuno1rzYPb1xLa4yktw| | 89|type-challenges/type-challenges !2025-03-284488511|Collection of TypeScript type challenges with online judge| | 90|lllyasviel/Fooocus !2025-03-284402716|Focus on prompting and generating| | 91|RVC-Boss/GPT-SoVITS !2025-03-284327738|1 min voice data can also be used to train a good TTS model! (few shot voice cloning)| | 92|rasbt/LLMs-from-scratch !2025-03-284320667|Implementing a ChatGPT-like LLM from scratch, step by step| | 93|oobabooga/text-generation-webui !2025-03-284302012 |A gradio web UI for running Large Language Models like LLaMA, llama.cpp, GPT-J, OPT, and GALACTICA.| | 94|vllm-project/vllm !2025-03-2842982102|A high-throughput and memory-efficient inference and serving engine for LLMs| | 95|dani-garcia/vaultwarden !2025-03-284297121|Unofficial Bitwarden compatible server written in Rust, formerly known as bitwarden_rs| | 96|microsoft/autogen !2025-03-284233049|Enable Next-Gen Large Language Model Applications. Join our Discord: https://discord.gg/pAbnFJrkgZ| | 97|jeecgboot/JeecgBoot !2025-03-284205920|🔥「企业级低代码平台」前后端分离架构SpringBoot 2.x/3.x,SpringCloud,Ant Design&Vue3,Mybatis,Shiro,JWT。强大的代码生成器让前后端代码一键生成,无需写任何代码! 引领新的开发模式OnlineCoding->代码生成->手工MERGE,帮助Java项目解决70%重复工作,让开发更关注业务,既能快速提高效率,帮助公司节省成本,同时又不失灵活性。| | 98|Mintplex-Labs/anything-llm !2025-03-284186955|A full-stack application that turns any documents into an intelligent chatbot with a sleek UI and easier way to manage your workspaces.| | 99|THUDM/ChatGLM-6B !2025-03-28410192 |ChatGLM-6B: An Open Bilingual Dialogue Language Model| | 100|hpcaitech/ColossalAI !2025-03-28406902|Making large AI models cheaper, faster and more accessible| | 101|Stability-AI/stablediffusion !2025-03-28406337|High-Resolution Image Synthesis with Latent Diffusion Models| | 102|mingrammer/diagrams !2025-03-28405063|🎨 Diagram as Code for prototyping cloud system architectures| | 103|Kong/kong !2025-03-28404616|🦍 The Cloud-Native API Gateway and AI Gateway.| | 104|getsentry/sentry !2025-03-284040913|Developer-first error tracking and performance monitoring| | 105| karpathy/nanoGPT !2025-03-284034613 |The simplest, fastest repository for training/finetuning medium-sized GPTs| | 106|fastlane/fastlane !2025-03-2840014-1|🚀 The easiest way to automate building and releasing your iOS and Android apps| | 107|psf/black !2025-03-28399765|The uncompromising Python code formatter| | 108|OpenBB-finance/OpenBBTerminal !2025-03-283972074 |Investment Research for Everyone, Anywhere.| | 109|2dust/v2rayNG !2025-03-283943415|A V2Ray client for Android, support Xray core and v2fly core| | 110|apache/airflow !2025-03-283937314|Apache Airflow - A platform to programmatically author, schedule, and monitor workflows| | 111|KRTirtho/spotube !2025-03-283902746|🎧 Open source Spotify client that doesn't require Premium nor uses Electron! Available for both desktop & mobile!| | 112|coqui-ai/TTS !2025-03-283889719 |🐸💬 - a deep learning toolkit for Text-to-Speech, battle-tested in research and production| | 113|ggerganov/whisper.cpp !2025-03-283882116|Port of OpenAI's Whisper model in C/C++| | 114|ultralytics/ultralytics !2025-03-283866951|NEW - YOLOv8 🚀 in PyTorch > ONNX > OpenVINO > CoreML > TFLite| | 115|typst/typst !2025-03-283863914|A new markup-based typesetting system that is powerful and easy to learn.| | 116|streamlit/streamlit !2025-03-283845828|Streamlit — A faster way to build and share data apps.| | 117|LC044/WeChatMsg !2025-03-283836931|提取微信聊天记录,将其导出成HTML、Word、Excel文档永久保存,对聊天记录进行分析生成年度聊天报告,用聊天数据训练专属于个人的AI聊天助手| | 118|lm-sys/FastChat !2025-03-283822112 |An open platform for training, serving, and evaluating large languages. Release repo for Vicuna and FastChat-T5.| | 119|NaiboWang/EasySpider !2025-03-283819013|A visual no-code/code-free web crawler/spider易采集:一个可视化浏览器自动化测试/数据采集/爬虫软件,可以无代码图形化的设计和执行爬虫任务。别名:ServiceWrapper面向Web应用的智能化服务封装系统。| | 120|microsoft/DeepSpeed !2025-03-283765816 |A deep learning optimization library that makes distributed training and inference easy, efficient, and effective| | 121|QuivrHQ/quivr !2025-03-28376067|Your GenAI Second Brain 🧠 A personal productivity assistant (RAG) ⚡️🤖 Chat with your docs (PDF, CSV, ...) & apps using Langchain, GPT 3.5 / 4 turbo, Private, Anthropic, VertexAI, Ollama, LLMs, that you can share with users ! Local & Private alternative to OpenAI GPTs & ChatGPT powered by retrieval-augmented generation.| | 122|freqtrade/freqtrade !2025-03-283757817 |Free, open source crypto trading bot| | 123|suno-ai/bark !2025-03-28373178 |🔊 Text-Prompted Generative Audio Model| | 124|🔥cline/cline !2025-03-2837307282|Autonomous coding agent right in your IDE, capable of creating/editing files, executing commands, and more with your permission every step of the way.| | 125|LAION-AI/Open-Assistant !2025-03-28372712 |OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.| | 126|penpot/penpot !2025-03-283716217|Penpot: The open-source design tool for design and code collaboration| | 127|gradio-app/gradio !2025-03-283713320|Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work!| | 128|FlowiseAI/Flowise !2025-03-283667135 |Drag & drop UI to build your customized LLM flow using LangchainJS| | 129|SimplifyJobs/Summer2025-Internships !2025-03-28366506|Collection of Summer 2025 tech internships!| | 130|TencentARC/GFPGAN !2025-03-28365027 |GFPGAN aims at developing Practical Algorithms for Real-world Face Restoration.| | 131|ray-project/ray !2025-03-283626819|Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a toolkit of libraries (Ray AIR) for accelerating ML workloads.| | 132|babysor/MockingBird !2025-03-28360498|🚀AI拟声: 5秒内克隆您的声音并生成任意语音内容 Clone a voice in 5 seconds to generate arbitrary speech in real-time| | 133|unslothai/unsloth !2025-03-283603691|5X faster 50% less memory LLM finetuning| | 134|zhayujie/chatgpt-on-wechat !2025-03-283600124 |Wechat robot based on ChatGPT, which uses OpenAI api and itchat library| | 135|upscayl/upscayl !2025-03-283599824|🆙 Upscayl - Free and Open Source AI Image Upscaler for Linux, MacOS and Windows built with Linux-First philosophy.| | 136|freeCodeCamp/devdocs !2025-03-28359738|API Documentation Browser| | 137|XingangPan/DragGAN !2025-03-28359043 |Code for DragGAN (SIGGRAPH 2023)| | 138|2noise/ChatTTS !2025-03-283543922|ChatTTS is a generative speech model for daily dialogue.| | 139|google-research/google-research !2025-03-28352207 |Google Research| | 140|karanpratapsingh/system-design !2025-03-28351003|Learn how to design systems at scale and prepare for system design interviews| | 141|lapce/lapce !2025-03-28350855|Lightning-fast and Powerful Code Editor written in Rust| | 142| microsoft/TaskMatrix !2025-03-2834500-3 | Talking, Drawing and Editing with Visual Foundation Models| | 143|chatchat-space/Langchain-Chatchat !2025-03-283442020|Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM) QA app with langchain| | 144|unclecode/crawl4ai !2025-03-283434163|🔥🕷️ Crawl4AI: Open-source LLM Friendly Web Crawler & Scrapper| | 145|Bin-Huang/chatbox !2025-03-283374733 |A desktop app for GPT-4 / GPT-3.5 (OpenAI API) that supports Windows, Mac & Linux| | 146|milvus-io/milvus !2025-03-283366525 |A cloud-native vector database, storage for next generation AI applications| | 147|mendableai/firecrawl !2025-03-2833297128|🔥 Turn entire websites into LLM-ready markdown| | 148|pola-rs/polars !2025-03-283269320|Fast multi-threaded, hybrid-out-of-core query engine focussing on DataFrame front-ends| | 149|Pythagora-io/gpt-pilot !2025-03-28325321|PoC for a scalable dev tool that writes entire apps from scratch while the developer oversees the implementation| | 150|hashicorp/vault !2025-03-28320797|A tool for secrets management, encryption as a service, and privileged access management| | 151|shardeum/shardeum !2025-03-28319580|Shardeum is an EVM based autoscaling blockchain| | 152|Chanzhaoyu/chatgpt-web !2025-03-28319242 |A demonstration website built with Express and Vue3 called ChatGPT| | 153|lllyasviel/ControlNet !2025-03-283186413 |Let us control diffusion models!| | 154|google/jax !2025-03-28317727|Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more| | 155|facebookresearch/detectron2 !2025-03-28315987|Detectron2 is a platform for object detection, segmentation and other visual recognition tasks.| | 156|myshell-ai/OpenVoice !2025-03-28315233|Instant voice cloning by MyShell| | 157|TheAlgorithms/C-Plus-Plus !2025-03-283151411|Collection of various algorithms in mathematics, machine learning, computer science and physics implemented in C++ for educational purposes.| | 158|hiroi-sora/Umi-OCR !2025-03-283138129|OCR图片转文字识别软件,完全离线。截屏/批量导入图片,支持多国语言、合并段落、竖排文字。可排除水印区域,提取干净的文本。基于 PaddleOCR 。| | 159|mudler/LocalAI !2025-03-283127815|🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.| | 160|facebookresearch/fairseq !2025-03-28312124 |Facebook AI Research Sequence-to-Sequence Toolkit written in Python.| | 161|alibaba/nacos !2025-03-28310559|an easy-to-use dynamic service discovery, configuration and service management platform for building cloud native applications.| | 162|yunjey/pytorch-tutorial !2025-03-28310326|PyTorch Tutorial for Deep Learning Researchers| | 163|v2fly/v2ray-core !2025-03-28307448|A platform for building proxies to bypass network restrictions.| | 164|mckaywrigley/chatbot-ui !2025-03-283067714|The open-source AI chat interface for everyone.| | 165|TabbyML/tabby !2025-03-28305949 |Self-hosted AI coding assistant| | 166|deepseek-ai/awesome-deepseek-integration !2025-03-283053193|| | 167|danielmiessler/fabric !2025-03-283028914|fabric is an open-source framework for augmenting humans using AI.| | 168|xinntao/Real-ESRGAN !2025-03-283026623 |Real-ESRGAN aims at developing Practical Algorithms for General Image/Video Restoration.| | 169|paul-gauthier/aider !2025-03-283014642|aider is GPT powered coding in your terminal| | 170|tatsu-lab/stanfordalpaca !2025-03-28299022 |Code and documentation to train Stanford's Alpaca models, and generate the data.| | 171|DataTalksClub/data-engineering-zoomcamp !2025-03-282971817|Free Data Engineering course!| | 172|HeyPuter/puter !2025-03-282967014|🌐 The Internet OS! Free, Open-Source, and Self-Hostable.| | 173|mli/paper-reading !2025-03-282962314|Classic Deep Learning and In-Depth Reading of New Papers Paragraph by Paragraph| | 174|linexjlin/GPTs !2025-03-28295568|leaked prompts of GPTs| | 175|s0md3v/roop !2025-03-28295286 |one-click deepfake (face swap)| | 176|JushBJJ/Mr.-Ranedeer-AI-Tutor !2025-03-2829465-1 |A GPT-4 AI Tutor Prompt for customizable personalized learning experiences.| | 177|opendatalab/MinerU !2025-03-282927074|A one-stop, open-source, high-quality data extraction tool, supports PDF/webpage/e-book extraction.一站式开源高质量数据提取工具,支持PDF/网页/多格式电子书提取。| | 178|mouredev/Hello-Python !2025-03-282920720|Curso para aprender el lenguaje de programación Python desde cero y para principiantes. 75 clases, 37 horas en vídeo, código, proyectos y grupo de chat. Fundamentos, frontend, backend, testing, IA...| | 179|Lightning-AI/pytorch-lightning !2025-03-28292039|Pretrain, finetune and deploy AI models on multiple GPUs, TPUs with zero code changes.| | 180|crewAIInc/crewAI !2025-03-282919344|Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.| | 181|facebook/folly !2025-03-282916612|An open-source C++ library developed and used at Facebook.| | 182|google-ai-edge/mediapipe !2025-03-28291519|Cross-platform, customizable ML solutions for live and streaming media.| | 183| getcursor/cursor !2025-03-282892025 | An editor made for programming with AI| | 184|chatanywhere/GPTAPIfree !2025-03-282856424|Free ChatGPT API Key, Free ChatGPT API, supports GPT-4 API (free), ChatGPT offers a free domestic forwarding API that allows direct connections without the need for a proxy. It can be used in conjunction with software/plugins like ChatBox, significantly reducing interface usage costs. Enjoy unlimited and unrestricted chatting within China| | 185|meta-llama/llama3 !2025-03-28285552|The official Meta Llama 3 GitHub site| | 186|tinygrad/tinygrad !2025-03-282845811|You like pytorch? You like micrograd? You love tinygrad! ❤️| | 187|google-research/tuningplaybook !2025-03-282841514|A playbook for systematically maximizing the performance of deep learning models.| | 188|huggingface/diffusers !2025-03-282830222|🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.| | 189|tokio-rs/tokio !2025-03-28282408|A runtime for writing reliable asynchronous applications with Rust. Provides I/O, networking, scheduling, timers, ...| | 190|RVC-Project/Retrieval-based-Voice-Conversion-WebUI !2025-03-282823817|Voice data !2025-03-282822612|Jan is an open source alternative to ChatGPT that runs 100% offline on your computer| | 192|openai/CLIP !2025-03-282814720|CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image| | 193|🔥khoj-ai/khoj !2025-03-2828112313|Your AI second brain. A copilot to get answers to your questions, whether they be from your own notes or from the internet. Use powerful, online (e.g gpt4) or private, local (e.g mistral) LLMs. Self-host locally or use our web app. Access from Obsidian, Emacs, Desktop app, Web or Whatsapp.| | 194| acheong08/ChatGPT !2025-03-2828054-2 | Reverse engineered ChatGPT API | | 195|iperov/DeepFaceLive !2025-03-28279345 |Real-time face swap for PC streaming or video calls| | 196|eugeneyan/applied-ml !2025-03-28278471|📚 Papers & tech blogs by companies sharing their work on data science & machine learning in production.| | 197|XTLS/Xray-core !2025-03-282778213|Xray, Penetrates Everything. Also the best v2ray-core, with XTLS support. Fully compatible configuration.| | 198|feder-cr/JobsApplierAIAgent !2025-03-282776410|AutoJobsApplierAI_Agent aims to easy job hunt process by automating the job application process. Utilizing artificial intelligence, it enables users to apply for multiple jobs in an automated and personalized way.| | 199|mindsdb/mindsdb !2025-03-282750631|The platform for customizing AI from enterprise data| | 200|DataExpert-io/data-engineer-handbook !2025-03-282721611|This is a repo with links to everything you'd ever want to learn about data engineering| | 201|exo-explore/exo !2025-03-282721633|Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚| | 202|taichi-dev/taichi !2025-03-2826926-1|Productive, portable, and performant GPU programming in Python.| | 203|mem0ai/mem0 !2025-03-282689134|The memory layer for Personalized AI| | 204|svc-develop-team/so-vits-svc !2025-03-28268096 |SoftVC VITS Singing Voice Conversion| | 205|OpenBMB/ChatDev !2025-03-28265624|Create Customized Software using Natural Language Idea (through Multi-Agent Collaboration)| | 206|roboflow/supervision !2025-03-282632010|We write your reusable computer vision tools. 💜| | 207|drawdb-io/drawdb !2025-03-282626913|Free, simple, and intuitive online database design tool and SQL generator.| | 208|karpathy/llm.c !2025-03-28261633|LLM training in simple, raw C/CUDA| | 209|airbnb/lottie-ios !2025-03-28261431|An iOS library to natively render After Effects vector animations| | 210|openai/openai-python !2025-03-282607713|The OpenAI Python library provides convenient access to the OpenAI API from applications written in the Python language.| | 211|academic/awesome-datascience !2025-03-28259876|📝 An awesome Data Science repository to learn and apply for real world problems.| | 212|harry0703/MoneyPrinterTurbo !2025-03-282576618|Generate short videos with one click using a large model| | 213|gabime/spdlog !2025-03-282571511|Fast C++ logging library.| | 214|ocrmypdf/OCRmyPDF !2025-03-2825674217|OCRmyPDF adds an OCR text layer to scanned PDF files, allowing them to be searched| | 215|Vision-CAIR/MiniGPT-4 !2025-03-28256170 |Enhancing Vision-language Understanding with Advanced Large Language Models| | 216|Stability-AI/generative-models !2025-03-28255936|Generative Models by Stability AI| | 217|DS4SD/docling !2025-03-282555662|Get your docs ready for gen AI| | 218|PostHog/posthog !2025-03-282533227|🦔 PostHog provides open-source product analytics, session recording, feature flagging and A/B testing that you can self-host.| | 219|nrwl/nx !2025-03-282509612|Smart Monorepos · Fast CI| | 220|continuedev/continue !2025-03-282500737|⏩ the open-source copilot chat for software development—bring the power of ChatGPT to VS Code| | 221|opentofu/opentofu !2025-03-28247968|OpenTofu lets you declaratively manage your cloud infrastructure.| | 222|invoke-ai/InvokeAI !2025-03-28247293|InvokeAI is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The solution offers an industry leading WebUI, supports terminal use through a CLI, and serves as the foundation for multiple commercial products.| | 223|deepinsight/insightface !2025-03-282471615 |State-of-the-art 2D and 3D Face Analysis Project| | 224|apache/flink !2025-03-28246865|Apache Flink| | 225|ComposioHQ/composio !2025-03-28246436|Composio equips agents with well-crafted tools empowering them to tackle complex tasks| | 226|Genesis-Embodied-AI/Genesis !2025-03-282458314|A generative world for general-purpose robotics & embodied AI learning.| | 227|stretchr/testify !2025-03-28243184|A toolkit with common assertions and mocks that plays nicely with the standard library| | 228| yetone/openai-translator !2025-03-28242921 | Browser extension and cross-platform desktop application for translation based on ChatGPT API | | 229|frappe/erpnext !2025-03-282425211|Free and Open Source Enterprise Resource Planning (ERP)| | 230|songquanpeng/one-api !2025-03-282410034|OpenAI 接口管理 & 分发系统,支持 Azure、Anthropic Claude、Google PaLM 2 & Gemini、智谱 ChatGLM、百度文心一言、讯飞星火认知、阿里通义千问、360 智脑以及腾讯混元,可用于二次分发管理 key,仅单可执行文件,已打包好 Docker 镜像,一键部署,开箱即用. OpenAI key management & redistribution system, using a single API for all LLMs, and features an English UI.| | 231| microsoft/JARVIS !2025-03-28240604 | a system to connect LLMs with ML community | | 232|google/flatbuffers !2025-03-28239965|FlatBuffers: Memory Efficient Serialization Library| | 233|microsoft/graphrag !2025-03-282398928|A modular graph-based Retrieval-Augmented Generation (RAG) system| | 234|rancher/rancher !2025-03-28239675|Complete container management platform| | 235|bazelbuild/bazel !2025-03-282384618|a fast, scalable, multi-language and extensible build system| | 236|modularml/mojo !2025-03-28238236 |The Mojo Programming Language| | 237|danny-avila/LibreChat !2025-03-282378753|Enhanced ChatGPT Clone: Features OpenAI, GPT-4 Vision, Bing, Anthropic, OpenRouter, Google Gemini, AI model switching, message search, langchain, DALL-E-3, ChatGPT Plugins, OpenAI Functions, Secure Multi-User System, Presets, completely open-source for self-hosting. More features in development| |!green-up-arrow.svg 238|🔥🔥🔥Shubhamsaboo/awesome-llm-apps !2025-03-28237391211|Collection of awesome LLM apps with RAG using OpenAI, Anthropic, Gemini and opensource models.| |!red-down-arrow 239|microsoft/semantic-kernel !2025-03-282373611|Integrate cutting-edge LLM technology quickly and easily into your apps| |!red-down-arrow 240|TheAlgorithms/Rust !2025-03-28236995|All Algorithms implemented in Rust| | 241|stanford-oval/storm !2025-03-28236326|An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.| | 242|openai/gpt-2 !2025-03-28232483|Code for the paper "Language Models are Unsupervised Multitask Learners"| | 243|labring/FastGPT !2025-03-282319445|A platform that uses the OpenAI API to quickly build an AI knowledge base, supporting many-to-many relationships.| | 244|pathwaycom/llm-app !2025-03-2822928-10|Ready-to-run cloud templates for RAG, AI pipelines, and enterprise search with live data. 🐳Docker-friendly.⚡Always in sync with Sharepoint, Google Drive, S3, Kafka, PostgreSQL, real-time data APIs, and more.| | 245|warpdotdev/Warp !2025-03-282286825|Warp is a modern, Rust-based terminal with AI built in so you and your team can build great software, faster.| | 246|🔥agno-agi/agno !2025-03-2822833298|Agno is a lightweight library for building Multimodal Agents. It exposes LLMs as a unified API and gives them superpowers like memory, knowledge, tools and reasoning.| | 247|qdrant/qdrant !2025-03-282275214 |Qdrant - Vector Database for the next generation of AI applications. Also available in the cloud https://cloud.qdrant.io/| | 248|ashishpatel26/500-AI-Machine-learning-Deep-learning-Computer-vision-NLP-Projects-with-code !2025-03-282271815|500 AI Machine learning Deep learning Computer vision NLP Projects with code| | 249|stanfordnlp/dspy !2025-03-282268321|Stanford DSPy: The framework for programming—not prompting—foundation models| | 250|PaddlePaddle/Paddle !2025-03-28226246|PArallel Distributed Deep LEarning: Machine Learning Framework from Industrial Practice (『飞桨』核心框架,深度学习&机器学习高性能单机、分布式训练和跨平台部署)| | 251|zulip/zulip !2025-03-28225464|Zulip server and web application. Open-source team chat that helps teams stay productive and focused.| | 252|Hannibal046/Awesome-LLM !2025-03-282240721|Awesome-LLM: a curated list of Large Language Model| | 253|facefusion/facefusion !2025-03-282218812|Next generation face swapper and enhancer| | 254|Mozilla-Ocho/llamafile !2025-03-28220624|Distribute and run LLMs with a single file.| | 255|yuliskov/SmartTube !2025-03-282201614|SmartTube - an advanced player for set-top boxes and tvs running Android OS| | 256|haotian-liu/LLaVA !2025-03-282201316 |Large Language-and-Vision Assistant built towards multimodal GPT-4 level capabilities.| | 257|ashishps1/awesome-system-design-resources !2025-03-282189367|This repository contains System Design resources which are useful while preparing for interviews and learning Distributed Systems| | 258|Cinnamon/kotaemon !2025-03-28218248|An open-source RAG-based tool for chatting with your documents.| | 259|CodePhiliaX/Chat2DB !2025-03-282179757|🔥🔥🔥AI-driven database tool and SQL client, The hottest GUI client, supporting MySQL, Oracle, PostgreSQL, DB2, SQL Server, DB2, SQLite, H2, ClickHouse, and more.| | 260|blakeblackshear/frigate !2025-03-282177113|NVR with realtime local object detection for IP cameras| | 261|facebookresearch/audiocraft !2025-03-28217111|Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable music generation LM with textual and melodic conditioning.| | 262|karpathy/minGPT !2025-03-28216567|A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training| | 263|grpc/grpc-go !2025-03-282159510|The Go language implementation of gRPC. HTTP/2 based RPC| | 264|HumanSignal/label-studio !2025-03-282137618|Label Studio is a multi-type data labeling and annotation tool with standardized output format| | 265|yoheinakajima/babyagi !2025-03-28212764 |uses OpenAI and Pinecone APIs to create, prioritize, and execute tasks, This is a pared-down version of the original Task-Driven Autonomous Agent| | 266|deepseek-ai/DeepSeek-Coder !2025-03-282118210|DeepSeek Coder: Let the Code Write Itself| | 267|BuilderIO/gpt-crawler !2025-03-282118010|Crawl a site to generate knowledge files to create your own custom GPT from a URL| | 268| openai/chatgpt-retrieval-plugin !2025-03-2821152-1 | Plugins are chat extensions designed specifically for language models like ChatGPT, enabling them to access up-to-date information, run computations, or interact with third-party services in response to a user's request.| | 269|microsoft/OmniParser !2025-03-282113123|A simple screen parsing tool towards pure vision based GUI agent| | 270|black-forest-labs/flux !2025-03-282107219|Official inference repo for FLUX.1 models| | 271|ItzCrazyKns/Perplexica !2025-03-282099154|Perplexica is an AI-powered search engine. It is an Open source alternative to Perplexity AI| | 272|microsoft/unilm !2025-03-28209876|Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities| | 273|Sanster/lama-cleaner !2025-03-282077614|Image inpainting tool powered by SOTA AI Model. Remove any unwanted object, defect, people from your pictures or erase and replace(powered by stable diffusion) any thing on your pictures.| | 274|assafelovic/gpt-researcher !2025-03-282057222|GPT based autonomous agent that does online comprehensive research on any given topic| | 275|PromtEngineer/localGPT !2025-03-28204230 |Chat with your documents on your local device using GPT models. No data leaves your device and 100% private.| | 276|elastic/kibana !2025-03-28203482|Your window into the Elastic Stack| | 277|fishaudio/fish-speech !2025-03-282033222|Brand new TTS solution| | 278|mlc-ai/mlc-llm !2025-03-282028110 |Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.| | 279|deepset-ai/haystack !2025-03-282005320|🔍 Haystack is an open source NLP framework to interact with your data using Transformer models and LLMs (GPT-4, ChatGPT and alike). Haystack offers production-ready tools to quickly build complex question answering, semantic search, text generation applications, and more.| | 280|tree-sitter/tree-sitter !2025-03-28200487|An incremental parsing system for programming tools| | 281|Anjok07/ultimatevocalremovergui !2025-03-281999811|GUI for a Vocal Remover that uses Deep Neural Networks.| | 282|guidance-ai/guidance !2025-03-28199622|A guidance language for controlling large language models.| | 283|ml-explore/mlx !2025-03-28199619|MLX: An array framework for Apple silicon| | 284|mlflow/mlflow !2025-03-281995314|Open source platform for the machine learning lifecycle| | 285|ml-tooling/best-of-ml-python !2025-03-28198631|🏆 A ranked list of awesome machine learning Python libraries. Updated weekly.| | 286|BerriAI/litellm !2025-03-281981862|Call all LLM APIs using the OpenAI format. Use Bedrock, Azure, OpenAI, Cohere, Anthropic, Ollama, Sagemaker, HuggingFace, Replicate (100+ LLMs)| | 287|LazyVim/LazyVim !2025-03-281981320|Neovim config for the lazy| | 288|wez/wezterm !2025-03-281976018|A GPU-accelerated cross-platform terminal emulator and multiplexer written by @wez and implemented in Rust| | 289|valkey-io/valkey !2025-03-281970416|A flexible distributed key-value datastore that supports both caching and beyond caching workloads.| | 290|LiLittleCat/awesome-free-chatgpt !2025-03-28196185|🆓免费的 ChatGPT 镜像网站列表,持续更新。List of free ChatGPT mirror sites, continuously updated.| | 291|Byaidu/PDFMathTranslate !2025-03-281947645|PDF scientific paper translation with preserved formats - 基于 AI 完整保留排版的 PDF 文档全文双语翻译,支持 Google/DeepL/Ollama/OpenAI 等服务,提供 CLI/GUI/Docker| | 292|openai/swarm !2025-03-281947111|Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.| | 293|HqWu-HITCS/Awesome-Chinese-LLM !2025-03-281921423|Organizing smaller, cost-effective, privately deployable open-source Chinese language models, including related datasets and tutorials| | 294|stitionai/devika !2025-03-28190903|Devika is an Agentic AI Software Engineer that can understand high-level human instructions, break them down into steps, research relevant information, and write code to achieve the given objective. Devika aims to be a competitive open-source alternative to Devin by Cognition AI.| | 295|OpenBMB/MiniCPM-o !2025-03-28190887|MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone| | 296|samber/lo !2025-03-281904815|💥 A Lodash-style Go library based on Go 1.18+ Generics (map, filter, contains, find...)| | 297|chroma-core/chroma !2025-03-281895221 |the AI-native open-source embedding database| | 298|DarkFlippers/unleashed-firmware !2025-03-28189278|Flipper Zero Unleashed Firmware| | 299|brave/brave-browser !2025-03-281892710|Brave browser for Android, iOS, Linux, macOS, Windows.| | 300| tloen/alpaca-lora !2025-03-28188641 | Instruct-tune LLaMA on consumer hardware| | 301|VinciGit00/Scrapegraph-ai !2025-03-281884618|Python scraper based on AI| | 302|gitroomhq/postiz-app !2025-03-281879110|📨 Schedule social posts, measure them, exchange with other members and get a lot of help from AI 🚀| | 303|PrefectHQ/prefect !2025-03-281878715|Prefect is a workflow orchestration tool empowering developers to build, observe, and react to data pipelines| | 304|ymcui/Chinese-LLaMA-Alpaca !2025-03-28187723 |Chinese LLaMA & Alpaca LLMs| | 305|kenjihiranabe/The-Art-of-Linear-Algebra !2025-03-28187335|Graphic notes on Gilbert Strang's "Linear Algebra for Everyone"| | 306|joonspk-research/generativeagents !2025-03-28187288|Generative Agents: Interactive Simulacra of Human Behavior| | 307|renovatebot/renovate !2025-03-28186820|Universal dependency update tool that fits into your workflows.| | 308|gventuri/pandas-ai !2025-03-28186109 |Pandas AI is a Python library that integrates generative artificial intelligence capabilities into Pandas, making dataframes conversational| | 309|thingsboard/thingsboard !2025-03-28185184|Open-source IoT Platform - Device management, data collection, processing and visualization.| | 310|ente-io/ente !2025-03-28184722|Fully open source, End to End Encrypted alternative to Google Photos and Apple Photos| | 311|serengil/deepface !2025-03-281840113|A Lightweight Face Recognition and Facial Attribute Analysis (Age, Gender, Emotion and Race) Library for Python| | 312|Raphire/Win11Debloat !2025-03-281840132|A simple, easy to use PowerShell script to remove pre-installed apps from windows, disable telemetry, remove Bing from windows search as well as perform various other changes to declutter and improve your windows experience. This script works for both windows 10 and windows 11.| | 313|Avaiga/taipy !2025-03-28179235|Turns Data and AI algorithms into production-ready web applications in no time.| | 314|microsoft/qlib !2025-03-281784231|Qlib is an AI-oriented quantitative investment platform that aims to realize the potential, empower research, and create value using AI technologies in quantitative investment, from exploring ideas to implementing productions. Qlib supports diverse machine learning modeling paradigms. including supervised learning, market dynamics modeling, and RL.| | 315|CopilotKit/CopilotKit !2025-03-281778571|Build in-app AI chatbots 🤖, and AI-powered Textareas ✨, into react web apps.| | 316|QwenLM/Qwen-7B !2025-03-281766017|The official repo of Qwen-7B (通义千问-7B) chat & pretrained large language model proposed by Alibaba Cloud.| | 317|w-okada/voice-changer !2025-03-28176078 |リアルタイムボイスチェンジャー Realtime Voice Changer| | 318|rlabbe/Kalman-and-Bayesian-Filters-in-Python !2025-03-281756011|Kalman Filter book using Jupyter Notebook. Focuses on building intuition and experience, not formal proofs. Includes Kalman filters,extended Kalman filters, unscented Kalman filters, particle filters, and more. All exercises include solutions.| | 319|Mikubill/sd-webui-controlnet !2025-03-28174794 |WebUI extension for ControlNet| | 320|jingyaogong/minimind !2025-03-2817380116|「大模型」3小时完全从0训练26M的小参数GPT,个人显卡即可推理训练!| | 321|apify/crawlee !2025-03-28172696|Crawlee—A web scraping and browser automation library for Node.js to build reliable crawlers. In JavaScript and TypeScript. Extract data for AI, LLMs, RAG, or GPTs. Download HTML, PDF, JPG, PNG, and other files from websites. Works with Puppeteer, Playwright, Cheerio, JSDOM, and raw HTTP. Both headful and headless mode. With proxy rotation.| | 322|apple/ml-stable-diffusion !2025-03-28172395|Stable Diffusion with Core ML on Apple Silicon| | 323| transitive-bullshit/chatgpt-api !2025-03-28172095 | Node.js client for the official ChatGPT API. | | 324|teableio/teable !2025-03-281719222|✨ The Next Gen Airtable Alternative: No-Code Postgres| | 325| xx025/carrot !2025-03-28170900 | Free ChatGPT Site List | | 326|microsoft/LightGBM !2025-03-28170723|A fast, distributed, high-performance gradient boosting (GBT, GBDT, GBRT, GBM or MART) framework based on decision tree algorithms, used for ranking, classification and many other machine learning tasks.| | 327|VikParuchuri/surya !2025-03-28169827|Accurate line-level text detection and recognition (OCR) in any language| | 328|deepseek-ai/Janus !2025-03-281692825|Janus-Series: Unified Multimodal Understanding and Generation Models| | 329|ardalis/CleanArchitecture !2025-03-28168823|Clean Architecture Solution Template: A starting point for Clean Architecture with ASP.NET Core| | 330|neondatabase/neon !2025-03-28166466|Neon: Serverless Postgres. We separated storage and compute to offer autoscaling, code-like database branching, and scale to zero.| | 331|kestra-io/kestra !2025-03-281661313|⚡ Workflow Automation Platform. Orchestrate & Schedule code in any language, run anywhere, 500+ plugins. Alternative to Zapier, Rundeck, Camunda, Airflow...| | 332|Dao-AILab/flash-attention !2025-03-281659720|Fast and memory-efficient exact attention| | 333|RPCS3/rpcs3 !2025-03-281655712|PS3 emulator/debugger| | 334|meta-llama/llama-recipes !2025-03-28165486|Scripts for fine-tuning Llama2 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization & question answering. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment.Demo apps to showcase Llama2 for WhatsApp & Messenger| | 335|emilwallner/Screenshot-to-code !2025-03-28165180|A neural network that transforms a design mock-up into a static website.| | 336|datawhalechina/llm-cookbook !2025-03-281650922|面向开发者的 LLM 入门教程,吴恩达大模型系列课程中文版| | 337|e2b-dev/awesome-ai-agents !2025-03-281643923|A list of AI autonomous agents| | 338|QwenLM/Qwen2.5 !2025-03-281641114|Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.| | 339|dair-ai/ML-YouTube-Courses !2025-03-28164114|📺 Discover the latest machine learning / AI courses on YouTube.| | 340|pybind/pybind11 !2025-03-28163620|Seamless operability between C++11 and Python| | 341|graphdeco-inria/gaussian-splatting !2025-03-281627116|Original reference implementation of "3D Gaussian Splatting for Real-Time Radiance Field Rendering"| | 342|meta-llama/codellama !2025-03-28162531|Inference code for CodeLlama models| | 343|TransformerOptimus/SuperAGI !2025-03-28161292 | SuperAGI - A dev-first open source autonomous AI agent framework. Enabling developers to build, manage & run useful autonomous agents quickly and reliably.| | 344|microsoft/onnxruntime !2025-03-28161169|ONNX Runtime: cross-platform, high-performance ML inferencing and training accelerator| | 345|IDEA-Research/Grounded-Segment-Anything !2025-03-281601411 |Marrying Grounding DINO with Segment Anything & Stable Diffusion & BLIP - Automatically Detect, Segment and Generate Anything with Image and Text Inputs| | 346|ddbourgin/numpy-ml !2025-03-28160054|Machine learning, in numpy| | 347|eosphoros-ai/DB-GPT !2025-03-281585225|Revolutionizing Database Interactions with Private LLM Technology| | 348|Stability-AI/StableLM !2025-03-28158310 |Stability AI Language Models| | 349|openai/evals !2025-03-28157935 |Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.| | 350|THUDM/ChatGLM2-6B !2025-03-28157500|ChatGLM2-6B: An Open Bilingual Chat LLM | | 351|sunner/ChatALL !2025-03-28156761 |Concurrently chat with ChatGPT, Bing Chat, Bard, Alpaca, Vincuna, Claude, ChatGLM, MOSS, iFlytek Spark, ERNIE and more, discover the best answers| | 352|abseil/abseil-cpp !2025-03-28156656|Abseil Common Libraries (C++)| | 353|NVIDIA/open-gpu-kernel-modules !2025-03-28156531|NVIDIA Linux open GPU kernel module source| | 354|letta-ai/letta !2025-03-281563718|Letta (formerly MemGPT) is a framework for creating LLM services with memory.| | 355|typescript-eslint/typescript-eslint !2025-03-28156211|✨ Monorepo for all the tooling which enables ESLint to support TypeScript| | 356|umijs/umi !2025-03-28156211|A framework in react community ✨| | 357|AI4Finance-Foundation/FinGPT !2025-03-281561215|Data-Centric FinGPT. Open-source for open finance! Revolutionize 🔥 We'll soon release the trained model.| | 358|amplication/amplication !2025-03-28156022|🔥🔥🔥 The Only Production-Ready AI-Powered Backend Code Generation| | 359|KindXiaoming/pykan !2025-03-28155477|Kolmogorov Arnold Networks| | 360|arc53/DocsGPT !2025-03-28154900|GPT-powered chat for documentation, chat with your documents| | 361|influxdata/telegraf !2025-03-28154502|Agent for collecting, processing, aggregating, and writing metrics, logs, and other arbitrary data.| | 362|microsoft/Bringing-Old-Photos-Back-to-Life !2025-03-28154084|Bringing Old Photo Back to Life (CVPR 2020 oral)| | 363|GaiZhenbiao/ChuanhuChatGPT !2025-03-2815394-2|GUI for ChatGPT API and many LLMs. Supports agents, file-based QA, GPT finetuning and query with web search. All with a neat UI.| | 364|Zeyi-Lin/HivisionIDPhotos !2025-03-281529710|⚡️HivisionIDPhotos: a lightweight and efficient AI ID photos tools. 一个轻量级的AI证件照制作算法。| | 365| mayooear/gpt4-pdf-chatbot-langchain !2025-03-281529518 | GPT4 & LangChain Chatbot for large PDF docs | | 366|1Panel-dev/MaxKB !2025-03-2815277148|? Based on LLM large language model knowledge base Q&A system. Ready to use out of the box, supports quick integration into third-party business systems. Officially produced by 1Panel| | 367|ai16z/eliza !2025-03-281526811|Conversational Agent for Twitter and Discord| | 368|apache/arrow !2025-03-28151684|Apache Arrow is a multi-language toolbox for accelerated data interchange and in-memory processing| | 369|princeton-nlp/SWE-agent !2025-03-281516119|SWE-agent: Agent Computer Interfaces Enable Software Engineering Language Models| | 370|mlc-ai/web-llm !2025-03-281509311 |Bringing large-language models and chat to web browsers. Everything runs inside the browser with no server support.| | 371|guillaumekln/faster-whisper !2025-03-281507117 |Faster Whisper transcription with CTranslate2| | 372|overleaf/overleaf !2025-03-28150316|A web-based collaborative LaTeX editor| | 373|triton-lang/triton !2025-03-28150169|Development repository for the Triton language and compiler| | 374|soxoj/maigret !2025-03-281500410|🕵️‍♂️ Collect a dossier on a person by username from thousands of sites| | 375|alibaba/lowcode-engine !2025-03-28149841|An enterprise-class low-code technology stack with scale-out design / 一套面向扩展设计的企业级低代码技术体系| | 376|espressif/esp-idf !2025-03-28148545|Espressif IoT Development Framework. Official development framework for Espressif SoCs.| | 377|pgvector/pgvector !2025-03-281484913|Open-source vector similarity search for Postgres| | 378|datawhalechina/leedl-tutorial !2025-03-28148246|《李宏毅深度学习教程》(李宏毅老师推荐👍),PDF下载地址:https://github.com/datawhalechina/leedl-tutorial/releases| | 379|xcanwin/KeepChatGPT !2025-03-28147972 |Using ChatGPT is more efficient and smoother, perfectly solving ChatGPT network errors. No longer do you need to frequently refresh the webpage, saving over 10 unnecessary steps| | 380|m-bain/whisperX !2025-03-281471313|WhisperX: Automatic Speech Recognition with Word-level Timestamps (& Diarization)| | 381|HumanAIGC/AnimateAnyone !2025-03-2814706-1|Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation| |!green-up-arrow.svg 382|naklecha/llama3-from-scratch !2025-03-281469024|llama3 implementation one matrix multiplication at a time| |!red-down-arrow 383| fauxpilot/fauxpilot !2025-03-28146871 | An open-source GitHub Copilot server | | 384|LlamaFamily/Llama-Chinese !2025-03-28145111|Llama Chinese Community, the best Chinese Llama large model, fully open source and commercially available| | 385|BradyFU/Awesome-Multimodal-Large-Language-Models !2025-03-281450121|Latest Papers and Datasets on Multimodal Large Language Models| | 386|vanna-ai/vanna !2025-03-281449819|🤖 Chat with your SQL database 📊. Accurate Text-to-SQL Generation via LLMs using RAG 🔄.| | 387|bleedline/aimoneyhunter !2025-03-28144845|AI Side Hustle Money Mega Collection: Teaching You How to Utilize AI for Various Side Projects to Earn Extra Income.| | 388|stefan-jansen/machine-learning-for-trading !2025-03-28144629|Code for Machine Learning for Algorithmic Trading, 2nd edition.| | 389|state-spaces/mamba !2025-03-28144139|Mamba: Linear-Time Sequence Modeling with Selective State Spaces| | 390|vercel/ai-chatbot !2025-03-281434614|A full-featured, hackable Next.js AI chatbot built by Vercel| | 391|steven-tey/novel !2025-03-281428410|Notion-style WYSIWYG editor with AI-powered autocompletions| | 392|unifyai/ivy !2025-03-281409348|Unified AI| | 393|chidiwilliams/buzz !2025-03-281402411 |Buzz transcribes and translates audio offline on your personal computer. Powered by OpenAI's Whisper.| | 394|lukas-blecher/LaTeX-OCR !2025-03-28139769|pix2tex: Using a ViT to convert images of equations into LaTeX code.| | 395|openai/tiktoken !2025-03-28139599|tiktoken is a fast BPE tokeniser for use with OpenAI's models.| | 396|nocobase/nocobase !2025-03-281391522|NocoBase is a scalability-first, open-source no-code/low-code platform for building business applications and enterprise solutions.| | 397|neonbjb/tortoise-tts !2025-03-28139010 |A multi-voice TTS system trained with an emphasis on quality| | 398|yamadashy/repomix !2025-03-281382036|📦 Repomix (formerly Repopack) is a powerful tool that packs your entire repository into a single, AI-friendly file. Perfect for when you need to feed your codebase to Large Language Models (LLMs) or other AI tools like Claude, ChatGPT, and Gemini.| | 399|adobe/react-spectrum !2025-03-28136766|A collection of libraries and tools that help you build adaptive, accessible, and robust user experiences.| | 400|THUDM/ChatGLM3 !2025-03-28136684|ChatGLM3 series: Open Bilingual Chat LLMs | | 401|NVIDIA/NeMo !2025-03-28134837|A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)| | 402|BlinkDL/RWKV-LM !2025-03-28134346 |RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it combines the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.| | 403| fuergaosi233/wechat-chatgpt !2025-03-28133330 | Use ChatGPT On Wechat via wechaty | | 404|udecode/plate !2025-03-28133325|A rich-text editor powered by AI| | 405|xenova/transformers.js !2025-03-281331219|State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!| | 406|stas00/ml-engineering !2025-03-281325615|Machine Learning Engineering Guides and Tools| | 407| wong2/chatgpt-google-extension !2025-03-2813241-1 | A browser extension that enhances search engines with ChatGPT, this repos will not be updated from 2023-02-20| | 408|mrdbourke/pytorch-deep-learning !2025-03-281317520|Materials for the Learn PyTorch for Deep Learning: Zero to Mastery course.| | 409|Koenkk/zigbee2mqtt !2025-03-28131544|Zigbee 🐝 to MQTT bridge 🌉, get rid of your proprietary Zigbee bridges 🔨| | 410|vercel-labs/ai !2025-03-281298528|Build AI-powered applications with React, Svelte, and Vue| | 411|netease-youdao/QAnything !2025-03-28129318|Question and Answer based on Anything.| | 412|huggingface/trl !2025-03-281289622|Train transformer language models with reinforcement learning.| | 413|microsoft/BitNet !2025-03-28128503|Official inference framework for 1-bit LLMs| | 414|mediar-ai/screenpipe !2025-03-281283915|24/7 local AI screen & mic recording. Build AI apps that have the full context. Works with Ollama. Alternative to Rewind.ai. Open. Secure. You own your data. Rust.| | 415|Skyvern-AI/skyvern !2025-03-281277612|Automate browser-based workflows with LLMs and Computer Vision| | 416|pytube/pytube !2025-03-28126591|A lightweight, dependency-free Python library (and command-line utility) for downloading YouTube Videos.| | 417|official-stockfish/Stockfish !2025-03-28126574|UCI chess engine| | 418|sgl-project/sglang !2025-03-281260143|SGLang is a structured generation language designed for large language models (LLMs). It makes your interaction with LLMs faster and more controllable.| | 419|plasma-umass/scalene !2025-03-28125535|Scalene: a high-performance, high-precision CPU, GPU, and memory profiler for Python with AI-powered optimization proposals| | 420|danswer-ai/danswer !2025-03-28125503|Ask Questions in natural language and get Answers backed by private sources. Connects to tools like Slack, GitHub, Confluence, etc.| | 421|OpenTalker/SadTalker !2025-03-28125226|[CVPR 2023] SadTalker:Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation| | 422|facebookresearch/AnimatedDrawings !2025-03-28123693 |Code to accompany "A Method for Animating Children's Drawings of the Human Figure"| | 423|activepieces/activepieces !2025-03-28123609|Your friendliest open source all-in-one automation tool ✨ Workflow automation tool 100+ integration / Enterprise automation tool / Zapier Alternative| | 424|ggerganov/ggml !2025-03-28121992 |Tensor library for machine learning| | 425|bytebase/bytebase !2025-03-28121694|World's most advanced database DevOps and CI/CD for Developer, DBA and Platform Engineering teams. The GitLab/GitHub for database DevOps.| | 426| willwulfken/MidJourney-Styles-and-Keywords-Reference !2025-03-28120971 | A reference containing Styles and Keywords that you can use with MidJourney AI| | 427|Huanshere/VideoLingo !2025-03-281207013|Netflix-level subtitle cutting, translation, alignment, and even dubbing - one-click fully automated AI video subtitle team | | 428|OpenLMLab/MOSS !2025-03-28120330 |An open-source tool-augmented conversational language model from Fudan University| | 429|llmware-ai/llmware !2025-03-281200727|Providing enterprise-grade LLM-based development framework, tools, and fine-tuned models.| | 430|PKU-YuanGroup/Open-Sora-Plan !2025-03-28119362|This project aim to reproduce Sora (Open AI T2V model), but we only have limited resource. We deeply wish the all open source community can contribute to this project.| | 431|ShishirPatil/gorilla !2025-03-28119332 |Gorilla: An API store for LLMs| | 432|NVIDIA/Megatron-LM !2025-03-281192716|Ongoing research training transformer models at scale| | 433|illacloud/illa-builder !2025-03-28119192|Create AI-Driven Apps like Assembling Blocks| | 434|marimo-team/marimo !2025-03-281191521|A reactive notebook for Python — run reproducible experiments, execute as a script, deploy as an app, and version with git.| | 435|smol-ai/developer !2025-03-28119111 | With 100k context windows on the way, it's now feasible for every dev to have their own smol developer| | 436|Lightning-AI/litgpt !2025-03-28118878|Pretrain, finetune, deploy 20+ LLMs on your own data. Uses state-of-the-art techniques: flash attention, FSDP, 4-bit, LoRA, and more.| | 437|openai/shap-e !2025-03-28118474 |Generate 3D objects conditioned on text or images| | 438|eugeneyan/open-llms !2025-03-28118451 |A list of open LLMs available for commercial use.| | 439|andrewyng/aisuite !2025-03-28118124|Simple, unified interface to multiple Generative AI providers| | 440|hajimehoshi/ebiten !2025-03-28117816|Ebitengine - A dead simple 2D game engine for Go| | 441|kgrzybek/modular-monolith-with-ddd !2025-03-28117493|Full Modular Monolith application with Domain-Driven Design approach.| | 442|h2oai/h2ogpt !2025-03-2811736-1 |Come join the movement to make the world's best open source GPT led by H2O.ai - 100% private chat and document search, no data leaks, Apache 2.0| | 443|owainlewis/awesome-artificial-intelligence !2025-03-28117332|A curated list of Artificial Intelligence (AI) courses, books, video lectures and papers.| | 444|DataTalksClub/mlops-zoomcamp !2025-03-28116643|Free MLOps course from DataTalks.Club| | 445|Rudrabha/Wav2Lip !2025-03-281163410|This repository contains the codes of "A Lip Sync Expert Is All You Need for Speech to Lip Generation In the Wild", published at ACM Multimedia 2020.| | 446|aishwaryanr/awesome-generative-ai-guide !2025-03-281152810|A one stop repository for generative AI research updates, interview resources, notebooks and much more!| | 447|karpathy/micrograd !2025-03-28115146|A tiny scalar-valued autograd engine and a neural net library on top of it with PyTorch-like API| | 448|InstantID/InstantID !2025-03-28115111|InstantID : Zero-shot Identity-Preserving Generation in Seconds 🔥| | 449|facebookresearch/seamlesscommunication !2025-03-28114434|Foundational Models for State-of-the-Art Speech and Text Translation| | 450|anthropics/anthropic-cookbook !2025-03-281140112|A collection of notebooks/recipes showcasing some fun and effective ways of using Claude.| | 451|mastra-ai/mastra !2025-03-281139240|the TypeScript AI agent framework| | 452|NVIDIA/TensorRT !2025-03-28113864|NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.| | 453|plandex-ai/plandex !2025-03-28113645|An AI coding engine for complex tasks| | 454|RUCAIBox/LLMSurvey !2025-03-28112735 |A collection of papers and resources related to Large Language Models.| | 455|kubeshark/kubeshark !2025-03-28112711|The API traffic analyzer for Kubernetes providing real-time K8s protocol-level visibility, capturing and monitoring all traffic and payloads going in, out and across containers, pods, nodes and clusters. Inspired by Wireshark, purposely built for Kubernetes| | 456|electric-sql/pglite !2025-03-28112617|Lightweight Postgres packaged as WASM into a TypeScript library for the browser, Node.js, Bun and Deno from https://electric-sql.com| | 457|lightaime/camel !2025-03-281124441 |🐫 CAMEL: Communicative Agents for “Mind” Exploration of Large Scale Language Model Society| | 458|huggingface/lerobot !2025-03-281120184|🤗 LeRobot: State-of-the-art Machine Learning for Real-World Robotics in Pytorch| | 459|normal-computing/outlines !2025-03-28111657|Generative Model Programming| | 460|libretro/RetroArch !2025-03-28110701|Cross-platform, sophisticated frontend for the libretro API. Licensed GPLv3.| | 461|THUDM/CogVideo !2025-03-28110599|Text-to-video generation: CogVideoX (2024) and CogVideo (ICLR 2023)| | 462|bentoml/OpenLLM !2025-03-28110495|An open platform for operating large language models (LLMs) in production. Fine-tune, serve, deploy, and monitor any LLMs with ease.| | 463|vosen/ZLUDA !2025-03-28110429|CUDA on AMD GPUs| | 464|dair-ai/ML-Papers-of-the-Week !2025-03-28110304 |🔥Highlighting the top ML papers every week.| | 465|WordPress/gutenberg !2025-03-28110212|The Block Editor project for WordPress and beyond. Plugin is available from the official repository.| | 466|microsoft/data-formulator !2025-03-281099827|🪄 Create rich visualizations with AI| | 467|LibreTranslate/LibreTranslate !2025-03-28109887|Free and Open Source Machine Translation API. Self-hosted, offline capable and easy to setup.| | 468|block/goose !2025-03-281097737|an open-source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM| | 469|getumbrel/llama-gpt !2025-03-28109553|A self-hosted, offline, ChatGPT-like chatbot. Powered by Llama 2. 100% private, with no data leaving your device.| | 470|HigherOrderCO/HVM !2025-03-28109182|A massively parallel, optimal functional runtime in Rust| | 471|databrickslabs/dolly !2025-03-2810812-3 | A large language model trained on the Databricks Machine Learning Platform| | 472|srush/GPU-Puzzles !2025-03-28108014|Solve puzzles. Learn CUDA.| | 473|Z3Prover/z3 !2025-03-28107952|The Z3 Theorem Prover| | 474|UFund-Me/Qbot !2025-03-281079313 |Qbot is an AI-oriented quantitative investment platform, which aims to realize the potential, empower AI technologies in quantitative investment| | 475|langchain-ai/langgraph !2025-03-281077336|| | 476|lz4/lz4 !2025-03-28107647|Extremely Fast Compression algorithm| | 477|magic-research/magic-animate !2025-03-28107160|MagicAnimate: Temporally Consistent Human Image Animation using Diffusion Model| | 478|PaperMC/Paper !2025-03-281071410|The most widely used, high performance Minecraft server that aims to fix gameplay and mechanics inconsistencies| | 479|getomni-ai/zerox !2025-03-281071015|Zero shot pdf OCR with gpt-4o-mini| |!green-up-arrow.svg 480|🔥NirDiamant/GenAIAgents !2025-03-2810693318|This repository provides tutorials and implementations for various Generative AI Agent techniques, from basic to advanced. It serves as a comprehensive guide for building intelligent, interactive AI systems.| |!red-down-arrow 481|Unstructured-IO/unstructured !2025-03-28106889|Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.| | 482|apache/thrift !2025-03-28106610|Apache Thrift| | 483| TheR1D/shellgpt !2025-03-28106097 | A command-line productivity tool powered by ChatGPT, will help you accomplish your tasks faster and more efficiently | | 484|TheRamU/Fay !2025-03-281060312 |Fay is a complete open source project that includes Fay controller and numeral models, which can be used in different applications such as virtual hosts, live promotion, numeral human interaction and so on| | 485|zyronon/douyin !2025-03-28105566|Vue3 + Pinia + Vite5 仿抖音,Vue 在移动端的最佳实践 . Imitate TikTok ,Vue Best practices on Mobile| | 486|THU-MIG/yolov10 !2025-03-28105485|YOLOv10: Real-Time End-to-End Object Detection| | 487|idootop/mi-gpt !2025-03-281052522|? Transform XiaoAi speaker into a personal voice assistant with ChatGPT and DouBao integration.| | 488|SakanaAI/AI-Scientist !2025-03-281051310|The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery 🧑‍🔬| | 489|szimek/sharedrop !2025-03-28105101|Easy P2P file transfer powered by WebRTC - inspired by Apple AirDrop| | 490|salesforce/LAVIS !2025-03-28103942 |LAVIS - A One-stop Library for Language-Vision Intelligence| | 491|aws/amazon-sagemaker-examples !2025-03-28103654|Example 📓 Jupyter notebooks that demonstrate how to build, train, and deploy machine learning models using 🧠 Amazon SageMaker.| | 492|artidoro/qlora !2025-03-28103402 |QLoRA: Efficient Finetuning of Quantized LLMs| | 493|lllyasviel/stable-diffusion-webui-forge !2025-03-281029314| a platform on top of Stable Diffusion WebUI (based on Gradio) to make development easier, optimize resource management, and speed up inference| | 494|NielsRogge/Transformers-Tutorials !2025-03-28102487|This repository contains demos I made with the Transformers library by HuggingFace.| | 495|kedro-org/kedro !2025-03-28102371|Kedro is a toolbox for production-ready data science. It uses software engineering best practices to help you create data engineering and data science pipelines that are reproducible, maintainable, and modular.| | 496| chathub-dev/chathub !2025-03-28102301 | All-in-one chatbot client | | 497|microsoft/promptflow !2025-03-28101612|Build high-quality LLM apps - from prototyping, testing to production deployment and monitoring.| | 498|mistralai/mistral-src !2025-03-28101372|Reference implementation of Mistral AI 7B v0.1 model.| | 499|burn-rs/burn !2025-03-28101183|Burn - A Flexible and Comprehensive Deep Learning Framework in Rust| | 500|AIGC-Audio/AudioGPT !2025-03-28101150 |AudioGPT: Understanding and Generating Speech, Music, Sound, and Talking Head| | 501|facebookresearch/dinov2 !2025-03-281011210 |PyTorch code and models for the DINOv2 self-supervised learning method.| | 502|RockChinQ/LangBot !2025-03-281008455|😎丰富生态、🧩支持扩展、🦄多模态 - 大模型原生即时通信机器人平台 🤖 | | 503|78/xiaozhi-esp32 !2025-03-281008180|Build your own AI friend| | 504|cumulo-autumn/StreamDiffusion !2025-03-28100761|StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation| | 505|DataTalksClub/machine-learning-zoomcamp !2025-03-28100664|The code from the Machine Learning Bookcamp book and a free course based on the book| | 506|nerfstudio-project/nerfstudio !2025-03-28100343|A collaboration friendly studio for NeRFs| | 507|cupy/cupy !2025-03-28100344|NumPy & SciPy for GPU| | 508|NVIDIA/TensorRT-LLM !2025-03-281000823|TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.| | 509|wasp-lang/open-saas !2025-03-2899665|A free, open-source SaaS app starter for React & Node.js with superpowers. Production-ready. Community-driven.| | 510|huggingface/text-generation-inference !2025-03-2899383|Large Language Model Text Generation Inference| | 511|jxnl/instructor !2025-03-2899224|structured outputs for llms| | 512|GoogleCloudPlatform/generative-ai !2025-03-2899086|Sample code and notebooks for Generative AI on Google Cloud| | 513|manticoresoftware/manticoresearch !2025-03-2898799|Easy to use open source fast database for search | | 514|langfuse/langfuse !2025-03-28985134|🪢 Open source LLM engineering platform. Observability, metrics, evals, prompt management, testing, prompt playground, datasets, LLM evaluations -- 🍊YC W23 🤖 integrate via Typescript, Python / Decorators, OpenAI, Langchain, LlamaIndex, Litellm, Instructor, Mistral, Perplexity, Claude, Gemini, Vertex| | 515|keephq/keep !2025-03-2897949|The open-source alert management and AIOps platform| | 516|sashabaranov/go-openai !2025-03-2897843|OpenAI ChatGPT, GPT-3, GPT-4, DALL·E, Whisper API wrapper for Go| | 517|autowarefoundation/autoware !2025-03-2897766|Autoware - the world's leading open-source software project for autonomous driving| | 518|anthropics/courses !2025-03-2897269|Anthropic's educational courses| | 519|popcorn-official/popcorn-desktop !2025-03-2896853|Popcorn Time is a multi-platform, free software BitTorrent client that includes an integrated media player ( Windows / Mac / Linux ) A Butter-Project Fork| | 520|getmaxun/maxun !2025-03-28968515|🔥 Open-source no-code web data extraction platform. Turn websites to APIs and spreadsheets with no-code robots in minutes! [In Beta]| | 521|wandb/wandb !2025-03-2896763|🔥 A tool for visualizing and tracking your machine learning experiments. This repo contains the CLI and Python API.| | 522|karpathy/minbpe !2025-03-2895353|Minimal, clean, code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.| | 523|bigscience-workshop/petals !2025-03-2895142|🌸 Run large language models at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading| | 524|OthersideAI/self-operating-computer !2025-03-2894931|A framework to enable multimodal models to operate a computer.| | 525|mshumer/gpt-prompt-engineer !2025-03-2894911|| | 526| BloopAI/bloop !2025-03-2894710 | A fast code search engine written in Rust| | 527|BlinkDL/ChatRWKV !2025-03-289467-1 |ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.| | 528|timlrx/tailwind-nextjs-starter-blog !2025-03-2894677|This is a Next.js, Tailwind CSS blogging starter template. Comes out of the box configured with the latest technologies to make technical writing a breeze. Easily configurable and customizable. Perfect as a replacement to existing Jekyll and Hugo individual blogs.| | 529|google/benchmark !2025-03-2893634|A microbenchmark support library| | 530|facebookresearch/nougat !2025-03-2893603|Implementation of Nougat Neural Optical Understanding for Academic Documents| | 531|modelscope/facechain !2025-03-2893536|FaceChain is a deep-learning toolchain for generating your Digital-Twin.| | 532|DrewThomasson/ebook2audiobook !2025-03-2893388|Convert ebooks to audiobooks with chapters and metadata using dynamic AI models and voice cloning. Supports 1,107+ languages!| | 533|RayTracing/raytracing.github.io !2025-03-2893035|Main Web Site (Online Books)| | 534|QwenLM/Qwen2.5-VL !2025-03-28930249|Qwen2.5-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.| | 535|WongKinYiu/yolov9 !2025-03-2892201|Implementation of paper - YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information| | 536|alibaba-damo-academy/FunASR !2025-03-28920222|A Fundamental End-to-End Speech Recognition Toolkit and Open Source SOTA Pretrained Models.| | 537|Visualize-ML/Book4Power-of-Matrix !2025-03-2891931|Book4 'Power of Matrix' | | 538|dice2o/BingGPT !2025-03-289185-1 |Desktop application of new Bing's AI-powered chat (Windows, macOS and Linux)| | 539|browserbase/stagehand !2025-03-28917621|An AI web browsing framework focused on simplicity and extensibility.| | 540|FlagOpen/FlagEmbedding !2025-03-28914111|Dense Retrieval and Retrieval-augmented LLMs| | 541|Const-me/Whisper !2025-03-2890979|High-performance GPGPU inference of OpenAI's Whisper automatic speech recognition (ASR) model| | 542|lucidrains/denoising-diffusion-pytorch !2025-03-2890942|Implementation of Denoising Diffusion Probabilistic Model in Pytorch| | 543|Chainlit/chainlit !2025-03-28904422|Build Conversational AI in minutes ⚡️| | 544|togethercomputer/OpenChatKit !2025-03-2890160 |OpenChatKit provides a powerful, open-source base to create both specialized and general purpose chatbots for various applications| | 545|Stability-AI/StableStudio !2025-03-2889631 |Community interface for generative AI| | 546|voicepaw/so-vits-svc-fork !2025-03-2889482 |so-vits-svc fork with realtime support, improved interface and more features.| | 547|pymc-devs/pymc !2025-03-2889413|Bayesian Modeling and Probabilistic Programming in Python| | 548|espnet/espnet !2025-03-2889302|End-to-End Speech Processing Toolkit| | 549|kedacore/keda !2025-03-2888991|KEDA is a Kubernetes-based Event Driven Autoscaling component. It provides event driven scale for any container running in Kubernetes| | 550|open-mmlab/Amphion !2025-03-28886911|Amphion (/æmˈfaɪən/) is a toolkit for Audio, Music, and Speech Generation. Its purpose is to support reproducible research and help junior researchers and engineers get started in the field of audio, music, and speech generation research and development.| | 551|gorse-io/gorse !2025-03-2888451|Gorse open source recommender system engine| | 552|adams549659584/go-proxy-bingai !2025-03-288768-1 |A Microsoft New Bing demo site built with Vue3 and Go, providing a consistent UI experience, supporting ChatGPT prompts, and accessible within China| | 553|open-mmlab/mmsegmentation !2025-03-2887513|OpenMMLab Semantic Segmentation Toolbox and Benchmark.| | 554|bytedance/monolith !2025-03-2887223|ByteDance's Recommendation System| | 555|LouisShark/chatgptsystemprompt !2025-03-2887216|store all agent's system prompt| | 556|brexhq/prompt-engineering !2025-03-2887080 |Tips and tricks for working with Large Language Models like OpenAI's GPT-4.| | 557|erincatto/box2d !2025-03-2886841|Box2D is a 2D physics engine for games| | 558|🔥microsoft/ai-agents-for-beginners !2025-03-288669323|10 Lessons to Get Started Building AI Agents| | 559|nashsu/FreeAskInternet !2025-03-2886102|FreeAskInternet is a completely free, private and locally running search aggregator & answer generate using LLM, without GPU needed. The user can ask a question and the system will make a multi engine search and combine the search result to the ChatGPT3.5 LLM and generate the answer based on search results.| | 560|goldmansachs/gs-quant !2025-03-2885981|Python toolkit for quantitative finance| | 561|srbhr/Resume-Matcher !2025-03-2885800|Open Source Free ATS Tool to compare Resumes with Job Descriptions and create a score to rank them.| | 562|facebookresearch/ImageBind !2025-03-2885681 |ImageBind One Embedding Space to Bind Them All| | 563|ashawkey/stable-dreamfusion !2025-03-2885481 |A pytorch implementation of text-to-3D dreamfusion, powered by stable diffusion.| | 564|meetecho/janus-gateway !2025-03-2885232|Janus WebRTC Server| | 565|google/magika !2025-03-2885003|Detect file content types with deep learning| | 566|huggingface/chat-ui !2025-03-2884871 |Open source codebase powering the HuggingChat app| | 567|EleutherAI/lm-evaluation-harness !2025-03-28843012|A framework for few-shot evaluation of autoregressive language models.| | 568|jina-ai/reader !2025-03-2884089|Convert any URL to an LLM-friendly input with a simple prefix https://r.jina.ai/| | 569|microsoft/TypeChat !2025-03-288406-1|TypeChat is a library that makes it easy to build natural language interfaces using types.| | 570|thuml/Time-Series-Library !2025-03-28839715|A Library for Advanced Deep Time Series Models.| | 571|OptimalScale/LMFlow !2025-03-2883882|An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Model for All.| | 572|baptisteArno/typebot.io !2025-03-2883845|💬 Typebot is a powerful chatbot builder that you can self-host.| | 573|jzhang38/TinyLlama !2025-03-2883504|The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.| | 574|fishaudio/Bert-VITS2 !2025-03-2883472|vits2 backbone with multilingual-bert| | 575|OpenBMB/XAgent !2025-03-2882683|An Autonomous LLM Agent for Complex Task Solving| | 576|Acly/krita-ai-diffusion !2025-03-2882387|Streamlined interface for generating images with AI in Krita. Inpaint and outpaint with optional text prompt, no tweaking required.| | 577|jasonppy/VoiceCraft !2025-03-2882151|Zero-Shot Speech Editing and Text-to-Speech in the Wild| | 578|SJTU-IPADS/PowerInfer !2025-03-2881693|High-speed Large Language Model Serving on PCs with Consumer-grade GPUs| | 579|modelscope/DiffSynth-Studio !2025-03-28814713|Enjoy the magic of Diffusion models!| | 580|o3de/o3de !2025-03-2881443|Open 3D Engine (O3DE) is an Apache 2.0-licensed multi-platform 3D engine that enables developers and content creators to build AAA games, cinema-quality 3D worlds, and high-fidelity simulations without any fees or commercial obligations.| | 581|zmh-program/chatnio !2025-03-2881325|🚀 Next Generation AI One-Stop Internationalization Solution. 🚀 下一代 AI 一站式 B/C 端解决方案,支持 OpenAI,Midjourney,Claude,讯飞星火,Stable Diffusion,DALL·E,ChatGLM,通义千问,腾讯混元,360 智脑,百川 AI,火山方舟,新必应,Gemini,Moonshot 等模型,支持对话分享,自定义预设,云端同步,模型市场,支持弹性计费和订阅计划模式,支持图片解析,支持联网搜索,支持模型缓存,丰富美观的后台管理与仪表盘数据统计。| | 582|leptonai/searchwithlepton !2025-03-2880632|Building a quick conversation-based search demo with Lepton AI.| | 583|sebastianstarke/AI4Animation !2025-03-2880620|Bringing Characters to Life with Computer Brains in Unity| | 584|wangrongding/wechat-bot !2025-03-2880528|🤖一个基于 WeChaty 结合 DeepSeek / ChatGPT / Kimi / 讯飞等Ai服务实现的微信机器人 ,可以用来帮助你自动回复微信消息,或者管理微信群/好友,检测僵尸粉等...| | 585|openvinotoolkit/openvino !2025-03-2880528|OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference| | 586|steven2358/awesome-generative-ai !2025-03-28802610|A curated list of modern Generative Artificial Intelligence projects and services| | 587|adam-maj/tiny-gpu !2025-03-2880234|A minimal GPU design in Verilog to learn how GPUs work from the ground up| | 588| anse-app/chatgpt-demo !2025-03-2880180 | A demo repo based on OpenAI API (gpt-3.5-turbo) | | 589| acheong08/EdgeGPT !2025-03-288015-1 |Reverse engineered API of Microsoft's Bing Chat | | 590|ai-collection/ai-collection !2025-03-2879994 |The Generative AI Landscape - A Collection of Awesome Generative AI Applications| | 591|GreyDGL/PentestGPT !2025-03-2879953 |A GPT-empowered penetration testing tool| | 592|delta-io/delta !2025-03-2879112|An open-source storage framework that enables building a Lakehouse architecture with compute engines including Spark, PrestoDB, Flink, Trino, and Hive and APIs| | 593|dataelement/bisheng !2025-03-2879085|Bisheng is an open LLM devops platform for next generation AI applications.| | 594|e2b-dev/e2b !2025-03-2878447 |Vercel for AI agents. We help developers to build, deploy, and monitor AI agents. Focusing on specialized AI agents that build software for you - your personal software developers.| | 595|01-ai/Yi !2025-03-2878311|A series of large language models trained from scratch by developers @01-ai| | 596|Plachtaa/VALL-E-X !2025-03-287830-1|An open source implementation of Microsoft's VALL-E X zero-shot TTS model. The demo is available at https://plachtaa.github.io| | 597|abhishekkrthakur/approachingalmost !2025-03-2878204|Approaching (Almost) Any Machine Learning Problem| | 598|pydantic/pydantic-ai !2025-03-28781041|Agent Framework / shim to use Pydantic with LLMs| | 599|rany2/edge-tts !2025-03-2877901|Use Microsoft Edge's online text-to-speech service from Python WITHOUT needing Microsoft Edge or Windows or an API key| | 600|CASIA-IVA-Lab/FastSAM !2025-03-2877881|Fast Segment Anything| | 601|netease-youdao/EmotiVoice !2025-03-2877817|EmotiVoice 😊: a Multi-Voice and Prompt-Controlled TTS Engine| | 602|lllyasviel/IC-Light !2025-03-2877804|More relighting!| | 603|kroma-network/tachyon !2025-03-287774-1|Modular ZK(Zero Knowledge) backend accelerated by GPU| | 604|deep-floyd/IF !2025-03-2877731 |A novel state-of-the-art open-source text-to-image model with a high degree of photorealism and language understanding| | 605|oumi-ai/oumi !2025-03-2877705|Everything you need to build state-of-the-art foundation models, end-to-end.| | 606|reorproject/reor !2025-03-2877681|AI note-taking app that runs models locally.| | 607|lightpanda-io/browser !2025-03-28775813|Lightpanda: the headless browser designed for AI and automation| | 608|xiangsx/gpt4free-ts !2025-03-287755-1|Providing a free OpenAI GPT-4 API ! This is a replication project for the typescript version of xtekky/gpt4free| | 609|IDEA-Research/GroundingDINO !2025-03-28773311|Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"| | 610|bunkerity/bunkerweb !2025-03-2877326|🛡️ Make your web services secure by default !| | 611|vikhyat/moondream !2025-03-2877057|tiny vision language model| | 612|firmai/financial-machine-learning !2025-03-287703-1|A curated list of practical financial machine learning tools and applications.| | 613|n8n-io/self-hosted-ai-starter-kit !2025-03-28765121|The Self-hosted AI Starter Kit is an open-source template that quickly sets up a local AI environment. Curated by n8n, it provides essential tools for creating secure, self-hosted AI workflows.| | 614|intel-analytics/ipex-llm !2025-03-2876507|Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with llama.cpp, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, ModelScope, etc.| | 615|jrouwe/JoltPhysics !2025-03-28764510|A multi core friendly rigid body physics and collision detection library. Written in C++. Suitable for games and VR applications. Used by Horizon Forbidden West.| | 616|THUDM/CodeGeeX2 !2025-03-2876270|CodeGeeX2: A More Powerful Multilingual Code Generation Model| | 617|meta-llama/llama-stack !2025-03-2875866|Composable building blocks to build Llama Apps| | 618|sweepai/sweep !2025-03-287530-1|Sweep is an AI junior developer| | 619|lllyasviel/Omost !2025-03-2875301|Your image is almost there!| | 620|ahmedbahaaeldin/From-0-to-Research-Scientist-resources-guide !2025-03-2875050|Detailed and tailored guide for undergraduate students or anybody want to dig deep into the field of AI with solid foundation.| | 621|dair-ai/ML-Papers-Explained !2025-03-2875050|Explanation to key concepts in ML| | 622|zaidmukaddam/scira !2025-03-28750110|Scira (Formerly MiniPerplx) is a minimalistic AI-powered search engine that helps you find information on the internet. Powered by Vercel AI SDK! Search with models like Grok 2.0.| | 623|Portkey-AI/gateway !2025-03-28749416|A Blazing Fast AI Gateway. Route to 100+ LLMs with 1 fast & friendly API.| | 624|web-infra-dev/midscene !2025-03-28748729|An AI-powered automation SDK can control the page, perform assertions, and extract data in JSON format using natural language.| | 625|zilliztech/GPTCache !2025-03-2874801 |GPTCache is a library for creating semantic cache to store responses from LLM queries.| | 626|niedev/RTranslator !2025-03-2874742|RTranslator is the world's first open source real-time translation app.| |!green-up-arrow.svg 627|roboflow/notebooks !2025-03-2874666|Examples and tutorials on using SOTA computer vision models and techniques. Learn everything from old-school ResNet, through YOLO and object-detection transformers like DETR, to the latest models like Grounding DINO and SAM.| |!red-down-arrow 628|openlm-research/openllama !2025-03-2874652|OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset| | 629|LiheYoung/Depth-Anything !2025-03-2874155|Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data| | 630|enso-org/enso !2025-03-2874040|Hybrid visual and textual functional programming.| | 631|bigcode-project/starcoder !2025-03-287401-1 |Home of StarCoder: fine-tuning & inference!| | 632|git-ecosystem/git-credential-manager !2025-03-2873975|Secure, cross-platform Git credential storage with authentication to GitHub, Azure Repos, and other popular Git hosting services.| | 633|OpenGVLab/InternVL !2025-03-2873634|[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4V. 接近GPT-4V表现的可商用开源模型| | 634|WooooDyy/LLM-Agent-Paper-List !2025-03-2873551|The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.| | 635|lencx/Noi !2025-03-2873157|🦄 AI + Tools + Plugins + Community| | 636|udlbook/udlbook !2025-03-2873075|Understanding Deep Learning - Simon J.D. Prince| | 637|OpenBMB/MiniCPM !2025-03-2872841|MiniCPM-2B: An end-side LLM outperforms Llama2-13B.| | 638|jaywalnut310/vits !2025-03-2872815 |VITS: Conditional Variational Autoencoder with Adversarial Learning for End-to-End Text-to-Speech| | 639|xorbitsai/inference !2025-03-28727528|Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.| | 640|PWhiddy/PokemonRedExperiments !2025-03-2872492|Playing Pokemon Red with Reinforcement Learning| | 641|Canner/WrenAI !2025-03-28723213|🤖 Open-source AI Agent that empowers data-driven teams to chat with their data to generate Text-to-SQL, charts, spreadsheets, reports, and BI. 📈📊📋🧑‍💻| | 642|miurla/morphic !2025-03-2872258|An AI-powered answer engine with a generative UI| | 643|ml-explore/mlx-examples !2025-03-2872168|Examples in the MLX framework| | 644|PKU-YuanGroup/ChatLaw !2025-03-2872010|Chinese Legal Large Model| | 645|NVIDIA/cutlass !2025-03-2871883|CUDA Templates for Linear Algebra Subroutines| | 646|FoundationVision/VAR !2025-03-28717444|[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction"| | 647|ymcui/Chinese-LLaMA-Alpaca-2 !2025-03-2871561|Chinese LLaMA-2 & Alpaca-2 LLMs| | 648|nadermx/backgroundremover !2025-03-2871514 |Background Remover lets you Remove Background from images and video using AI with a simple command line interface that is free and open source.| | 649|onuratakan/gpt-computer-assistant !2025-03-28714514|gpt-4o for windows, macos and ubuntu| | 650|graviraja/MLOps-Basics !2025-03-2871326|| | 651|Future-House/paper-qa !2025-03-287118-1|High accuracy RAG for answering questions from scientific documents with citations| | 652|open-mmlab/mmagic !2025-03-2871102 |OpenMMLab Multimodal Advanced, Generative, and Intelligent Creation Toolbox| | 653|bhaskatripathi/pdfGPT !2025-03-2870941 |PDF GPT allows you to chat with the contents of your PDF file by using GPT capabilities. The only open source solution to turn your pdf files in a chatbot!| | 654|ollama/ollama-python !2025-03-28709117|Ollama Python library| | 655|facebookresearch/DiT !2025-03-2870376|Official PyTorch Implementation of "Scalable Diffusion Models with Transformers"| | 656|geekyutao/Inpaint-Anything !2025-03-2870262 |Inpaint anything using Segment Anything and inpainting models.| | 657|AbdullahAlfaraj/Auto-Photoshop-StableDiffusion-Plugin !2025-03-2870160 |A user-friendly plug-in that makes it easy to generate stable diffusion images inside Photoshop using Automatic1111-sd-webui as a backend.| | 658|apple/corenet !2025-03-2869990|CoreNet: A library for training deep neural networks| | 659|openstatusHQ/openstatus !2025-03-2869926|🏓 The open-source synthetic monitoring platform 🏓| | 660|weaviate/Verba !2025-03-2869772|Retrieval Augmented Generation (RAG) chatbot powered by Weaviate| | 661|meshery/meshery !2025-03-2869630|Meshery, the cloud native manager| | 662|OpenTalker/video-retalking !2025-03-2869530|[SIGGRAPH Asia 2022] VideoReTalking: Audio-based Lip Synchronization for Talking Head Video Editing In the Wild| | 663|digitalinnovationone/dio-lab-open-source !2025-03-28689013|Repositório do lab "Contribuindo em um Projeto Open Source no GitHub" da Digital Innovation One.| | 664|jianchang512/ChatTTS-ui !2025-03-2868842|一个简单的本地网页界面,直接使用ChatTTS将文字合成为语音,同时支持对外提供API接口。| | 665|patchy631/ai-engineering-hub !2025-03-28686434|In-depth tutorials on LLMs, RAGs and real-world AI agent applications.| | 666|gunnarmorling/1brc !2025-03-2868512|1️⃣🐝🏎️ The One Billion Row Challenge -- A fun exploration of how quickly 1B rows from a text file can be aggregated with Java| | 667|Azure-Samples/azure-search-openai-demo !2025-03-2868482 |A sample app for the Retrieval-Augmented Generation pattern running in Azure, using Azure Cognitive Search for retrieval and Azure OpenAI large language models to power ChatGPT-style and Q&A experiences.| | 668|mit-han-lab/streaming-llm !2025-03-2868382|Efficient Streaming Language Models with Attention Sinks| | 669|InternLM/InternLM !2025-03-2868352|InternLM has open-sourced a 7 billion parameter base model, a chat model tailored for practical scenarios and the training system.| | 670|dependency-check/DependencyCheck !2025-03-2868191|OWASP dependency-check is a software composition analysis utility that detects publicly disclosed vulnerabilities in application dependencies.| | 671|Soulter/AstrBot !2025-03-28678643|✨易上手的多平台 LLM 聊天机器人及开发框架✨。支持 QQ、QQ频道、Telegram、微信平台(Gewechat, 企业微信)、内置 Web Chat,OpenAI GPT、DeepSeek、Ollama、Llama、GLM、Gemini、OneAPI、LLMTuner,支持 LLM Agent 插件开发,可视化面板。一键部署。支持 Dify 工作流、代码执行器、Whisper 语音转文字。| | 672|react-native-webview/react-native-webview !2025-03-2867792|React Native Cross-Platform WebView| | 673|modelscope/agentscope !2025-03-28676916|Start building LLM-empowered multi-agent applications in an easier way.| | 674|mylxsw/aidea !2025-03-2867381|AIdea is a versatile app that supports GPT and domestic large language models,also supports "Stable Diffusion" text-to-image generation, image-to-image generation, SDXL 1.0, super-resolution, and image colorization| | 675|langchain-ai/ollama-deep-researcher !2025-03-28668635|Fully local web research and report writing assistant| | 676|threestudio-project/threestudio !2025-03-2866653|A unified framework for 3D content generation.| | 677|gaomingqi/Track-Anything !2025-03-2866631 |A flexible and interactive tool for video object tracking and segmentation, based on Segment Anything, XMem, and E2FGVI.| | 678|spdustin/ChatGPT-AutoExpert !2025-03-2866570|🚀🧠💬 Supercharged Custom Instructions for ChatGPT (non-coding) and ChatGPT Advanced Data Analysis (coding).| | 679|HariSekhon/DevOps-Bash-tools !2025-03-2866463|1000+ DevOps Bash Scripts - AWS, GCP, Kubernetes, Docker, CI/CD, APIs, SQL, PostgreSQL, MySQL, Hive, Impala, Kafka, Hadoop, Jenkins, GitHub, GitLab, BitBucket, Azure DevOps, TeamCity, Spotify, MP3, LDAP, Code/Build Linting, pkg mgmt for Linux, Mac, Python, Perl, Ruby, NodeJS, Golang, Advanced dotfiles: .bashrc, .vimrc, .gitconfig, .screenrc, tmux..| | 680|modelscope/swift !2025-03-28661530|ms-swift: Use PEFT or Full-parameter to finetune 200+ LLMs or 15+ MLLMs| | 681|langchain-ai/opengpts !2025-03-2866080|This is an open source effort to create a similar experience to OpenAI's GPTs and Assistants API| | 682| yihong0618/xiaogpt !2025-03-2865131 | Play ChatGPT with xiaomi ai speaker | | 683| civitai/civitai !2025-03-2865111 | Build a platform where people can share their stable diffusion models | | 684|KoljaB/RealtimeSTT !2025-03-28649513|A robust, efficient, low-latency speech-to-text library with advanced voice activity detection, wake word activation and instant transcription.| | 685|qunash/chatgpt-advanced !2025-03-2864910 | A browser extension that augments your ChatGPT prompts with web results.| | 686|Licoy/ChatGPT-Midjourney !2025-03-2864850|🎨 Own your own ChatGPT+Midjourney web service with one click| | 687|friuns2/BlackFriday-GPTs-Prompts !2025-03-2864744|List of free GPTs that doesn't require plus subscription| | 688|PixarAnimationStudios/OpenUSD !2025-03-2864700|Universal Scene Description| | 689|linyiLYi/street-fighter-ai !2025-03-2864630 |This is an AI agent for Street Fighter II Champion Edition.| | 690|run-llama/rags !2025-03-2864380|Build ChatGPT over your data, all with natural language| | 691|frdel/agent-zero !2025-03-2864154|Agent Zero AI framework| | 692|microsoft/DeepSpeedExamples !2025-03-2863911 |Example models using DeepSpeed| | 693|k8sgpt-ai/k8sgpt !2025-03-2863882|Giving Kubernetes Superpowers to everyone| | 694|open-metadata/OpenMetadata !2025-03-2863514|OpenMetadata is a unified platform for discovery, observability, and governance powered by a central metadata repository, in-depth lineage, and seamless team collaboration.| | 695|google/gemma.cpp !2025-03-2863163|lightweight, standalone C++ inference engine for Google's Gemma models.| | 696|RayVentura/ShortGPT !2025-03-286314-1|🚀🎬 ShortGPT - An experimental AI framework for automated short/video content creation. Enables creators to rapidly produce, manage, and deliver content using AI and automation.| | 697|openai/consistencymodels !2025-03-2862940 |Official repo for consistency models.| | 698|yangjianxin1/Firefly !2025-03-2862924|Firefly: Chinese conversational large language model (full-scale fine-tuning + QLoRA), supporting fine-tuning of Llma2, Llama, Baichuan, InternLM, Ziya, Bloom, and other large models| | 699|enricoros/big-AGI !2025-03-2862665|Generative AI suite powered by state-of-the-art models and providing advanced AI/AGI functions. It features AI personas, AGI functions, multi-model chats, text-to-image, voice, response streaming, code highlighting and execution, PDF import, presets for developers, much more. Deploy on-prem or in the cloud.| | 700|aptos-labs/aptos-core !2025-03-2862633|Aptos is a layer 1 blockchain built to support the widespread use of blockchain through better technology and user experience.| | 701|wenda-LLM/wenda !2025-03-286262-1 |Wenda: An LLM invocation platform. Its objective is to achieve efficient content generation tailored to specific environments while considering the limited computing resources of individuals and small businesses, as well as knowledge security and privacy concerns| | 702|Project-MONAI/MONAI !2025-03-2862603|AI Toolkit for Healthcare Imaging| | 703|HVision-NKU/StoryDiffusion !2025-03-2862470|Create Magic Story!| | 704|deepseek-ai/DeepSeek-LLM !2025-03-2862463|DeepSeek LLM: Let there be answers| | 705|Tohrusky/Final2x !2025-03-2862393|2^x Image Super-Resolution| | 706|OpenSPG/KAG !2025-03-28619611|KAG is a logical form-guided reasoning and retrieval framework based on OpenSPG engine and LLMs. It is used to build logical reasoning and factual Q&A solutions for professional domain knowledge bases. It can effectively overcome the shortcomings of the traditional RAG vector similarity calculation model.| | 707|Moonvy/OpenPromptStudio !2025-03-2861861 |AIGC Hint Word Visualization Editor| | 708|levihsu/OOTDiffusion !2025-03-2861761|Official implementation of OOTDiffusion| | 709|tmc/langchaingo !2025-03-2861729|LangChain for Go, the easiest way to write LLM-based programs in Go| | 710|vladmandic/automatic !2025-03-2861374|SD.Next: Advanced Implementation of Stable Diffusion and other Diffusion-based generative image models| | 711|clovaai/donut !2025-03-2861231 |Official Implementation of OCR-free Document Understanding Transformer (Donut) and Synthetic Document Generator (SynthDoG), ECCV 2022| | 712|Shaunwei/RealChar !2025-03-286121-1|🎙️🤖Create, Customize and Talk to your AI Character/Companion in Realtime(All in One Codebase!). Have a natural seamless conversation with AI everywhere(mobile, web and terminal) using LLM OpenAI GPT3.5/4, Anthropic Claude2, Chroma Vector DB, Whisper Speech2Text, ElevenLabs Text2Speech🎙️🤖| | 713|microsoft/TinyTroupe !2025-03-2861142|LLM-powered multiagent persona simulation for imagination enhancement and business insights.| | 714| rustformers/llm !2025-03-2861010 | Run inference for Large Language Models on CPU, with Rust| | 715|firebase/firebase-ios-sdk !2025-03-2860950|Firebase SDK for Apple App Development| | 716|vespa-engine/vespa !2025-03-2860824|The open big data serving engine. https://vespa.ai| | 717|n4ze3m/page-assist !2025-03-28607610|Use your locally running AI models to assist you in your web browsing| | 718|Dooy/chatgpt-web-midjourney-proxy !2025-03-2860646|chatgpt web, midjourney, gpts,tts, whisper 一套ui全搞定| | 719|ethereum-optimism/optimism !2025-03-2860213|Optimism is Ethereum, scaled.| | 720|sczhou/ProPainter !2025-03-2859971|[ICCV 2023] ProPainter: Improving Propagation and Transformer for Video Inpainting| | 721|MineDojo/Voyager !2025-03-2859951 |An Open-Ended Embodied Agent with Large Language Models| | 722|lavague-ai/LaVague !2025-03-2859800|Automate automation with Large Action Model framework| | 723|SevaSk/ecoute !2025-03-2859770 |Ecoute is a live transcription tool that provides real-time transcripts for both the user's microphone input (You) and the user's speakers output (Speaker) in a textbox. It also generates a suggested response using OpenAI's GPT-3.5 for the user to say based on the live transcription of the conversation.| | 724|google/mesop !2025-03-2859661|| | 725|pengxiao-song/LaWGPT !2025-03-2859542 |Repo for LaWGPT, Chinese-Llama tuned with Chinese Legal knowledge| | 726|fr0gger/Awesome-GPT-Agents !2025-03-2859434|A curated list of GPT agents for cybersecurity| | 727|google-deepmind/graphcast !2025-03-2859412|| | 728|comet-ml/opik !2025-03-28594126|Open-source end-to-end LLM Development Platform| | 729|SciPhi-AI/R2R !2025-03-28594033|A framework for rapid development and deployment of production-ready RAG systems| | 730|SkalskiP/courses !2025-03-2859272 |This repository is a curated collection of links to various courses and resources about Artificial Intelligence (AI)| | 731|QuivrHQ/MegaParse !2025-03-2859122|File Parser optimised for LLM Ingestion with no loss 🧠 Parse PDFs, Docx, PPTx in a format that is ideal for LLMs.| | 732|pytorch-labs/gpt-fast !2025-03-2858971|Simple and efficient pytorch-native transformer text generation in !2025-03-2858886|Curated list of chatgpt prompts from the top-rated GPTs in the GPTs Store. Prompt Engineering, prompt attack & prompt protect. Advanced Prompt Engineering papers.| | 734|nilsherzig/LLocalSearch !2025-03-2858852|LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a chain of LLMs to find the answer. The user can see the progress of the agents and the final answer. No OpenAI or Google API keys are needed.| | 735|kuafuai/DevOpsGPT !2025-03-285874-2|Multi agent system for AI-driven software development. Convert natural language requirements into working software. Supports any development language and extends the existing base code.| | 736|myshell-ai/MeloTTS !2025-03-2858486|High-quality multi-lingual text-to-speech library by MyShell.ai. Support English, Spanish, French, Chinese, Japanese and Korean.| | 737|OpenGVLab/LLaMA-Adapter !2025-03-2858421 |Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters| | 738|volcengine/verl !2025-03-28582563|veRL: Volcano Engine Reinforcement Learning for LLM| | 739|a16z-infra/companion-app !2025-03-2858171|AI companions with memory: a lightweight stack to create and host your own AI companions| | 740|HumanAIGC/OutfitAnyone !2025-03-285816-1|Outfit Anyone: Ultra-high quality virtual try-on for Any Clothing and Any Person| | 741|josStorer/RWKV-Runner !2025-03-2857472|A RWKV management and startup tool, full automation, only 8MB. And provides an interface compatible with the OpenAI API. RWKV is a large language model that is fully open source and available for commercial use.| | 742|648540858/wvp-GB28181-pro !2025-03-2857414|WEB VIDEO PLATFORM是一个基于GB28181-2016标准实现的网络视频平台,支持NAT穿透,支持海康、大华、宇视等品牌的IPC、NVR、DVR接入。支持国标级联,支持rtsp/rtmp等视频流转发到国标平台,支持rtsp/rtmp等推流转发到国标平台。| | 743|ToonCrafter/ToonCrafter !2025-03-2857345|a research paper for generative cartoon interpolation| | 744|PawanOsman/ChatGPT !2025-03-2857191|OpenAI API Free Reverse Proxy| | 745|apache/hudi !2025-03-2857091|Upserts, Deletes And Incremental Processing on Big Data.| | 746| nsarrazin/serge !2025-03-2857081 | A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API| | 747|homanp/superagent !2025-03-2857021|🥷 Superagent - Build, deploy, and manage LLM-powered agents| | 748|ramonvc/freegpt-webui !2025-03-2856910|GPT 3.5/4 with a Chat Web UI. No API key is required.| | 749|baichuan-inc/baichuan-7B !2025-03-2856901|A large-scale 7B pretraining language model developed by BaiChuan-Inc.| | 750|Azure/azure-sdk-for-net !2025-03-2856792|This repository is for active development of the Azure SDK for .NET. For consumers of the SDK we recommend visiting our public developer docs at https://learn.microsoft.com/dotnet/azure/ or our versioned developer docs at https://azure.github.io/azure-sdk-for-net.| | 751|mnotgod96/AppAgent !2025-03-2856643|AppAgent: Multimodal Agents as Smartphone Users, an LLM-based multimodal agent framework designed to operate smartphone apps.| | 752|microsoft/TaskWeaver !2025-03-2856243|A code-first agent framework for seamlessly planning and executing data analytics tasks.| | 753| yetone/bob-plugin-openai-translator !2025-03-285600-1 | A Bob Plugin base ChatGPT API | | 754|PrefectHQ/marvin !2025-03-2855840 |A batteries-included library for building AI-powered software| | 755|microsoft/promptbase !2025-03-2855832|All things prompt engineering| | 756|fullstackhero/dotnet-starter-kit !2025-03-2855560|Production Grade Cloud-Ready .NET 8 Starter Kit (Web API + Blazor Client) with Multitenancy Support, and Clean/Modular Architecture that saves roughly 200+ Development Hours! All Batteries Included.| | 757|deepseek-ai/DeepSeek-Coder-V2 !2025-03-2855435|DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence| | 758|aiwaves-cn/agents !2025-03-2855391|An Open-source Framework for Autonomous Language Agents| | 759|microsoft/Mastering-GitHub-Copilot-for-Paired-Programming !2025-03-2855158|A 6 Lesson course teaching everything you need to know about harnessing GitHub Copilot and an AI Paired Programing resource.| | 760|allenai/OLMo !2025-03-2854506|Modeling, training, eval, and inference code for OLMo| | 761|apify/crawlee-python !2025-03-2854493|Crawlee—A web scraping and browser automation library for Python to build reliable crawlers. Extract data for AI, LLMs, RAG, or GPTs. Download HTML, PDF, JPG, PNG, and other files from websites. Works with BeautifulSoup, Playwright, and raw HTTP. Both headful and headless mode. With proxy rotation.| | 762|k2-fsa/sherpa-onnx !2025-03-28541520|Speech-to-text, text-to-speech, and speaker recongition using next-gen Kaldi with onnxruntime without Internet connection. Support embedded systems, Android, iOS, Raspberry Pi, RISC-V, x86_64 servers, websocket server/client, C/C++, Python, Kotlin, C#, Go, NodeJS, Java, Swift| | 763|TEN-framework/TEN-Agent !2025-03-28541411|TEN Agent is a realtime conversational AI agent powered by TEN. It seamlessly integrates the OpenAI Realtime API, RTC capabilities, and advanced features like weather updates, web search, computer vision, and Retrieval-Augmented Generation (RAG).| | 764|google/gemmapytorch !2025-03-2854010|The official PyTorch implementation of Google's Gemma models| | 765|snakers4/silero-vad !2025-03-2853858|Silero VAD: pre-trained enterprise-grade Voice Activity Detector| | 766|livekit/agents !2025-03-2853836|Build real-time multimodal AI applications 🤖🎙️📹| | 767|pipecat-ai/pipecat !2025-03-28537811|Open Source framework for voice and multimodal conversational AI| | 768|EricLBuehler/mistral.rs !2025-03-28536324|Blazingly fast LLM inference.| | 769|asg017/sqlite-vec !2025-03-28535810|Work-in-progress vector search SQLite extension that runs anywhere.| | 770|albertan017/LLM4Decompile !2025-03-2853563|Reverse Engineering: Decompiling Binary Code with Large Language Models| | 771|Permify/permify !2025-03-2853235|An open-source authorization as a service inspired by Google Zanzibar, designed to build and manage fine-grained and scalable authorization systems for any application.| | 772|imoneoi/openchat !2025-03-2853171|OpenChat: Advancing Open-source Language Models with Imperfect Data| | 773|mosaicml/composer !2025-03-2853140|Train neural networks up to 7x faster| | 774|dsdanielpark/Bard-API !2025-03-285277-1 |The python package that returns a response of Google Bard through API.| | 775|lxfater/inpaint-web !2025-03-2852552|A free and open-source inpainting & image-upscaling tool powered by webgpu and wasm on the browser。| | 776|leanprover/lean4 !2025-03-2852441|Lean 4 programming language and theorem prover| | 777|AILab-CVC/YOLO-World !2025-03-2852415|Real-Time Open-Vocabulary Object Detection| | 778|openchatai/OpenChat !2025-03-2852260 |Run and create custom ChatGPT-like bots with OpenChat, embed and share these bots anywhere, the open-source chatbot console.| | 779|mufeedvh/code2prompt !2025-03-28519414|A CLI tool to convert your codebase into a single LLM prompt with source tree, prompt templating, and token counting.| | 780|biobootloader/wolverine !2025-03-2851700 |Automatically repair python scripts through GPT-4 to give them regenerative abilities.| | 781|huggingface/parler-tts !2025-03-2851671|Inference and training library for high-quality TTS models.| | 782|Akegarasu/lora-scripts !2025-03-2851308 |LoRA training scripts use kohya-ss's trainer, for diffusion model.| | 783|openchatai/OpenCopilot !2025-03-285128-3|🤖 🔥 Let your users chat with your product features and execute things by text - open source Shopify sidekick| | 784|e2b-dev/fragments !2025-03-2851228|Open-source Next.js template for building apps that are fully generated by AI. By E2B.| | 785|microsoft/SynapseML !2025-03-2851132|Simple and Distributed Machine Learning| | 786|aigc-apps/sd-webui-EasyPhoto !2025-03-285108-1|📷 EasyPhoto | | 787|ChaoningZhang/MobileSAM !2025-03-2850944|This is the official code for Faster Segment Anything (MobileSAM) project that makes SAM lightweight| | 788|huggingface/alignment-handbook !2025-03-2850932|Robust recipes for to align language models with human and AI preferences| | 789|alpkeskin/mosint !2025-03-2850920|An automated e-mail OSINT tool| | 790|TaskingAI/TaskingAI !2025-03-2850891|The open source platform for AI-native application development.| | 791|lipku/metahuman-stream !2025-03-28507615|Real time interactive streaming digital human| | 792|OpenInterpreter/01 !2025-03-2850530|The open-source language model computer| | 793|open-compass/opencompass !2025-03-28505111|OpenCompass is an LLM evaluation platform, supporting a wide range of models (InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.| | 794|xxlong0/Wonder3D !2025-03-2850491|A cross-domain diffusion model for 3D reconstruction from a single image| | 795|pytorch/torchtune !2025-03-2850342|A Native-PyTorch Library for LLM Fine-tuning| | 796|SuperDuperDB/superduperdb !2025-03-2850192|🔮 SuperDuperDB: Bring AI to your database: Integrate, train and manage any AI models and APIs directly with your database and your data.| | 797|WhiskeySockets/Baileys !2025-03-2850057|Lightweight full-featured typescript/javascript WhatsApp Web API| | 798| mpociot/chatgpt-vscode !2025-03-2849890 | A VSCode extension that allows you to use ChatGPT | | 799|OpenGVLab/DragGAN !2025-03-2849880|Unofficial Implementation of DragGAN - "Drag Your GAN: Interactive Point-based Manipulation on the Generative Image Manifold" (DragGAN 全功能实现,在线Demo,本地部署试用,代码、模型已全部开源,支持Windows, macOS, Linux)| | 800|microsoft/LLMLingua !2025-03-2849824|To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.| | 801|Zipstack/unstract !2025-03-2849745|No-code LLM Platform to launch APIs and ETL Pipelines to structure unstructured documents| | 802|OpenBMB/ToolBench !2025-03-2849621|An open platform for training, serving, and evaluating large language model for tool learning.| | 803|Fanghua-Yu/SUPIR !2025-03-2849593|SUPIR aims at developing Practical Algorithms for Photo-Realistic Image Restoration In the Wild| | 804|GaiaNet-AI/gaianet-node !2025-03-2849360|Install and run your own AI agent service| | 805|qodo-ai/qodo-cover !2025-03-284922-1|Qodo-Cover: An AI-Powered Tool for Automated Test Generation and Code Coverage Enhancement! 💻🤖🧪🐞| | 806|Zejun-Yang/AniPortrait !2025-03-2849042|AniPortrait: Audio-Driven Synthesis of Photorealistic Portrait Animation| | 807|lvwzhen/law-cn-ai !2025-03-2848901 |⚖️ AI Legal Assistant| | 808|developersdigest/llm-answer-engine !2025-03-2848740|Build a Perplexity-Inspired Answer Engine Using Next.js, Groq, Mixtral, Langchain, OpenAI, Brave & Serper| | 809|Plachtaa/VITS-fast-fine-tuning !2025-03-2848640|This repo is a pipeline of VITS finetuning for fast speaker adaptation TTS, and many-to-many voice conversion| | 810|espeak-ng/espeak-ng !2025-03-2848601|eSpeak NG is an open source speech synthesizer that supports more than hundred languages and accents.| | 811|ant-research/CoDeF !2025-03-2848581|[CVPR'24 Highlight] Official PyTorch implementation of CoDeF: Content Deformation Fields for Temporally Consistent Video Processing| | 812|deepseek-ai/DeepSeek-V2 !2025-03-2848512|| | 813|XRPLF/rippled !2025-03-2848210|Decentralized cryptocurrency blockchain daemon implementing the XRP Ledger protocol in C++| | 814|AutoMQ/automq !2025-03-28478721|AutoMQ is a cloud-first alternative to Kafka by decoupling durability to S3 and EBS. 10x cost-effective. Autoscale in seconds. Single-digit ms latency.| | 815|AILab-CVC/VideoCrafter !2025-03-2847800|VideoCrafter1: Open Diffusion Models for High-Quality Video Generation| | 816|nautechsystems/nautilustrader !2025-03-2847702|A high-performance algorithmic trading platform and event-driven backtester| | 817|kyegomez/swarms !2025-03-2847563|The Enterprise-Grade Production-Ready Multi-Agent Orchestration Framework Join our Community: https://discord.com/servers/agora-999382051935506503| | 818|Deci-AI/super-gradients !2025-03-2847310 |Easily train or fine-tune SOTA computer vision models with one open source training library. The home of Yolo-NAS.| | 819|QwenLM/Qwen2.5-Coder !2025-03-2847236|Qwen2.5-Coder is the code version of Qwen2.5, the large language model series developed by Qwen team, Alibaba Cloud.| | 820|SCIR-HI/Huatuo-Llama-Med-Chinese !2025-03-2847191 |Repo for HuaTuo (华驼), Llama-7B tuned with Chinese medical knowledge| | 821|togethercomputer/RedPajama-Data !2025-03-2846841 |code for preparing large datasets for training large language models| | 822|mishushakov/llm-scraper !2025-03-2846704|Turn any webpage into structured data using LLMs| | 823|1rgs/jsonformer !2025-03-2846663 |A Bulletproof Way to Generate Structured JSON from Language Models| | 824|anti-work/shortest !2025-03-2846565|QA via natural language AI tests| | 825|dnhkng/GlaDOS !2025-03-2846510|This is the Personality Core for GLaDOS, the first steps towards a real-life implementation of the AI from the Portal series by Valve.| | 826|Nukem9/dlssg-to-fsr3 !2025-03-2846380|Adds AMD FSR3 Frame Generation to games by replacing Nvidia DLSS-G Frame Generation (nvngx_dlssg).| | 827|BuilderIO/ai-shell !2025-03-2846373 |A CLI that converts natural language to shell commands.| | 828|facebookincubator/AITemplate !2025-03-2846220 |AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.| | 829|terraform-aws-modules/terraform-aws-eks !2025-03-2846030|Terraform module to create AWS Elastic Kubernetes (EKS) resources 🇺🇦| | 830|timescale/pgai !2025-03-2845915|A suite of tools to develop RAG, semantic search, and other AI applications more easily with PostgreSQL| | 831|awslabs/multi-agent-orchestrator !2025-03-2845788|Flexible and powerful framework for managing multiple AI agents and handling complex conversations| | 832|sanchit-gandhi/whisper-jax !2025-03-2845771 |Optimised JAX code for OpenAI's Whisper Model, largely built on the Hugging Face Transformers Whisper implementation| | 833|NVIDIA/NeMo-Guardrails !2025-03-2845755|NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.| | 834|PathOfBuildingCommunity/PathOfBuilding !2025-03-2845480|Offline build planner for Path of Exile.| | 835|UX-Decoder/Segment-Everything-Everywhere-All-At-Once !2025-03-2845412 |Official implementation of the paper "Segment Everything Everywhere All at Once"| | 836|build-trust/ockam !2025-03-2845171|Orchestrate end-to-end encryption, cryptographic identities, mutual authentication, and authorization policies between distributed applications – at massive scale.| | 837|google-research/timesfm !2025-03-2845135|TimesFM (Time Series Foundation Model) is a pretrained time-series foundation model developed by Google Research for time-series forecasting.| | 838|luosiallen/latent-consistency-model !2025-03-2844842|Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference| | 839|NVlabs/neuralangelo !2025-03-2844740|Official implementation of "Neuralangelo: High-Fidelity Neural Surface Reconstruction" (CVPR 2023)| | 840|kyegomez/tree-of-thoughts !2025-03-2844720 |Plug in and Play Implementation of Tree of Thoughts: Deliberate Problem Solving with Large Language Models that Elevates Model Reasoning by atleast 70%| | 841|sjvasquez/handwriting-synthesis !2025-03-2844720 |Handwriting Synthesis with RNNs ✏️| | 842| madawei2699/myGPTReader !2025-03-2844420 | A slack bot that can read any webpage, ebook or document and summarize it with chatGPT | | 843|OpenBMB/AgentVerse !2025-03-2844413|🤖 AgentVerse 🪐 provides a flexible framework that simplifies the process of building custom multi-agent environments for large language models (LLMs).| | 844|argmaxinc/WhisperKit !2025-03-2844395|Swift native speech recognition on-device for iOS and macOS applications.| | 845|landing-ai/vision-agent !2025-03-2844346|Vision agent| | 846|InternLM/xtuner !2025-03-2844273|An efficient, flexible and full-featured toolkit for fine-tuning large models (InternLM, Llama, Baichuan, Qwen, ChatGLM)| | 847|google-deepmind/alphageometry !2025-03-284421-1|Solving Olympiad Geometry without Human Demonstrations| | 848|ostris/ai-toolkit !2025-03-2844093|Various AI scripts. Mostly Stable Diffusion stuff.| | 849|LLM-Red-Team/kimi-free-api !2025-03-2844004|🚀 KIMI AI 长文本大模型白嫖服务,支持高速流式输出、联网搜索、长文档解读、图像解析、多轮对话,零配置部署,多路token支持,自动清理会话痕迹。| | 850|argilla-io/argilla !2025-03-2843991|Argilla is a collaboration platform for AI engineers and domain experts that require high-quality outputs, full data ownership, and overall efficiency.| | 851|spring-projects/spring-ai !2025-03-28438419|An Application Framework for AI Engineering| | 852|alibaba-damo-academy/FunClip !2025-03-2843555|Open-source, accurate and easy-to-use video clipping tool, LLM based AI clipping intergrated | | 853|yisol/IDM-VTON !2025-03-2843541|IDM-VTON : Improving Diffusion Models for Authentic Virtual Try-on in the Wild| | 854|fchollet/ARC-AGI !2025-03-2843368|The Abstraction and Reasoning Corpus| | 855|MahmoudAshraf97/whisper-diarization !2025-03-2843064|Automatic Speech Recognition with Speaker Diarization based on OpenAI Whisper| | 856|Speykious/cve-rs !2025-03-2843047|Blazingly 🔥 fast 🚀 memory vulnerabilities, written in 100% safe Rust. 🦀| | 857|Blealtan/efficient-kan !2025-03-2842770|An efficient pure-PyTorch implementation of Kolmogorov-Arnold Network (KAN).| | 858|smol-ai/GodMode !2025-03-284249-1|AI Chat Browser: Fast, Full webapp access to ChatGPT / Claude / Bard / Bing / Llama2! I use this 20 times a day.| | 859|openai/plugins-quickstart !2025-03-284235-4 |Get a ChatGPT plugin up and running in under 5 minutes!| | 860|Doriandarko/maestro !2025-03-2842260|A framework for Claude Opus to intelligently orchestrate subagents.| | 861|philz1337x/clarity-upscaler !2025-03-2842204|Clarity-Upscaler: Reimagined image upscaling for everyone| | 862|facebookresearch/co-tracker !2025-03-2842142|CoTracker is a model for tracking any point (pixel) on a video.| | 863|xlang-ai/OpenAgents !2025-03-2842031|OpenAgents: An Open Platform for Language Agents in the Wild| | 864|alibaba/higress !2025-03-28419514|🤖 AI Gateway | | 865|ray-project/llm-numbers !2025-03-2841920 |Numbers every LLM developer should know| | 866|fudan-generative-vision/champ !2025-03-2841820|Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance| | 867|NVIDIA/garak !2025-03-2841795|the LLM vulnerability scanner| | 868|leetcode-mafia/cheetah !2025-03-2841740 |Whisper & GPT-based app for passing remote SWE interviews| | 869|ragapp/ragapp !2025-03-2841710|The easiest way to use Agentic RAG in any enterprise| | 870|collabora/WhisperSpeech !2025-03-2841692|An Open Source text-to-speech system built by inverting Whisper.| | 871|Facico/Chinese-Vicuna !2025-03-2841520 |Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model| | 872|openai/grok !2025-03-2841381|| | 873|CrazyBoyM/llama3-Chinese-chat !2025-03-2841361|Llama3 Chinese Repository with modified versions, and training and deployment resources| | 874|luban-agi/Awesome-AIGC-Tutorials !2025-03-2841301|Curated tutorials and resources for Large Language Models, AI Painting, and more.| | 875|damo-vilab/AnyDoor !2025-03-2841192|Official implementations for paper: Anydoor: zero-shot object-level image customization| | 876|raspberrypi/pico-sdk !2025-03-2841072|| | 877|mshumer/gpt-llm-trainer !2025-03-284097-1|| | 878|metavoiceio/metavoice-src !2025-03-284076-1|AI for human-level speech intelligence| | 879|intelowlproject/IntelOwl !2025-03-2840763|IntelOwl: manage your Threat Intelligence at scale| | 880|a16z-infra/ai-getting-started !2025-03-2840682|A Javascript AI getting started stack for weekend projects, including image/text models, vector stores, auth, and deployment configs| | 881|MarkFzp/mobile-aloha !2025-03-2840641|Mobile ALOHA: Learning Bimanual Mobile Manipulation with Low-Cost Whole-Body Teleoperation| | 882| keijiro/AICommand !2025-03-2840380 | ChatGPT integration with Unity Editor | | 883|Tencent/HunyuanDiT !2025-03-2840214|Hunyuan-DiT : A Powerful Multi-Resolution Diffusion Transformer with Fine-Grained Chinese Understanding| | 884|hengyoush/kyanos !2025-03-2840061|Visualize the time packets spend in the kernel, watch & analyze in command line.| | 885|agiresearch/AIOS !2025-03-2840045|AIOS: LLM Agent Operating System| | 886|truefoundry/cognita !2025-03-2839773|RAG (Retrieval Augmented Generation) Framework for building modular, open source applications for production by TrueFoundry| | 887|X-PLUG/MobileAgent !2025-03-2839557|Mobile-Agent: Autonomous Multi-Modal Mobile Device Agent with Visual Perception| | 888|jackMort/ChatGPT.nvim !2025-03-2839231|ChatGPT Neovim Plugin: Effortless Natural Language Generation with OpenAI's ChatGPT API| | 889|microsoft/RD-Agent !2025-03-28388422|Research and development (R&D) is crucial for the enhancement of industrial productivity, especially in the AI era, where the core aspects of R&D are mainly focused on data and models. We are committed to automate these high-value generic R&D processes through our open source R&D automation tool RD-Agent, which let AI drive data-driven AI.| | 890|Significant-Gravitas/Auto-GPT-Plugins !2025-03-283882-1 |Plugins for Auto-GPT| | 891|apple/ml-mgie !2025-03-2838770|| | 892|OpenDriveLab/UniAD !2025-03-2838727|[CVPR 2023 Best Paper] Planning-oriented Autonomous Driving| | 893|llSourcell/DoctorGPT !2025-03-2838640|DoctorGPT is an LLM that can pass the US Medical Licensing Exam. It works offline, it's cross-platform, & your health data stays private.| | 894|FlagAI-Open/FlagAI !2025-03-2838601|FlagAI (Fast LArge-scale General AI models) is a fast, easy-to-use and extensible toolkit for large-scale model.| | 895|krishnaik06/Roadmap-To-Learn-Generative-AI-In-2024 !2025-03-2838513|Roadmap To Learn Generative AI In 2024| | 896|SysCV/sam-hq !2025-03-2838491|Segment Anything in High Quality| | 897|google/security-research !2025-03-2838420|This project hosts security advisories and their accompanying proof-of-concepts related to research conducted at Google which impact non-Google owned code.| | 898|shroominic/codeinterpreter-api !2025-03-2838330|Open source implementation of the ChatGPT Code Interpreter 👾| | 899|Yonom/assistant-ui !2025-03-2838308|React Components for AI Chat 💬 🚀| | 900|nucleuscloud/neosync !2025-03-2838262|Open source data anonymization and synthetic data orchestration for developers. Create high fidelity synthetic data and sync it across your environments.| | 901|ravenscroftj/turbopilot !2025-03-2838230 |Turbopilot is an open source large-language-model based code completion engine that runs locally on CPU| | 902|NVlabs/Sana !2025-03-28380810|SANA: Efficient High-Resolution Image Synthesis with Linear Diffusion Transformer| | 903|huggingface/distil-whisper !2025-03-2838061|Distilled variant of Whisper for speech recognition. 6x faster, 50% smaller, within 1% word error rate.| | 904|Codium-ai/AlphaCodium !2025-03-2837971|code generation tool that surpasses most human competitors in CodeContests| | 905|fixie-ai/ultravox !2025-03-2837710|A fast multimodal LLM for real-time voice| | 906|unit-mesh/auto-dev !2025-03-28375715|🧙‍AutoDev: The AI-powered coding wizard with multilingual support 🌐, auto code generation 🏗️, and a helpful bug-slaying assistant 🐞! Customizable prompts 🎨 and a magic Auto Dev/Testing/Document/Agent feature 🧪 included! 🚀| | 907|Marker-Inc-Korea/AutoRAG !2025-03-2837432|AutoML tool for RAG| | 908|deepseek-ai/DeepSeek-VL !2025-03-283734-1|DeepSeek-VL: Towards Real-World Vision-Language Understanding| | 909|hiyouga/ChatGLM-Efficient-Tuning !2025-03-283692-1|Fine-tuning ChatGLM-6B with PEFT | | 910| Yue-Yang/ChatGPT-Siri !2025-03-2836921 | Shortcuts for Siri using ChatGPT API gpt-3.5-turbo model | | 911|0hq/WebGPT !2025-03-2836901 |Run GPT model on the browser with WebGPU. An implementation of GPT inference in less than ~2000 lines of vanilla Javascript.| | 912|cvg/LightGlue !2025-03-2836903|LightGlue: Local Feature Matching at Light Speed (ICCV 2023)| | 913|deanxv/coze-discord-proxy !2025-03-2836791|代理Discord-Bot对话Coze-Bot,实现API形式请求GPT4对话模型/微调模型| | 914|MervinPraison/PraisonAI !2025-03-2836764|PraisonAI application combines AutoGen and CrewAI or similar frameworks into a low-code solution for building and managing multi-agent LLM systems, focusing on simplicity, customisation, and efficient human-agent collaboration.| | 915|Ironclad/rivet !2025-03-2836345 |The open-source visual AI programming environment and TypeScript library| | 916|BasedHardware/OpenGlass !2025-03-2835851|Turn any glasses into AI-powered smart glasses| | 917|ricklamers/gpt-code-ui !2025-03-2835840 |An open source implementation of OpenAI's ChatGPT Code interpreter| | 918|whoiskatrin/chart-gpt !2025-03-2835830 |AI tool to build charts based on text input| | 919|github/CopilotForXcode !2025-03-2835788|Xcode extension for GitHub Copilot| | 920|hemansnation/God-Level-Data-Science-ML-Full-Stack !2025-03-2835570 |A collection of scientific methods, processes, algorithms, and systems to build stories & models. This roadmap contains 16 Chapters, whether you are a fresher in the field or an experienced professional who wants to transition into Data Science & AI| | 921|pytorch/torchchat !2025-03-2835461|Run PyTorch LLMs locally on servers, desktop and mobile| | 922| Kent0n-Li/ChatDoctor !2025-03-2835451 | A Medical Chat Model Fine-tuned on LLaMA Model using Medical Domain Knowledge | | 923|xtekky/chatgpt-clone !2025-03-283519-1 |ChatGPT interface with better UI| | 924|jupyterlab/jupyter-ai !2025-03-2835120|A generative AI extension for JupyterLab| | 925|pytorch/torchtitan !2025-03-2835064|A native PyTorch Library for large model training| | 926|minimaxir/simpleaichat !2025-03-2835031|Python package for easily interfacing with chat apps, with robust features and minimal code complexity.| | 927|srush/Tensor-Puzzles !2025-03-2834930|Solve puzzles. Improve your pytorch.| | 928|Helicone/helicone !2025-03-2834918|🧊 Open source LLM-Observability Platform for Developers. One-line integration for monitoring, metrics, evals, agent tracing, prompt management, playground, etc. Supports OpenAI SDK, Vercel AI SDK, Anthropic SDK, LiteLLM, LLamaIndex, LangChain, and more. 🍓 YC W23| | 929|run-llama/llama-hub !2025-03-2834740|A library of data loaders for LLMs made by the community -- to be used with LlamaIndex and/or LangChain| | 930|NExT-GPT/NExT-GPT !2025-03-2834700|Code and models for NExT-GPT: Any-to-Any Multimodal Large Language Model| | 931|souzatharsis/podcastfy !2025-03-2834661|An Open Source Python alternative to NotebookLM's podcast feature: Transforming Multimodal Content into Captivating Multilingual Audio Conversations with GenAI| | 932|Dataherald/dataherald !2025-03-2834450|Interact with your SQL database, Natural Language to SQL using LLMs| | 933|iryna-kondr/scikit-llm !2025-03-2834350 |Seamlessly integrate powerful language models like ChatGPT into scikit-learn for enhanced text analysis tasks.| | 934|Netflix/maestro !2025-03-2834230|Maestro: Netflix’s Workflow Orchestrator| | 935|CanadaHonk/porffor !2025-03-2833560|A from-scratch experimental AOT JS engine, written in JS| | 936|hustvl/Vim !2025-03-2833323|Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model| | 937|pashpashpash/vault-ai !2025-03-2833250 |OP Vault ChatGPT: Give ChatGPT long-term memory using the OP Stack (OpenAI + Pinecone Vector Database). Upload your own custom knowledge base files (PDF, txt, etc) using a simple React frontend.| | 938|tencentmusic/supersonic !2025-03-28330611|SuperSonic is the next-generation BI platform that integrates Chat BI (powered by LLM) and Headless BI (powered by semantic layer) paradigms.| | 939|billmei/every-chatgpt-gui !2025-03-2832981|Every front-end GUI client for ChatGPT| | 940|microsoft/torchgeo !2025-03-2832772|TorchGeo: datasets, samplers, transforms, and pre-trained models for geospatial data| | 941|LLMBook-zh/LLMBook-zh.github.io !2025-03-28326110|《大语言模型》作者:赵鑫,李军毅,周昆,唐天一,文继荣| | 942|dvlab-research/MiniGemini !2025-03-2832601|Official implementation for Mini-Gemini| | 943|rashadphz/farfalle !2025-03-2832460|🔍 AI search engine - self-host with local or cloud LLMs| | 944|Luodian/Otter !2025-03-2832450|🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.| | 945|AprilNEA/ChatGPT-Admin-Web !2025-03-2832370 | ChatGPT WebUI with user management and admin dashboard system| | 946|MarkFzp/act-plus-plus !2025-03-2832365|Imitation Learning algorithms with Co-traing for Mobile ALOHA: ACT, Diffusion Policy, VINN| | 947|ethen8181/machine-learning !2025-03-2832310|🌎 machine learning tutorials (mainly in Python3)| | 948|opengeos/segment-geospatial !2025-03-2832312 |A Python package for segmenting geospatial data with the Segment Anything Model (SAM)| | 949|iusztinpaul/hands-on-llms !2025-03-283225-2|🦖 𝗟𝗲𝗮𝗿𝗻 about 𝗟𝗟𝗠𝘀, 𝗟𝗟𝗠𝗢𝗽𝘀, and 𝘃𝗲𝗰𝘁𝗼𝗿 𝗗𝗕𝘀 for free by designing, training, and deploying a real-time financial advisor LLM system ~ 𝘴𝘰𝘶𝘳𝘤𝘦 𝘤𝘰𝘥𝘦 + 𝘷𝘪𝘥𝘦𝘰 & 𝘳𝘦𝘢𝘥𝘪𝘯𝘨 𝘮𝘢𝘵𝘦𝘳𝘪𝘢𝘭𝘴| | 950|ToTheBeginning/PuLID !2025-03-2832221|Official code for PuLID: Pure and Lightning ID Customization via Contrastive Alignment| | 951|neo4j-labs/llm-graph-builder !2025-03-2832164|Neo4j graph construction from unstructured data using LLMs| | 952|OpenGVLab/InternGPT !2025-03-2832150 |InternGPT (iGPT) is an open source demo platform where you can easily showcase your AI models. Now it supports DragGAN, ChatGPT, ImageBind, multimodal chat like GPT-4, SAM, interactive image editing, etc. Try it at igpt.opengvlab.com (支持DragGAN、ChatGPT、ImageBind、SAM的在线Demo系统)| | 953|PKU-YuanGroup/Video-LLaVA !2025-03-2832060 |Video-LLaVA: Learning United Visual Representation by Alignment Before Projection| | 954|DataTalksClub/llm-zoomcamp !2025-03-2832030|LLM Zoomcamp - a free online course about building an AI bot that can answer questions about your knowledge base| | 955|gptscript-ai/gptscript !2025-03-2832010|Natural Language Programming| |!green-up-arrow.svg 956|isaac-sim/IsaacLab !2025-03-28320113|Unified framework for robot learning built on NVIDIA Isaac Sim| |!red-down-arrow 957|ai-boost/Awesome-GPTs !2025-03-2832003|Curated list of awesome GPTs 👍.| | 958|huggingface/safetensors !2025-03-2831901|Simple, safe way to store and distribute tensors| | 959|linyiLYi/bilibot !2025-03-2831771|A local chatbot fine-tuned by bilibili user comments.| | 960| project-baize/baize-chatbot !2025-03-283168-1 | Let ChatGPT teach your own chatbot in hours with a single GPU! | | 961|Azure-Samples/cognitive-services-speech-sdk !2025-03-2831280|Sample code for the Microsoft Cognitive Services Speech SDK| | 962|microsoft/Phi-3CookBook !2025-03-2831231|This is a Phi-3 book for getting started with Phi-3. Phi-3, a family of open AI models developed by Microsoft. Phi-3 models are the most capable and cost-effective small language models (SLMs) available, outperforming models of the same size and next size up across a variety of language, reasoning, coding, and math benchmarks.| | 963|neuralmagic/deepsparse !2025-03-2831180|Sparsity-aware deep learning inference runtime for CPUs| | 964|sugarforever/chat-ollama !2025-03-2831000|ChatOllama is an open source chatbot based on LLMs. It supports a wide range of language models, and knowledge base management.| | 965|amazon-science/chronos-forecasting !2025-03-2830974|Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting| | 966|damo-vilab/i2vgen-xl !2025-03-2830902|Official repo for VGen: a holistic video generation ecosystem for video generation building on diffusion models| | 967|google-deepmind/gemma !2025-03-2830733|Open weights LLM from Google DeepMind.| | 968|iree-org/iree !2025-03-2830733|A retargetable MLIR-based machine learning compiler and runtime toolkit.| | 969|NVlabs/VILA !2025-03-2830724|VILA - a multi-image visual language model with training, inference and evaluation recipe, deployable from cloud to edge (Jetson Orin and laptops)| | 970|microsoft/torchscale !2025-03-2830661|Foundation Architecture for (M)LLMs| | 971|openai/openai-realtime-console !2025-03-2830656|React app for inspecting, building and debugging with the Realtime API| | 972|daveshap/OpenAIAgentSwarm !2025-03-2830610|HAAS = Hierarchical Autonomous Agent Swarm - "Resistance is futile!"| | 973|microsoft/PromptWizard !2025-03-2830555|Task-Aware Agent-driven Prompt Optimization Framework| | 974|CVI-SZU/Linly !2025-03-2830490 |Chinese-LLaMA basic model; ChatFlow Chinese conversation model; NLP pre-training/command fine-tuning dataset| | 975|cohere-ai/cohere-toolkit !2025-03-2830130|Toolkit is a collection of prebuilt components enabling users to quickly build and deploy RAG applications.| | 976|adamcohenhillel/ADeus !2025-03-2830131|An open source AI wearable device that captures what you say and hear in the real world and then transcribes and stores it on your own server. You can then chat with Adeus using the app, and it will have all the right context about what you want to talk about - a truly personalized, personal AI.| | 977|Lightning-AI/LitServe !2025-03-2830132|Lightning-fast serving engine for AI models. Flexible. Easy. Enterprise-scale.| | 978|potpie-ai/potpie !2025-03-2829973|Prompt-To-Agent : Create custom engineering agents for your codebase| | 979|ant-design/x !2025-03-28299529|Craft AI-driven interfaces effortlessly 🤖| | 980|meta-llama/PurpleLlama !2025-03-2829832|Set of tools to assess and improve LLM security.| | 981|williamyang1991/RerenderAVideo !2025-03-2829800|[SIGGRAPH Asia 2023] Rerender A Video: Zero-Shot Text-Guided Video-to-Video Translation| | 982|baichuan-inc/Baichuan-13B !2025-03-2829790|A 13B large language model developed by Baichuan Intelligent Technology| | 983|Stability-AI/stable-audio-tools !2025-03-2829761|Generative models for conditional audio generation| | 984|li-plus/chatglm.cpp !2025-03-2829720|C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & more LLMs| | 985|NVIDIA/GenerativeAIExamples !2025-03-2829546|Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.| | 986|Josh-XT/AGiXT !2025-03-2829521 |AGiXT is a dynamic AI Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.| | 987|MrForExample/ComfyUI-3D-Pack !2025-03-2829515|An extensive node suite that enables ComfyUI to process 3D inputs (Mesh & UV Texture, etc) using cutting edge algorithms (3DGS, NeRF, etc.)| | 988|olimorris/codecompanion.nvim !2025-03-28295111|✨ AI-powered coding, seamlessly in Neovim. Supports Anthropic, Copilot, Gemini, Ollama, OpenAI and xAI LLMs| | 989|salesforce/CodeT5 !2025-03-282940-1 |Home of CodeT5: Open Code LLMs for Code Understanding and Generation| | 990|facebookresearch/ijepa !2025-03-2829391|Official codebase for I-JEPA, the Image-based Joint-Embedding Predictive Architecture. First outlined in the CVPR paper, "Self-supervised learning from images with a joint-embedding predictive architecture."| | 991|eureka-research/Eureka !2025-03-2829351|Official Repository for "Eureka: Human-Level Reward Design via Coding Large Language Models"| | 992|NVIDIA/trt-llm-rag-windows !2025-03-282934-1|A developer reference project for creating Retrieval Augmented Generation (RAG) chatbots on Windows using TensorRT-LLM| | 993|gmpetrov/databerry !2025-03-282930-1|The no-code platform for building custom LLM Agents| | 994|AI4Finance-Foundation/FinRobot !2025-03-28291946|FinRobot: An Open-Source AI Agent Platform for Financial Applications using LLMs 🚀 🚀 🚀| | 995|nus-apr/auto-code-rover !2025-03-2829013|A project structure aware autonomous software engineer aiming for autonomous program improvement| | 996|deepseek-ai/DreamCraft3D !2025-03-2828921|[ICLR 2024] Official implementation of DreamCraft3D: Hierarchical 3D Generation with Bootstrapped Diffusion Prior| | 997|mlabonne/llm-datasets !2025-03-2828848|High-quality datasets, tools, and concepts for LLM fine-tuning.| | 998|facebookresearch/jepa !2025-03-2828712|PyTorch code and models for V-JEPA self-supervised learning from video.| | 999|facebookresearch/habitat-sim !2025-03-2828604|A flexible, high-performance 3D simulator for Embodied AI research.| | 1000|xenova/whisper-web !2025-03-2828581|ML-powered speech recognition directly in your browser| | 1001|cvlab-columbia/zero123 !2025-03-2828530|Zero-1-to-3: Zero-shot One Image to 3D Object: https://zero123.cs.columbia.edu/| | 1002|yuruotong1/autoMate !2025-03-28285121|Like Manus, Computer Use Agent(CUA) and Omniparser, we are computer-using agents.AI-driven local automation assistant that uses natural language to make computers work by themselves| | 1003|muellerberndt/mini-agi !2025-03-282845-1 |A minimal generic autonomous agent based on GPT3.5/4. Can analyze stock prices, perform network security tests, create art, and order pizza.| | 1004|allenai/open-instruct !2025-03-2828432|| | 1005|CodingChallengesFYI/SharedSolutions !2025-03-2828360|Publicly shared solutions to Coding Challenges| | 1006|hegelai/prompttools !2025-03-2828220|Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate).| | 1007|mazzzystar/Queryable !2025-03-2828222|Run CLIP on iPhone to Search Photos.| | 1008|Doubiiu/DynamiCrafter !2025-03-2828173|DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors| | 1009|SamurAIGPT/privateGPT !2025-03-282805-1 |An app to interact privately with your documents using the power of GPT, 100% privately, no data leaks| | 1010|facebookresearch/Pearl !2025-03-2827951|A Production-ready Reinforcement Learning AI Agent Library brought by the Applied Reinforcement Learning team at Meta.| | 1011|intuitem/ciso-assistant-community !2025-03-2827954|CISO Assistant is a one-stop-shop for GRC, covering Risk, AppSec and Audit Management and supporting +70 frameworks worldwide with auto-mapping: NIST CSF, ISO 27001, SOC2, CIS, PCI DSS, NIS2, CMMC, PSPF, GDPR, HIPAA, Essential Eight, NYDFS-500, DORA, NIST AI RMF, 800-53, 800-171, CyFun, CJIS, AirCyber, NCSC, ECC, SCF and so much more| | 1012|facebookresearch/audio2photoreal !2025-03-2827840|Code and dataset for photorealistic Codec Avatars driven from audio| | 1013|Azure/azure-rest-api-specs !2025-03-2827770|The source for REST API specifications for Microsoft Azure.| | 1014|SCUTlihaoyu/open-chat-video-editor !2025-03-2827690 |Open source short video automatic generation tool| | 1015|Alpha-VLLM/LLaMA2-Accessory !2025-03-2827642|An Open-source Toolkit for LLM Development| | 1016|johnma2006/mamba-minimal !2025-03-2827601|Simple, minimal implementation of the Mamba SSM in one file of PyTorch.| | 1017|nerfstudio-project/gsplat !2025-03-2827576|CUDA accelerated rasterization of gaussian splatting| | 1018|Physical-Intelligence/openpi !2025-03-28274617|| | 1019|leptonai/leptonai !2025-03-2827246|A Pythonic framework to simplify AI service building| |!green-up-arrow.svg 1020|joanrod/star-vector !2025-03-28271149|StarVector is a foundation model for SVG generation that transforms vectorization into a code generation task. Using a vision-language modeling architecture, StarVector processes both visual and textual inputs to produce high-quality SVG code with remarkable precision.| |!red-down-arrow 1021|jqnatividad/qsv !2025-03-2827092|CSVs sliced, diced & analyzed.| | 1022|FranxYao/chain-of-thought-hub !2025-03-2826991|Benchmarking large language models' complex reasoning ability with chain-of-thought prompting| | 1023|princeton-nlp/SWE-bench !2025-03-2826965|[ICLR 2024] SWE-Bench: Can Language Models Resolve Real-world Github Issues?| | 1024|elastic/otel-profiling-agent !2025-03-2826930|The production-scale datacenter profiler| | 1025|src-d/hercules !2025-03-2826900|Gaining advanced insights from Git repository history.| | 1026|lanqian528/chat2api !2025-03-2826695|A service that can convert ChatGPT on the web to OpenAI API format.| | 1027|ishan0102/vimGPT !2025-03-2826681|Browse the web with GPT-4V and Vimium| | 1028|TMElyralab/MuseV !2025-03-2826650|MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising| | 1029|georgia-tech-db/eva !2025-03-2826600 |AI-Relational Database System | | 1030|kubernetes-sigs/controller-runtime !2025-03-2826590|Repo for the controller-runtime subproject of kubebuilder (sig-apimachinery)| | 1031|gptlink/gptlink !2025-03-2826550 |Build your own free commercial ChatGPT environment in 10 minutes. The setup is simple and includes features such as user management, orders, tasks, and payments| | 1032|pytorch/executorch !2025-03-2826534|On-device AI across mobile, embedded and edge for PyTorch| | 1033|NVIDIA/nv-ingest !2025-03-2826290|NVIDIA Ingest is an early access set of microservices for parsing hundreds of thousands of complex, messy unstructured PDFs and other enterprise documents into metadata and text to embed into retrieval systems.| | 1034|SuperTux/supertux !2025-03-2826081|SuperTux source code| | 1035|abi/secret-llama !2025-03-2826050|Fully private LLM chatbot that runs entirely with a browser with no server needed. Supports Mistral and LLama 3.| | 1036|liou666/polyglot !2025-03-2825841 |Desktop AI Language Practice Application| | 1037|janhq/nitro !2025-03-2825821|A fast, lightweight, embeddable inference engine to supercharge your apps with local AI. OpenAI-compatible API| | 1038|deepseek-ai/DeepSeek-Math !2025-03-2825825|DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models| | 1039|anthropics/prompt-eng-interactive-tutorial !2025-03-2825781|Anthropic's Interactive Prompt Engineering Tutorial| | 1040|microsoft/promptbench !2025-03-2825741|A unified evaluation framework for large language models| | 1041|baaivision/Painter !2025-03-2825580 |Painter & SegGPT Series: Vision Foundation Models from BAAI| | 1042|OpenPipe/OpenPipe !2025-03-2825581|Turn expensive prompts into cheap fine-tuned models| | 1043|TracecatHQ/tracecat !2025-03-2825531|😼 The AI-native, open source alternative to Tines / Splunk SOAR.| | 1044|JoshuaC215/agent-service-toolkit !2025-03-2825528|Full toolkit for running an AI agent service built with LangGraph, FastAPI and Streamlit| | 1045|databricks/dbrx !2025-03-2825460|Code examples and resources for DBRX, a large language model developed by Databricks| | 1046|lamini-ai/lamini !2025-03-2825271 |Official repo for Lamini's data generator for generating instructions to train instruction-following LLMs| | 1047|mshumer/gpt-author !2025-03-282510-1|| | 1048|TMElyralab/MusePose !2025-03-2824971|MusePose: a Pose-Driven Image-to-Video Framework for Virtual Human Generation| | 1049|Kludex/fastapi-tips !2025-03-2824974|FastAPI Tips by The FastAPI Expert!| | 1050|openai/simple-evals !2025-03-2824813|| | 1051|iterative/datachain !2025-03-2824732|AI-data warehouse to enrich, transform and analyze data from cloud storages| | 1052|girafe-ai/ml-course !2025-03-2824703|Open Machine Learning course| | 1053|kevmo314/magic-copy !2025-03-2824620 |Magic Copy is a Chrome extension that uses Meta's Segment Anything Model to extract a foreground object from an image and copy it to the clipboard.| | 1054|Eladlev/AutoPrompt !2025-03-2824432|A framework for prompt tuning using Intent-based Prompt Calibration| | 1055|OpenBMB/CPM-Bee !2025-03-282434-1 |A bilingual large-scale model with trillions of parameters| | 1056|IDEA-Research/T-Rex !2025-03-2824310|T-Rex2: Towards Generic Object Detection via Text-Visual Prompt Synergy| | 1057|microsoft/genaiscript !2025-03-2824202|Automatable GenAI Scripting| | 1058|paulpierre/RasaGPT !2025-03-2824090 |💬 RasaGPT is the first headless LLM chatbot platform built on top of Rasa and Langchain. Built w/ Rasa, FastAPI, Langchain, LlamaIndex, SQLModel, pgvector, ngrok, telegram| | 1059|ashishpatel26/LLM-Finetuning !2025-03-2823911|LLM Finetuning with peft| | 1060|SoraWebui/SoraWebui !2025-03-2823570|SoraWebui is an open-source Sora web client, enabling users to easily create videos from text with OpenAI's Sora model.| | 1061|6drf21e/ChatTTScolab !2025-03-2823491|🚀 一键部署(含离线整合包)!基于 ChatTTS ,支持音色抽卡、长音频生成和分角色朗读。简单易用,无需复杂安装。| | 1062|Azure/PyRIT !2025-03-2823343|The Python Risk Identification Tool for generative AI (PyRIT) is an open access automation framework to empower security professionals and machine learning engineers to proactively find risks in their generative AI systems.| | 1063|tencent-ailab/V-Express !2025-03-2823201|V-Express aims to generate a talking head video under the control of a reference image, an audio, and a sequence of V-Kps images.| | 1064|THUDM/CogVLM2 !2025-03-2823170|GPT4V-level open-source multi-modal model based on Llama3-8B| | 1065|dvmazur/mixtral-offloading !2025-03-2823001|Run Mixtral-8x7B models in Colab or consumer desktops| | 1066|semanser/codel !2025-03-2822950|✨ Fully autonomous AI Agent that can perform complicated tasks and projects using terminal, browser, and editor.| | 1067|mshumer/gpt-investor !2025-03-2822590|| | 1068|aixcoder-plugin/aiXcoder-7B !2025-03-2822550|official repository of aiXcoder-7B Code Large Language Model| | 1069|Azure-Samples/graphrag-accelerator !2025-03-2822503|One-click deploy of a Knowledge Graph powered RAG (GraphRAG) in Azure| | 1070|emcf/engshell !2025-03-2821830 |An English-language shell for any OS, powered by LLMs| | 1071|hncboy/chatgpt-web-java !2025-03-2821771|ChatGPT project developed in Java, based on Spring Boot 3 and JDK 17, supports both AccessToken and ApiKey modes| | 1072|openai/consistencydecoder !2025-03-2821692|Consistency Distilled Diff VAE| | 1073|Alpha-VLLM/Lumina-T2X !2025-03-2821681|Lumina-T2X is a unified framework for Text to Any Modality Generation| | 1074|bghira/SimpleTuner !2025-03-2821612|A general fine-tuning kit geared toward Stable Diffusion 2.1, Stable Diffusion 3, DeepFloyd, and SDXL.| | 1075|JiauZhang/DragGAN !2025-03-2821530 |Implementation of DragGAN: Interactive Point-based Manipulation on the Generative Image Manifold| | 1076|cgpotts/cs224u !2025-03-2821390|Code for Stanford CS224u| | 1077|PKU-YuanGroup/MoE-LLaVA !2025-03-2821300|Mixture-of-Experts for Large Vision-Language Models| | 1078|darrenburns/elia !2025-03-2820831|A snappy, keyboard-centric terminal user interface for interacting with large language models. Chat with ChatGPT, Claude, Llama 3, Phi 3, Mistral, Gemma and more.| | 1079|ageerle/ruoyi-ai !2025-03-28207898|RuoYi AI 是一个全栈式 AI 开发平台,旨在帮助开发者快速构建和部署个性化的 AI 应用。| | 1080|NVIDIA/gpu-operator !2025-03-2820510|NVIDIA GPU Operator creates/configures/manages GPUs atop Kubernetes| | 1081|BAAI-Agents/Cradle !2025-03-2820481|The Cradle framework is a first attempt at General Computer Control (GCC). Cradle supports agents to ace any computer task by enabling strong reasoning abilities, self-improvment, and skill curation, in a standardized general environment with minimal requirements.| | 1082|microsoft/aici !2025-03-2820080|AICI: Prompts as (Wasm) Programs| | 1083|PRIS-CV/DemoFusion !2025-03-2820040|Let us democratise high-resolution generation! (arXiv 2023)| | 1084|apple/axlearn !2025-03-2820012|An Extensible Deep Learning Library| | 1085|naver/mast3r !2025-03-2819685|Grounding Image Matching in 3D with MASt3R| | 1086|liltom-eth/llama2-webui !2025-03-281958-1|Run Llama 2 locally with gradio UI on GPU or CPU from anywhere (Linux/Windows/Mac). Supporting Llama-2-7B/13B/70B with 8-bit, 4-bit. Supporting GPU inference (6 GB VRAM) and CPU inference.| | 1087|GaParmar/img2img-turbo !2025-03-2819582|One-step image-to-image with Stable Diffusion turbo: sketch2image, day2night, and more| | 1088|Niek/chatgpt-web !2025-03-2819560|ChatGPT web interface using the OpenAI API| | 1089|huggingface/cookbook !2025-03-2819421|Open-source AI cookbook| | 1090|pytorch/ao !2025-03-2819241|PyTorch native quantization and sparsity for training and inference| | 1091|emcie-co/parlant !2025-03-2819053|The behavior guidance framework for customer-facing LLM agents| | 1092|ymcui/Chinese-LLaMA-Alpaca-3 !2025-03-2818980|中文羊驼大模型三期项目 (Chinese Llama-3 LLMs) developed from Meta Llama 3| | 1093|Nutlope/notesGPT !2025-03-2818811|Record voice notes & transcribe, summarize, and get tasks| | 1094|InstantStyle/InstantStyle !2025-03-2818791|InstantStyle: Free Lunch towards Style-Preserving in Text-to-Image Generation 🔥| | 1095|idaholab/moose !2025-03-2818771|Multiphysics Object Oriented Simulation Environment| | 1096|The-OpenROAD-Project/OpenROAD !2025-03-2818351|OpenROAD's unified application implementing an RTL-to-GDS Flow. Documentation at https://openroad.readthedocs.io/en/latest/| | 1097|alibaba/spring-ai-alibaba !2025-03-281831121|Agentic AI Framework for Java Developers| | 1098|ytongbai/LVM !2025-03-2817990|Sequential Modeling Enables Scalable Learning for Large Vision Models| | 1099|microsoft/sample-app-aoai-chatGPT !2025-03-2817981|[PREVIEW] Sample code for a simple web chat experience targeting chatGPT through AOAI.| | 1100|AI-Citizen/SolidGPT !2025-03-2817830|Chat everything with your code repository, ask repository level code questions, and discuss your requirements. AI Scan and learning your code repository, provide you code repository level answer🧱 🧱| | 1101|YangLing0818/RPG-DiffusionMaster !2025-03-2817784|Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs (PRG)| | 1102|kyegomez/BitNet !2025-03-2817710|Implementation of "BitNet: Scaling 1-bit Transformers for Large Language Models" in pytorch| | 1103|eloialonso/diamond !2025-03-2817671|DIAMOND (DIffusion As a Model Of eNvironment Dreams) is a reinforcement learning agent trained in a diffusion world model.| | 1104|flowdriveai/flowpilot !2025-03-2817250|flow-pilot is an openpilot based driver assistance system that runs on linux, windows and android powered machines.| | 1105|xlang-ai/OSWorld !2025-03-2817200|OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments| | 1106|linyiLYi/snake-ai !2025-03-2817031|An AI agent that beats the classic game "Snake".| | 1107|baaivision/Emu !2025-03-2816991|Emu Series: Generative Multimodal Models from BAAI| | 1108|kevmo314/scuda !2025-03-2816870|SCUDA is a GPU over IP bridge allowing GPUs on remote machines to be attached to CPU-only machines.| | 1109|SharifiZarchi/IntroductiontoMachineLearning !2025-03-2816701|دوره‌ی مقدمه‌ای بر یادگیری ماشین، برای دانشجویان| | 1110|google/maxtext !2025-03-2816670|A simple, performant and scalable Jax LLM!| | 1111|ml-explore/mlx-swift-examples !2025-03-2816471|Examples using MLX Swift| | 1112|unitreerobotics/unitreerlgym !2025-03-2816256|| | 1113|collabora/WhisperFusion !2025-03-2815901|WhisperFusion builds upon the capabilities of WhisperLive and WhisperSpeech to provide a seamless conversations with an AI.| | 1114|lichao-sun/Mora !2025-03-2815520|Mora: More like Sora for Generalist Video Generation| | 1115|GoogleCloudPlatform/localllm !2025-03-2815370|Run LLMs locally on Cloud Workstations| | 1116|TencentARC/BrushNet !2025-03-2815330|The official implementation of paper "BrushNet: A Plug-and-Play Image Inpainting Model with Decomposed Dual-Branch Diffusion"| | 1117|ai-christianson/RA.Aid !2025-03-2815288|Develop software autonomously.| | 1118|stephansturges/WALDO !2025-03-2815170|Whereabouts Ascertainment for Low-lying Detectable Objects. The SOTA in FOSS AI for drones!| | 1119|skills/copilot-codespaces-vscode !2025-03-2815112|Develop with AI-powered code suggestions using GitHub Copilot and VS Code| | 1120|andrewnguonly/Lumos !2025-03-2814920|A RAG LLM co-pilot for browsing the web, powered by local LLMs| | 1121|TeamNewPipe/NewPipeExtractor !2025-03-2814811|NewPipe's core library for extracting data from streaming sites| | 1122|mhamilton723/FeatUp !2025-03-2814770|Official code for "FeatUp: A Model-Agnostic Frameworkfor Features at Any Resolution" ICLR 2024| | 1123|AnswerDotAI/fsdpqlora !2025-03-2814671|Training LLMs with QLoRA + FSDP| | 1124|jgravelle/AutoGroq !2025-03-2814330|| | 1125|OpenGenerativeAI/llm-colosseum !2025-03-2814130|Benchmark LLMs by fighting in Street Fighter 3! The new way to evaluate the quality of an LLM| | 1126|microsoft/vscode-ai-toolkit !2025-03-2814000|| | 1127|McGill-NLP/webllama !2025-03-2813930|Llama-3 agents that can browse the web by following instructions and talking to you| | 1128|lucidrains/self-rewarding-lm-pytorch !2025-03-2813760|Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI| | 1129|ishaan1013/sandbox !2025-03-2813650|A cloud-based code editing environment with an AI copilot and real-time collaboration.| | 1130|goatcorp/Dalamud !2025-03-2813275|FFXIV plugin framework and API| | 1131|Lightning-AI/lightning-thunder !2025-03-2813151|Make PyTorch models Lightning fast! Thunder is a source to source compiler for PyTorch. It enables using different hardware executors at once.| | 1132|PKU-YuanGroup/MagicTime !2025-03-2813052|MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators| | 1133|SakanaAI/evolutionary-model-merge !2025-03-2813000|Official repository of Evolutionary Optimization of Model Merging Recipes| | 1134|a-real-ai/pywinassistant !2025-03-2812950|The first open source Large Action Model generalist Artificial Narrow Intelligence that controls completely human user interfaces by only using natural language. PyWinAssistant utilizes Visualization-of-Thought Elicits Spatial Reasoning in Large Language Models.| | 1135|TraceMachina/nativelink !2025-03-2812630|NativeLink is an open source high-performance build cache and remote execution server, compatible with Bazel, Buck2, Reclient, and other RBE-compatible build systems. It offers drastically faster builds, reduced test flakiness, and significant infrastructure cost savings.| | 1136|MLSysOps/MLE-agent !2025-03-2812500|🤖 MLE-Agent: Your intelligent companion for seamless AI engineering and research. 🔍 Integrate with arxiv and paper with code to provide better code/research plans 🧰 OpenAI, Ollama, etc supported. 🎆 Code RAG| | 1137|wpilibsuite/allwpilib !2025-03-2811610|Official Repository of WPILibJ and WPILibC| | 1138|elfvingralf/macOSpilot-ai-assistant !2025-03-2811470|Voice + Vision powered AI assistant that answers questions about any application, in context and in audio.| | 1139|langchain-ai/langchain-extract !2025-03-2811210|🦜⛏️ Did you say you like data?| | 1140|FoundationVision/GLEE !2025-03-2811120|【CVPR2024】GLEE: General Object Foundation Model for Images and Videos at Scale| | 1141|Profluent-AI/OpenCRISPR !2025-03-2810990|AI-generated gene editing systems| | 1142|zju3dv/EasyVolcap !2025-03-2810821|[SIGGRAPH Asia 2023 (Technical Communications)] EasyVolcap: Accelerating Neural Volumetric Video Research| | 1143|PaddlePaddle/PaddleHelix !2025-03-2810560|Bio-Computing Platform Featuring Large-Scale Representation Learning and Multi-Task Deep Learning “螺旋桨”生物计算工具集| | 1144|myshell-ai/JetMoE !2025-03-289800|Reaching LLaMA2 Performance with 0.1M Dollars| | 1145|likejazz/llama3.np !2025-03-289770|llama3.np is pure NumPy implementation for Llama 3 model.| | 1146|mustafaaljadery/gemma-2B-10M !2025-03-289500|Gemma 2B with 10M context length using Infini-attention.| | 1147|HITsz-TMG/FilmAgent !2025-03-289382|Resources of our paper "FilmAgent: A Multi-Agent Framework for End-to-End Film Automation in Virtual 3D Spaces". New versions in the making!| | 1148|aws-samples/amazon-bedrock-samples !2025-03-289362|This repository contains examples for customers to get started using the Amazon Bedrock Service. This contains examples for all available foundational models| | 1149|Akkudoktor-EOS/EOS !2025-03-2893154|This repository features an Energy Optimization System (EOS) that optimizes energy distribution, usage for batteries, heat pumps& household devices. It includes predictive models for electricity prices (planned), load forecasting& dynamic optimization to maximize energy efficiency & minimize costs. Founder Dr. Andreas Schmitz (YouTube @akkudoktor)| Tip: | symbol| rule | | :----| :---- | |🔥 | 256 1k| |!green-up-arrow.svg !red-down-arrow | ranking up / down| |⭐ | on trending page today| [Back to Top] Tools | No. | Tool | Description | | ----:|:----------------------------------------------- |:------------------------------------------------------------------------------------------- | | 1 | ChatGPT | A sibling model to InstructGPT, which is trained to follow instructions in a prompt and provide a detailed response | | 2 | DALL·E 2 | Create original, realistic images and art from a text description | | 3 | Murf AI | AI enabled, real people's voices| | 4 | Midjourney | An independent research lab that produces an artificial intelligence program under the same name that creates images from textual descriptions, used in Discord | 5 | Make-A-Video | Make-A-Video is a state-of-the-art AI system that generates videos from text | | 6 | Creative Reality™ Studio by D-ID| Use generative AI to create future-facing videos| | 7 | chat.D-ID| The First App Enabling Face-to-Face Conversations with ChatGPT| | 8 | Notion AI| Access the limitless power of AI, right inside Notion. Work faster. Write better. Think bigger. | | 9 | Runway| Text to Video with Gen-2 | | 10 | Resemble AI| Resemble’s AI voice generator lets you create human–like voice overs in seconds | | 11 | Cursor| Write, edit, and chat about your code with a powerful AI | | 12 | Hugging Face| Build, train and deploy state of the art models powered by the reference open source in machine learning | | 13 | Claude | A next-generation AI assistant for your tasks, no matter the scale | | 14 | Poe| Poe lets you ask questions, get instant answers, and have back-and-forth conversations with AI. Gives access to GPT-4, gpt-3.5-turbo, Claude from Anthropic, and a variety of other bots| [Back to Top] Websites | No. | WebSite |Description | | ----:|:------------------------------------------ |:---------------------------------------------------------------------------------------- | | 1 | OpenAI | An artificial intelligence research lab | | 2 | Bard | Base Google's LaMDA chatbots and pull from internet | | 3 | ERNIE Bot | Baidu’s new generation knowledge-enhanced large language model is a new member of the Wenxin large model family | | 4 | DALL·E 2 | An AI system that can create realistic images and art from a description in natural language | | 5 | Whisper | A general-purpose speech recognition model | | 6| CivitAI| A platform that makes it easy for people to share and discover resources for creating AI art| | 7|D-ID| D-ID’s Generative AI enables users to transform any picture or video into extraordinary experiences| | 8| Nvidia eDiff-I| Text-to-Image Diffusion Models with Ensemble of Expert Denoisers | | 9| Stability AI| The world's leading open source generative AI company which opened source Stable Diffusion | | 10| Meta AI| Whether it be research, product or infrastructure development, we’re driven to innovate responsibly with AI to benefit the world | | 11| ANTHROPIC| AI research and products that put safety at the frontier | [Back to Top] Reports&Papers | No. | Report&Paper | Description | |:---- |:-------------------------------------------------------------------------------------------------------------- |:---------------------------------------------------- | | 1 | GPT-4 Technical Report | GPT-4 Technical Report | | 2 | mli/paper-reading | Deep learning classics and new papers are read carefully paragraph by paragraph. | | 3 | labmlai/annotateddeeplearningpaperimplementations| A collection of simple PyTorch implementations of neural networks and related algorithms, which are documented with explanations | | 4 | Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models | Talking, Drawing and Editing with Visual Foundation Models | | 5 | OpenAI Research | The latest research report and papers from OpenAI | | 6 | Make-A-Video: Text-to-Video Generation without Text-Video Data|Meta's Text-to-Video Generation| | 7 | eDiff-I: Text-to-Image Diffusion Models with Ensemble of Expert Denoisers| Nvidia eDiff-I - New generation of generative AI content creation tool | | 8 | Training an Assistant-style Chatbot with Large Scale Data Distillation from GPT-3.5-Turbo | 2023 GPT4All Technical Report | | 9 | Segment Anything| Meta Segment Anything | | 10 | LLaMA: Open and Efficient Foundation Language Models| LLaMA: a collection of foundation language models ranging from 7B to 65B parameters| | 11 | papers-we-love/papers-we-love |Papers from the computer science community to read and discuss| | 12 | CVPR 2023 papers |The most exciting and influential CVPR 2023 papers| [Back to Top] Tutorials | No. | Tutorial | Description| |:---- |:---------------------------------------------------------------- | --- | | 1 | Coursera - Machine Learning | The Machine Learning Specialization Course taught by Dr. Andrew Ng| | 2 | microsoft/ML-For-Beginners | 12 weeks, 26 lessons, 52 quizzes, classic Machine Learning for all| | 3 | ChatGPT Prompt Engineering for Developers | This short course taught by Isa Fulford (OpenAI) and Andrew Ng (DeepLearning.AI) will teach how to use a large language model (LLM) to quickly build new and powerful applications | | 4 | Dive into Deep Learning |Targeting Chinese readers, functional and open for discussion. The Chinese and English versions are used for teaching in over 400 universities across more than 60 countries | | 5 | AI Expert Roadmap | Roadmap to becoming an Artificial Intelligence Expert in 2022 | | 6 | Computer Science courses |List of Computer Science courses with video lectures| | 7 | Machine Learning with Python | Machine Learning with Python Certification on freeCodeCamp| | 8 | Building Systems with the ChatGPT API | This short course taught by Isa Fulford (OpenAI) and Andrew Ng (DeepLearning.AI), you will learn how to automate complex workflows using chain calls to a large language model| | 9 | LangChain for LLM Application Development | This short course taught by Harrison Chase (Co-Founder and CEO at LangChain) and Andrew Ng. you will gain essential skills in expanding the use cases and capabilities of language models in application development using the LangChain framework| | 10 | How Diffusion Models Work | This short course taught by Sharon Zhou (CEO, Co-founder, Lamini). you will gain a deep familiarity with the diffusion process and the models which carry it out. More than simply pulling in a pre-built model or using an API, this course will teach you to build a diffusion model from scratch| | 11 | Free Programming Books For AI |📚 Freely available programming books for AI | | 12 | microsoft/AI-For-Beginners |12 Weeks, 24 Lessons, AI for All!| | 13 | hemansnation/God-Level-Data-Science-ML-Full-Stack |A collection of scientific methods, processes, algorithms, and systems to build stories & models. This roadmap contains 16 Chapters, whether you are a fresher in the field or an experienced professional who wants to transition into Data Science & AI| | 14 | datawhalechina/prompt-engineering-for-developers |Chinese version of Andrew Ng's Big Model Series Courses, including "Prompt Engineering", "Building System", and "LangChain"| | 15 | ossu/computer-science |🎓 Path to a free self-taught education in Computer Science!| | 16 | microsoft/Data-Science-For-Beginners | 10 Weeks, 20 Lessons, Data Science for All! | |17 |jwasham/coding-interview-university !2023-09-29268215336 |A complete computer science study plan to become a software engineer.| [Back to Top] Thanks If this project has been helpful to you in any way, please give it a ⭐️ by clicking on the star.

aiortc
github
LLM Vibe Score0.507
Human Vibe Score0.11415188209660238
aiortcMar 28, 2025

aiortc

.. image:: docs/_static/aiortc.svg :width: 120px :alt: aiortc .. image:: https://img.shields.io/pypi/l/aiortc.svg :target: https://pypi.python.org/pypi/aiortc :alt: License .. image:: https://img.shields.io/pypi/v/aiortc.svg :target: https://pypi.python.org/pypi/aiortc :alt: Version .. image:: https://img.shields.io/pypi/pyversions/aiortc.svg :target: https://pypi.python.org/pypi/aiortc :alt: Python versions .. image:: https://github.com/aiortc/aiortc/workflows/tests/badge.svg :target: https://github.com/aiortc/aiortc/actions :alt: Tests .. image:: https://img.shields.io/codecov/c/github/aiortc/aiortc.svg :target: https://codecov.io/gh/aiortc/aiortc :alt: Coverage .. image:: https://readthedocs.org/projects/aiortc/badge/?version=latest :target: https://aiortc.readthedocs.io/ :alt: Documentation What is `aiortc? aiortc is a library for Web Real-Time Communication (WebRTC)_ and Object Real-Time Communication (ORTC)_ in Python. It is built on top of asyncio, Python's standard asynchronous I/O framework. The API closely follows its Javascript counterpart while using pythonic constructs: promises are replaced by coroutines events are emitted using pyee.EventEmitter To learn more about aiortc please read the documentation_. .. _Web Real-Time Communication (WebRTC): https://webrtc.org/ .. _Object Real-Time Communication (ORTC): https://ortc.org/ .. _read the documentation: https://aiortc.readthedocs.io/en/latest/ Why should I use aiortc? The main WebRTC and ORTC implementations are either built into web browsers, or come in the form of native code. While they are extensively battle tested, their internals are complex and they do not provide Python bindings. Furthermore they are tightly coupled to a media stack, making it hard to plug in audio or video processing algorithms. In contrast, the aiortc implementation is fairly simple and readable. As such it is a good starting point for programmers wishing to understand how WebRTC works or tinker with its internals. It is also easy to create innovative products by leveraging the extensive modules available in the Python ecosystem. For instance you can build a full server handling both signaling and data channels or apply computer vision algorithms to video frames using OpenCV. Furthermore, a lot of effort has gone into writing an extensive test suite for the aiortc code to ensure best-in-class code quality. Implementation status aiortc allows you to exchange audio, video and data channels and interoperability is regularly tested against both Chrome and Firefox. Here are some of its features: SDP generation / parsing Interactive Connectivity Establishment, with half-trickle and mDNS support DTLS key and certificate generation DTLS handshake, encryption / decryption (for SCTP) SRTP keying, encryption and decryption for RTP and RTCP Pure Python SCTP implementation Data Channels Sending and receiving audio (Opus / PCMU / PCMA) Sending and receiving video (VP8 / H.264) Bundling audio / video / data channels RTCP reports, including NACK / PLI to recover from packet loss Installing The easiest way to install aiortc is to run: .. code:: bash pip install aiortc Building from source If there are no wheels for your system or if you wish to build aiortc from source you will need a couple of libraries installed on your system: Opus for audio encoding / decoding LibVPX for video encoding / decoding Linux ..... On Debian/Ubuntu run: .. code:: bash apt install libopus-dev libvpx-dev OS X .... On OS X run: .. code:: bash brew install opus libvpx License aiortc is released under the BSD license`_. .. _BSD license: https://aiortc.readthedocs.io/en/latest/license.html

aima-python
github
LLM Vibe Score0.575
Human Vibe Score0.33114909407186394
aimacodeMar 28, 2025

aima-python

aima-python Python code for the book Artificial Intelligence: A Modern Approach. You can use this in conjunction with a course on AI, or for study on your own. We're looking for solid contributors to help. Updates for 4th Edition The 4th edition of the book as out now in 2020, and thus we are updating the code. All code here will reflect the 4th edition. Changes include: Move from Python 3.5 to 3.7. More emphasis on Jupyter (Ipython) notebooks. More projects using external packages (tensorflow, etc.). Structure of the Project When complete, this project will have Python implementations for all the pseudocode algorithms in the book, as well as tests and examples of use. For each major topic, such as search, we provide the following files: search.ipynb and search.py: Implementations of all the pseudocode algorithms, and necessary support functions/classes/data. The .py file is generated automatically from the .ipynb file; the idea is that it is easier to read the documentation in the .ipynb file. search_XX.ipynb: Notebooks that show how to use the code, broken out into various topics (the XX). tests/test_search.py: A lightweight test suite, using assert statements, designed for use with py.test, but also usable on their own. Python 3.7 and up The code for the 3rd edition was in Python 3.5; the current 4th edition code is in Python 3.7. It should also run in later versions, but does not run in Python 2. You can install Python or use a browser-based Python interpreter such as repl.it. You can run the code in an IDE, or from the command line with python -i filename.py where the -i option puts you in an interactive loop where you can run Python functions. All notebooks are available in a binder environment. Alternatively, visit jupyter.org for instructions on setting up your own Jupyter notebook environment. Features from Python 3.6 and 3.7 that we will be using for this version of the code: f-strings: all string formatting should be done with f'var = {var}', not with 'var = {}'.format(var) nor 'var = %s' % var. typing module: declare functions with type hints: def successors(state) -> List[State]:; that is, give type declarations, but omit them when it is obvious. I don't need to say state: State, but in another context it would make sense to say s: State. Underscores in numerics: write a million as 1000000 not as 1000000. dataclasses module: replace namedtuple with dataclass. [//]: (There is a sibling [aima-docker]https://github.com/rajatjain1997/aima-docker project that shows you how to use docker containers to run more complex problems in more complex software environments.) Installation Guide To download the repository: git clone https://github.com/aimacode/aima-python.git Then you need to install the basic dependencies to run the project on your system: You also need to fetch the datasets from the aima-data repository: Wait for the datasets to download, it may take a while. Once they are downloaded, you need to install pytest, so that you can run the test suite: pip install pytest Then to run the tests: py.test And you are good to go! Index of Algorithms Here is a table of algorithms, the figure, name of the algorithm in the book and in the repository, and the file where they are implemented in the repository. This chart was made for the third edition of the book and is being updated for the upcoming fourth edition. Empty implementations are a good place for contributors to look for an issue. The aima-pseudocode project describes all the algorithms from the book. An asterisk next to the file name denotes the algorithm is not fully implemented. Another great place for contributors to start is by adding tests and writing on the notebooks. You can see which algorithms have tests and notebook sections below. If the algorithm you want to work on is covered, don't worry! You can still add more tests and provide some examples of use in the notebook! | Figure | Name (in 3rd edition) | Name (in repository) | File | Tests | Notebook |:-------|:----------------------------------|:------------------------------|:--------------------------------|:-----|:---------| | 2 | Random-Vacuum-Agent | RandomVacuumAgent | [agents.py][agents] | Done | Included | | 2 | Model-Based-Vacuum-Agent | ModelBasedVacuumAgent | [agents.py][agents] | Done | Included | | 2.1 | Environment | Environment | [agents.py][agents] | Done | Included | | 2.1 | Agent | Agent | [agents.py][agents] | Done | Included | | 2.3 | Table-Driven-Vacuum-Agent | TableDrivenVacuumAgent | [agents.py][agents] | Done | Included | | 2.7 | Table-Driven-Agent | TableDrivenAgent | [agents.py][agents] | Done | Included | | 2.8 | Reflex-Vacuum-Agent | ReflexVacuumAgent | [agents.py][agents] | Done | Included | | 2.10 | Simple-Reflex-Agent | SimpleReflexAgent | [agents.py][agents] | Done | Included | | 2.12 | Model-Based-Reflex-Agent | ReflexAgentWithState | [agents.py][agents] | Done | Included | | 3 | Problem | Problem | [search.py][search] | Done | Included | | 3 | Node | Node | [search.py][search] | Done | Included | | 3 | Queue | Queue | [utils.py][utils] | Done | No Need | | 3.1 | Simple-Problem-Solving-Agent | SimpleProblemSolvingAgent | [search.py][search] | Done | Included | | 3.2 | Romania | romania | [search.py][search] | Done | Included | | 3.7 | Tree-Search | depth/breadthfirsttree_search | [search.py][search] | Done | Included | | 3.7 | Graph-Search | depth/breadthfirstgraph_search | [search.py][search] | Done | Included | | 3.11 | Breadth-First-Search | breadthfirstgraph_search | [search.py][search] | Done | Included | | 3.14 | Uniform-Cost-Search | uniformcostsearch | [search.py][search] | Done | Included | | 3.17 | Depth-Limited-Search | depthlimitedsearch | [search.py][search] | Done | Included | | 3.18 | Iterative-Deepening-Search | iterativedeepeningsearch | [search.py][search] | Done | Included | | 3.22 | Best-First-Search | bestfirstgraph_search | [search.py][search] | Done | Included | | 3.24 | A\*-Search | astar_search | [search.py][search] | Done | Included | | 3.26 | Recursive-Best-First-Search | recursivebestfirst_search | [search.py][search] | Done | Included | | 4.2 | Hill-Climbing | hill_climbing | [search.py][search] | Done | Included | | 4.5 | Simulated-Annealing | simulated_annealing | [search.py][search] | Done | Included | | 4.8 | Genetic-Algorithm | genetic_algorithm | [search.py][search] | Done | Included | | 4.11 | And-Or-Graph-Search | andorgraph_search | [search.py][search] | Done | Included | | 4.21 | Online-DFS-Agent | onlinedfsagent | [search.py][search] | Done | Included | | 4.24 | LRTA\*-Agent | LRTAStarAgent | [search.py][search] | Done | Included | | 5.3 | Minimax-Decision | minimax_decision | [games.py][games] | Done | Included | | 5.7 | Alpha-Beta-Search | alphabeta_search | [games.py][games] | Done | Included | | 6 | CSP | CSP | [csp.py][csp] | Done | Included | | 6.3 | AC-3 | AC3 | [csp.py][csp] | Done | Included | | 6.5 | Backtracking-Search | backtracking_search | [csp.py][csp] | Done | Included | | 6.8 | Min-Conflicts | min_conflicts | [csp.py][csp] | Done | Included | | 6.11 | Tree-CSP-Solver | treecspsolver | [csp.py][csp] | Done | Included | | 7 | KB | KB | [logic.py][logic] | Done | Included | | 7.1 | KB-Agent | KB_AgentProgram | [logic.py][logic] | Done | Included | | 7.7 | Propositional Logic Sentence | Expr | [utils.py][utils] | Done | Included | | 7.10 | TT-Entails | tt_entails | [logic.py][logic] | Done | Included | | 7.12 | PL-Resolution | pl_resolution | [logic.py][logic] | Done | Included | | 7.14 | Convert to CNF | to_cnf | [logic.py][logic] | Done | Included | | 7.15 | PL-FC-Entails? | plfcentails | [logic.py][logic] | Done | Included | | 7.17 | DPLL-Satisfiable? | dpll_satisfiable | [logic.py][logic] | Done | Included | | 7.18 | WalkSAT | WalkSAT | [logic.py][logic] | Done | Included | | 7.20 | Hybrid-Wumpus-Agent | HybridWumpusAgent | | | | | 7.22 | SATPlan | SAT_plan | [logic.py][logic] | Done | Included | | 9 | Subst | subst | [logic.py][logic] | Done | Included | | 9.1 | Unify | unify | [logic.py][logic] | Done | Included | | 9.3 | FOL-FC-Ask | folfcask | [logic.py][logic] | Done | Included | | 9.6 | FOL-BC-Ask | folbcask | [logic.py][logic] | Done | Included | | 10.1 | Air-Cargo-problem | air_cargo | [planning.py][planning] | Done | Included | | 10.2 | Spare-Tire-Problem | spare_tire | [planning.py][planning] | Done | Included | | 10.3 | Three-Block-Tower | threeblocktower | [planning.py][planning] | Done | Included | | 10.7 | Cake-Problem | havecakeandeatcake_too | [planning.py][planning] | Done | Included | | 10.9 | Graphplan | GraphPlan | [planning.py][planning] | Done | Included | | 10.13 | Partial-Order-Planner | PartialOrderPlanner | [planning.py][planning] | Done | Included | | 11.1 | Job-Shop-Problem-With-Resources | jobshopproblem | [planning.py][planning] | Done | Included | | 11.5 | Hierarchical-Search | hierarchical_search | [planning.py][planning] | Done | Included | | 11.8 | Angelic-Search | angelic_search | [planning.py][planning] | Done | Included | | 11.10 | Doubles-tennis | doubletennisproblem | [planning.py][planning] | Done | Included | | 13 | Discrete Probability Distribution | ProbDist | [probability.py][probability] | Done | Included | | 13.1 | DT-Agent | DTAgent | [probability.py][probability] | Done | Included | | 14.9 | Enumeration-Ask | enumeration_ask | [probability.py][probability] | Done | Included | | 14.11 | Elimination-Ask | elimination_ask | [probability.py][probability] | Done | Included | | 14.13 | Prior-Sample | prior_sample | [probability.py][probability] | Done | Included | | 14.14 | Rejection-Sampling | rejection_sampling | [probability.py][probability] | Done | Included | | 14.15 | Likelihood-Weighting | likelihood_weighting | [probability.py][probability] | Done | Included | | 14.16 | Gibbs-Ask | gibbs_ask | [probability.py][probability] | Done | Included | | 15.4 | Forward-Backward | forward_backward | [probability.py][probability] | Done | Included | | 15.6 | Fixed-Lag-Smoothing | fixedlagsmoothing | [probability.py][probability] | Done | Included | | 15.17 | Particle-Filtering | particle_filtering | [probability.py][probability] | Done | Included | | 16.9 | Information-Gathering-Agent | InformationGatheringAgent | [probability.py][probability] | Done | Included | | 17.4 | Value-Iteration | value_iteration | [mdp.py][mdp] | Done | Included | | 17.7 | Policy-Iteration | policy_iteration | [mdp.py][mdp] | Done | Included | | 17.9 | POMDP-Value-Iteration | pomdpvalueiteration | [mdp.py][mdp] | Done | Included | | 18.5 | Decision-Tree-Learning | DecisionTreeLearner | [learning.py][learning] | Done | Included | | 18.8 | Cross-Validation | cross_validation | [learning.py][learning]\* | | | | 18.11 | Decision-List-Learning | DecisionListLearner | [learning.py][learning]\* | | | | 18.24 | Back-Prop-Learning | BackPropagationLearner | [learning.py][learning] | Done | Included | | 18.34 | AdaBoost | AdaBoost | [learning.py][learning] | Done | Included | | 19.2 | Current-Best-Learning | currentbestlearning | knowledge.py | Done | Included | | 19.3 | Version-Space-Learning | versionspacelearning | knowledge.py | Done | Included | | 19.8 | Minimal-Consistent-Det | minimalconsistentdet | knowledge.py | Done | Included | | 19.12 | FOIL | FOIL_container | knowledge.py | Done | Included | | 21.2 | Passive-ADP-Agent | PassiveADPAgent | [rl.py][rl] | Done | Included | | 21.4 | Passive-TD-Agent | PassiveTDAgent | [rl.py][rl] | Done | Included | | 21.8 | Q-Learning-Agent | QLearningAgent | [rl.py][rl] | Done | Included | | 22.1 | HITS | HITS | [nlp.py][nlp] | Done | Included | | 23 | Chart-Parse | Chart | [nlp.py][nlp] | Done | Included | | 23.5 | CYK-Parse | CYK_parse | [nlp.py][nlp] | Done | Included | | 25.9 | Monte-Carlo-Localization | montecarlolocalization | [probability.py][probability] | Done | Included | Index of data structures Here is a table of the implemented data structures, the figure, name of the implementation in the repository, and the file where they are implemented. | Figure | Name (in repository) | File | |:-------|:--------------------------------|:--------------------------| | 3.2 | romania_map | [search.py][search] | | 4.9 | vacumm_world | [search.py][search] | | 4.23 | onedimstate_space | [search.py][search] | | 6.1 | australia_map | [search.py][search] | | 7.13 | wumpusworldinference | [logic.py][logic] | | 7.16 | hornclausesKB | [logic.py][logic] | | 17.1 | sequentialdecisionenvironment | [mdp.py][mdp] | | 18.2 | waitingdecisiontree | [learning.py][learning] | Acknowledgements Many thanks for contributions over the years. I got bug reports, corrected code, and other support from Darius Bacon, Phil Ruggera, Peng Shao, Amit Patil, Ted Nienstedt, Jim Martin, Ben Catanzariti, and others. Now that the project is on GitHub, you can see the contributors who are doing a great job of actively improving the project. Many thanks to all contributors, especially @darius, @SnShine, @reachtarunhere, @antmarakis, @Chipe1, @ad71 and @MariannaSpyrakou. [agents]:../master/agents.py [csp]:../master/csp.py [games]:../master/games.py [grid]:../master/grid.py [knowledge]:../master/knowledge.py [learning]:../master/learning.py [logic]:../master/logic.py [mdp]:../master/mdp.py [nlp]:../master/nlp.py [planning]:../master/planning.py [probability]:../master/probability.py [rl]:../master/rl.py [search]:../master/search.py [utils]:../master/utils.py [text]:../master/text.py

info8006-introduction-to-ai
github
LLM Vibe Score0.532
Human Vibe Score0.28780746199907875
glouppeMar 28, 2025

info8006-introduction-to-ai

INFO8006 Introduction to Artificial Intelligence Lectures for INFO8006 Introduction to Artificial Intelligence, ULiège, Fall 2024. Instructor: Gilles Louppe Teaching assistants: Gérôme Andry, Arnaud Delaunoy When: Fall 2024, Thursday 8:30 AM to 12:30 AM Classroom: B31/Laurent (4/89) Contact: info8006@montefiore.ulg.ac.be Discord: https://discord.gg/Y8UP2SBu2h Agenda | Date | Topic | | ---- | ----- | | September 19 | [Course syllabus][syllabus] [[PDF][syllabus-pdf]] Lecture 0: [Introduction to artificial intelligence][l0] [[PDF][l0-pdf]] Lecture 1: [Intelligent agents][l1] [[PDF][l1-pdf]] | | September 26 | Lecture 2: [Solving problems by searching][l2] [[PDF][l2-pdf]] Tutorial: Project 0 | | October 3 | Lecture 3: [Games and adversarial search][l3] [[PDF][l3-pdf]] Exercises 1: Solving problems by searching [[PDF][e1]] [[Solutions][e1s]] | | October 10 | Lecture 4: [Quantifying uncertainty][l4] [[PDF][l4-pdf]] Exercises 2: Games and adversarial search [[PDF][e2]] [[Solutions][e2s]]| | October 17 | Lecture 5: [Probabilistic reasoning][l5] [[PDF][l5-pdf]] Exercises 3: Quantifying uncertainty [[PDF][e3]] [[Solutions][e3s]]| | October 24 | Lecture 6: [Reasoning over time][l6] [[PDF][l6-pdf]]No exercises| | October 31 | No class | | November 3 | Deadline for Project 1 | | November 7 | Lecture 7: [Machine learning and neural networks][l7] [[PDF][l7-pdf]] Exercises 4: Probabilistic reasoning [[PDF][e4]] [[Solutions][e4s]]| | November 14 | Lecture 7: [Machine learning and neural networks][l7] (continued) [[PDF][l7-pdf]] Exercises 5: Reasoning over time [[PDF][e5]] [[Solutions][e5s]]| | November 21 |Lecture 8: [Making decisions][l8] [[PDF][l8-pdf]] Exercises 6: Reasoning over time (continued) [notebook] | | November 28 | Lecture 9: [Reinforcement Learning][l9] [[PDF][l9-pdf]] Exercises 7: Machine learning [[PDF][e6]] [[Solutions][e6s]]| | December 5 | No lecture Exercises 8: Making decisions & RL [[PDF][e7]] [[Solutions][e7s]]| | December 8 | Deadline for Project 2 | | December 12 | No class | | December 19 | No class | [syllabus]: https://glouppe.github.io/info8006-introduction-to-ai/?p=course-syllabus.md [syllabus-pdf]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/course-syllabus.pdf [l0]: https://glouppe.github.io/info8006-introduction-to-ai/?p=lecture0.md [l0-pdf]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/lec0.pdf [l1]: https://glouppe.github.io/info8006-introduction-to-ai/?p=lecture1.md [l1-pdf]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/lec1.pdf [l2]: https://glouppe.github.io/info8006-introduction-to-ai/?p=lecture2.md [l2-pdf]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/lec2.pdf [l3]: https://glouppe.github.io/info8006-introduction-to-ai/?p=lecture3.md [l3-pdf]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/lec3.pdf [l4]: https://glouppe.github.io/info8006-introduction-to-ai/?p=lecture4.md [l4-pdf]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/lec4.pdf [l5]: https://glouppe.github.io/info8006-introduction-to-ai/?p=lecture5.md [l5-pdf]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/lec5.pdf [l6]: https://glouppe.github.io/info8006-introduction-to-ai/?p=lecture6.md [l6-pdf]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/lec6.pdf [l7]: https://glouppe.github.io/info8006-introduction-to-ai/?p=lecture7.md [l7-pdf]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/lec7.pdf [l8]: https://glouppe.github.io/info8006-introduction-to-ai/?p=lecture8.md [l8-pdf]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/lec8.pdf [l9]: https://glouppe.github.io/info8006-introduction-to-ai/?p=lecture9.md [l9-pdf]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/lec9.pdf [e1]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-1.pdf [e1s]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-1-solutions.pdf [e2]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-2.pdf [e2s]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-2-solutions.pdf [e3]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-3.pdf [e3s]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-3-solutions.pdf [e4]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-4.pdf [e4s]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-4-solutions.pdf [e5]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-5.pdf [e5s]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-5-solutions.pdf [e6]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-6.pdf [e6s]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-6-solutions.pdf [e7]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-7.pdf [e7s]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-7-solutions.pdf [e8]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-8.pdf [e8s]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-8-solutions.pdf [e9]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-9.pdf [e9s]: https://glouppe.github.io/info8006-introduction-to-ai/pdf/exercises-9-solutions.pdf Pacman programming projects General instructions Python tutorial [video (Linux), video (Windows)] Part 0: Search algorithms (tutorial session in class) Part 1: Adversarial search (due by November 3) Part 2: Bayes filter (due by December 8) Previous exams January 2019 (solutions) August 2019 January 2020 August 2020 (solutions) January 2021 (solutions) August 2021 January 2022 (solutions) August 2022 January 2023 (solutions) August 2023 January 2024 August 2024 Materials covered by the exam are listed here. Archives Previous editions 2023-2024 2022-2023 2021-2022 2020-2021 2019-2020 2018-2019 2017-2018 Archived lectures Due to progress in the field, some of the lectures have become less relevant. However, they are still available for those who are interested. | Topic | | --- | | Lecture: Constraint satisfaction problems [PDF] | | Lecture: Inference in Bayesian networks [PDF] | | Lecture: Communication [PDF] | | Lecture: Artificial general intelligence and beyond [PDF] |

awesome-quantum-machine-learning
github
LLM Vibe Score0.64
Human Vibe Score1
krishnakumarsekarMar 27, 2025

awesome-quantum-machine-learning

Awesome Quantum Machine Learning A curated list of awesome quantum machine learning algorithms,study materials,libraries and software (by language). Table of Contents INTRODUCTION Why Quantum Machine Learning? BASICS What is Quantum Mechanics? What is Quantum Computing? What is Topological Quantum Computing? Quantum Computing vs Classical Computing QUANTUM COMPUTING Atom Structure Photon wave Electron Fluctuation or spin States SuperPosition SuperPosition specific for machine learning(Quantum Walks) Classical Bit Quantum Bit or Qubit or Qbit Basic Gates in Quantum Computing Quantum Diode Quantum Transistor Quantum Processor Quantum Registery QRAM Quantum Entanglement QUANTUM COMPUTING MACHINE LEARNING BRIDGE Complex Numbers Tensors Tensors Network Oracle Hadamard transform Hilbert Space eigenvalues and eigenvectors Schr¨odinger Operators Quantum lambda calculus Quantum Amplitute Phase Qubits Encode and Decode convert classical bit to qubit Quantum Dirac and Kets Quantum Complexity Arbitrary State Generation QUANTUM ALGORITHMS Quantum Fourier Transform Variational-Quantum-Eigensolver Grovers Algorithm Shor's algorithm Hamiltonian Oracle Model Bernstein-Vazirani Algorithm Simon’s Algorithm Deutsch-Jozsa Algorithm Gradient Descent Phase Estimation Haar Tansform Quantum Ridgelet Transform Quantum NP Problem QUANTUM MACHINE LEARNING ALGORITHMS Quantum K-Nearest Neighbour Quantum K-Means Quantum Fuzzy C-Means Quantum Support Vector Machine Quantum Genetic Algorithm Quantum Hidden Morkov Models Quantum state classification with Bayesian methods Quantum Ant Colony Optimization Quantum Cellular Automata Quantum Classification using Principle Component Analysis Quantum Inspired Evolutionary Algorithm Quantum Approximate Optimization Algorithm Quantum Elephant Herding Optimization Quantum-behaved Particle Swarm Optimization Quantum Annealing Expectation-Maximization QAUNTUM NEURAL NETWORK Quantum perceptrons Qurons Quantum Auto Encoder Quantum Annealing Photonic Implementation of Quantum Neural Network Quantum Feed Forward Neural Network Quantum Boltzman Neural Network Quantum Neural Net Weight Storage Quantum Upside Down Neural Net Quantum Hamiltonian Neural Net QANN QPN SAL Quantum Hamiltonian Learning Compressed Quantum Hamiltonian Learning QAUNTUM STATISTICAL DATA ANALYSIS Quantum Probability Theory Kolmogorovian Theory Quantum Measurement Problem Intuitionistic Logic Heyting Algebra Quantum Filtering Paradoxes Quantum Stochastic Process Double Negation Quantum Stochastic Calculus Hamiltonian Calculus Quantum Ito's Formula Quantum Stochastic Differential Equations(QSDE) Quantum Stochastic Integration Itō Integral Quasiprobability Distributions Quantum Wiener Processes Quantum Statistical Ensemble Quantum Density Operator or Density Matrix Gibbs Canonical Ensemble Quantum Mean Quantum Variance Envariance Polynomial Optimization Quadratic Unconstrained Binary Optimization Quantum Gradient Descent Quantum Based Newton's Method for Constrained Optimization Quantum Based Newton's Method for UnConstrained Optimization Quantum Ensemble Quantum Topology Quantum Topological Data Analysis Quantum Bayesian Hypothesis Quantum Statistical Decision Theory Quantum Minimax Theorem Quantum Hunt-Stein Theorem Quantum Locally Asymptotic Normality Quantum Ising Model Quantum Metropolis Sampling Quantum Monte Carlo Approximation Quantum Bootstrapping Quantum Bootstrap Aggregation Quantum Decision Tree Classifier Quantum Outlier Detection Cholesky-Decomposition for Quantum Chemistry Quantum Statistical Inference Asymptotic Quantum Statistical Inference Quantum Gaussian Mixture Modal Quantum t-design Quantum Central Limit Theorem Quantum Hypothesis Testing Quantum Chi-squared and Goodness of Fit Testing Quantum Estimation Theory Quantum Way of Linear Regression Asymptotic Properties of Quantum Outlier Detection in Quantum Concepts QAUNTUM ARTIFICIAL INTELLIGENCE Heuristic Quantum Mechanics Consistent Quantum Reasoning Quantum Reinforcement Learning QAUNTUM COMPUTER VISION QUANTUM PROGRAMMING LANGUAGES , TOOLs and SOFTWARES ALL QUANTUM ALGORITHMS SOURCE CODES , GITHUBS QUANTUM HOT TOPICS Quantum Cognition Quantum Camera Quantum Mathematics Quantum Information Processing Quantum Image Processing Quantum Cryptography Quantum Elastic Search Quantum DNA Computing Adiabetic Quantum Computing Topological Big Data Anlytics using Quantum Hamiltonian Time Based Quantum Computing Deep Quantum Learning Quantum Tunneling Quantum Entanglment Quantum Eigen Spectrum Quantum Dots Quantum elctro dynamics Quantum teleportation Quantum Supremacy Quantum Zeno Effect Quantum Cohomology Quantum Chromodynamics Quantum Darwinism Quantum Coherence Quantum Decoherence Topological Quantum Computing Topological Quantum Field Theory Quantum Knots Topological Entanglment Boson Sampling Quantum Convolutional Code Stabilizer Code Quantum Chaos Quantum Game Theory Quantum Channel Tensor Space Theory Quantum Leap Quantum Mechanics for Time Travel Quantum Secured Block Chain Quantum Internet Quantum Optical Network Quantum Interference Quantum Optical Network Quantum Operating System Electron Fractionalization Flip-Flop Quantum Computer Quantum Information with Gaussian States Quantum Anomaly Detection Distributed Secure Quantum Machine Learning Decentralized Quantum Machine Learning Artificial Agents for Quantum Designs Light Based Quantum Chips for AI Training QUANTUM STATE PREPARATION ALGORITHM FOR MACHINE LEARNING Pure Quantum State Product State Matrix Product State Greenberger–Horne–Zeilinger State W state AKLT model Majumdar–Ghosh Model Multistate Landau–Zener Models Projected entangled-pair States Infinite Projected entangled-pair States Corner Transfer Matrix Method Tensor-entanglement Renormalization Tree Tensor Network for Supervised Learning QUANTUM MACHINE LEARNING VS DEEP LEARNING QUANTUM MEETUPS QUANTUM GOOGLE GROUPS QUANTUM BASED COMPANIES QUANTUM LINKEDLIN QUANTUM BASED DEGREES CONSOLIDATED QUANTUM ML BOOKS CONSOLIDATED QUANTUM ML VIDEOS CONSOLIDATED QUANTUM ML Reserach Papers CONSOLIDATED QUANTUM ML Reserach Scientist RECENT QUANTUM UPDATES FORUM ,PAGES AND NEWSLETTER INTRODUCTION Why Quantum Machine Learning? Machine Learning(ML) is just a term in recent days but the work effort start from 18th century. What is Machine Learning ? , In Simple word the answer is making the computer or application to learn themselves . So its totally related with computing fields like computer science and IT ? ,The answer is not true . ML is a common platform which is mingled in all the aspects of the life from agriculture to mechanics . Computing is a key component to use ML easily and effectively . To be more clear ,Who is the mother of ML ?, As no option Mathematics is the mother of ML . The world tremendous invention complex numbers given birth to this field . Applying mathematics to the real life problem always gives a solution . From Neural Network to the complex DNA is running under some specific mathematical formulas and theorems. As computing technology growing faster and faster mathematics entered into this field and makes the solution via computing to the real world . In the computing technology timeline once a certain achievements reached peoples interested to use advanced mathematical ideas such as complex numbers ,eigen etc and its the kick start for the ML field such as Artificial Neural Network ,DNA Computing etc. Now the main question, why this field is getting boomed now a days ? , From the business perspective , 8-10 Years before during the kick start time for ML ,the big barrier is to merge mathematics into computing field . people knows well in computing has no idea on mathematics and research mathematician has no idea on what is computing . The education as well as the Job Opportunities is like that in that time . Even if a person tried to study both then the business value for making a product be not good. Then the top product companies like Google ,IBM ,Microsoft decided to form a team with mathematician ,a physician and a computer science person to come up with various ideas in this field . Success of this team made some wonderful products and they started by providing cloud services using this product . Now we are in this stage. So what's next ? , As mathematics reached the level of time travel concepts but the computing is still running under classical mechanics . the companies understood, the computing field must have a change from classical to quantum, and they started working on the big Quantum computing field, and the market named this field as Quantum Information Science .The kick start is from Google and IBM with the Quantum Computing processor (D-Wave) for making Quantum Neural Network .The field of Quantum Computer Science and Quantum Information Science will do a big change in AI in the next 10 years. Waiting to see that........... .(google, ibm). References D-Wave - Owner of a quantum processor Google - Quantum AI Lab IBM - Quantum Computer Lab Quora - Question Regarding future of quantum AI NASA - NASA Quantum Works Youtube - Google Video of a Quantum Processor external-link - MIT Review microsoft new product - Newly Launched Microsoft Quantum Language and Development Kit microsoft - Microsoft Quantum Related Works Google2 - Google Quantum Machine Learning Blog BBC - About Google Quantum Supremacy,IBM Quantum Computer and Microsoft Q Google Quantum Supremacy - Latest 2019 Google Quantum Supremacy Achievement IBM Quantum Supremacy - IBM Talk on Quantum Supremacy as a Primer VICE on the fight - IBM Message on Google Quantum Supremacy IBM Zurich Quantum Safe Cryptography - An interesting startup to replace all our Certificate Authority Via Cloud and IBM Q BASICS What is Quantum Mechanics? In a single line study of an electron moved out of the atom then its classical mechanic ,vibrates inside the atom its quantum mechanics WIKIPEDIA - Basic History and outline LIVESCIENCE. - A survey YOUTUBE - Simple Animation Video Explanining Great. What is Quantum Computing? A way of parallel execution of multiple processess in a same time using qubit ,It reduces the computation time and size of the processor probably in neuro size WIKIPEDIA - Basic History and outline WEBOPEDIA. - A survey YOUTUBE - Simple Animation Video Explanining Great. Quantum Computing vs Classical Computing LINK - Basic outline Quantum Computing Atom Structure one line : Electron Orbiting around the nucleous in an eliptical format YOUTUBE - A nice animation video about the basic atom structure Photon Wave one line : Light nornmally called as wave transmitted as photons as similar as atoms in solid particles YOUTUBE - A nice animation video about the basic photon 1 YOUTUBE - A nice animation video about the basic photon 2 Electron Fluctuation or spin one line : When a laser light collide with solid particles the electrons of the atom will get spin between the orbitary layers of the atom ) YOUTUBE - A nice animation video about the basic Electron Spin 1 YOUTUBE - A nice animation video about the basic Electron Spin 2 YOUTUBE - A nice animation video about the basic Electron Spin 3 States one line : Put a point on the spinning electron ,if the point is in the top then state 1 and its in bottom state 0 YOUTUBE - A nice animation video about the Quantum States SuperPosition two line : During the spin of the electron the point may be in the middle of upper and lower position, So an effective decision needs to take on the point location either 0 or 1 . Better option to analyse it along with other electrons using probability and is called superposition YOUTUBE - A nice animation video about the Quantum Superposition SuperPosition specific for machine learning(Quantum Walks) one line : As due to computational complexity ,quantum computing only consider superposition between limited electrons ,In case to merge more than one set quantum walk be the idea YOUTUBE - A nice video about the Quantum Walks Classical Bits one line : If electron moved from one one atom to other ,from ground state to excited state a bit value 1 is used else bit value 0 used Qubit one line : The superposition value of states of a set of electrons is Qubit YOUTUBE - A nice video about the Quantum Bits 1 YOUTUBE - A nice video about the Bits and Qubits 2 Basic Gates in Quantum Computing one line : As like NOT, OR and AND , Basic Gates like NOT, Hadamard gate , SWAP, Phase shift etc can be made with quantum gates YOUTUBE - A nice video about the Quantum Gates Quantum Diode one line : Quantum Diodes using a different idea from normal diode, A bunch of laser photons trigger the electron to spin and the quantum magnetic flux will capture the information YOUTUBE - A nice video about the Quantum Diode Quantum Transistors one line : A transistor default have Source ,drain and gate ,Here source is photon wave ,drain is flux and gate is classical to quantum bits QUORA -Discussion about the Quantum Transistor YOUTUBE - Well Explained Quantum Processor one line : A nano integration circuit performing the quantum gates operation sorrounded by cooling units to reduce the tremendous amount of heat YOUTUBE - Well Explained Quantum Registery QRAM one line : Comapring the normal ram ,its ultrafast and very small in size ,the address location can be access using qubits superposition value ,for a very large memory set coherent superposition(address of address) be used PDF - very Well Explained QUANTUM COMPUTING MACHINE LEARNING BRIDGE Complex Numbers one line : Normally Waves Interference is in n dimensional structure , to find a polynomial equation n order curves ,better option is complex number YOUTUBE - Wonderful Series very super Explained Tensors one line : Vectors have a direction in 2D vector space ,If on a n dimensional vector space ,vectors direction can be specify with the tensor ,The best solution to find the superposition of a n vector electrons spin space is representing vectors as tensors and doing tensor calculus YOUTUBE - Wonderful super Explained tensors basics YOUTUBE - Quantum tensors basics Tensors Network one line : As like connecting multiple vectors ,multple tensors form a network ,solving such a network reduce the complexity of processing qubits YOUTUBE - Tensors Network Some ideas specifically for quantum algorithms QUANTUM MACHINE LEARNING ALGORITHMS Quantum K-Nearest Neighbour info : Here the centroid(euclidean distance) can be detected using the swap gates test between two states of the qubit , As KNN is regerssive loss can be tally using the average PDF1 from Microsoft - Theory Explanation PDF2 - A Good Material to understand the basics Matlab - Yet to come soon Python - Yet to come soon Quantum K-Means info : Two Approaches possible ,1. FFT and iFFT to make an oracle and calculate the means of superposition 2. Adiobtic Hamiltonian generation and solve the hamiltonian to determine the cluster PDF1 - Applying Quantum Kmeans on Images in a nice way PDF2 - Theory PDF3 - Explaining well the K-means clustering using hamiltonian Matlab - Yet to come soon Python - Yet to come soon Quantum Fuzzy C-Means info : As similar to kmeans fcm also using the oracle dialect ,but instead of means,here oracle optimization followed by a rotation gate is giving a good result PDF1 - Theory Matlab - Yet to come soon Python - Yet to come soon Quantum Support Vector Machine info : A little different from above as here kernel preparation is via classical and the whole training be in oracles and oracle will do the classification, As SVM is linear ,An optimal Error(Optimum of the Least Squares Dual Formulation) Based regression is needed to improve the performance PDF1 - Nice Explanation but little hard to understand :) PDF2 - Nice Application of QSVM Matlab - Yet to come soon Python - Yet to come soon Quantum Genetic Algorithm info : One of the best algorithm suited for Quantum Field ,Here the chromosomes act as qubit vectors ,the crossover part carrying by an evaluation and the mutation part carrying by the rotation of gates ![Flow Chart]() PDF1 - Very Beautiful Article , well explained and superp PDF2 - A big theory :) PDF3 - Super Comparison Matlab - Simulation Python1 - Simulation Python2 - Yet to come Quantum Hidden Morkov Models info : As HMM is already state based ,Here the quantum states acts as normal for the markov chain and the shift between states is using quantum operation based on probability distribution ![Flow Chart]() PDF1 - Nice idea and explanation PDF2 - Nice but a different concept little Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come Quantum state classification with Bayesian methods info : Quantum Bayesian Network having the same states concept using quantum states,But here the states classification to make the training data as reusable is based on the density of the states(Interference) ![Bayesian Network Sample1]() ![Bayesian Network Sample2]() ![Bayesian Network Sample3]() PDF1 - Good Theory PDF2 - Good Explanation Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come Quantum Ant Colony Optimization info : A good algorithm to process multi dimensional equations, ACO is best suited for Sales man issue , QACO is best suited for Sales man in three or more dimension, Here the quantum rotation circuit is doing the peromene update and qubits based colony communicating all around the colony in complex space ![Ant Colony Optimization 1]() PDF1 - Good Concept PDF2 - Good Application Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come Quantum Cellular Automata info : One of the very complex algorithm with various types specifically used for polynomial equations and to design the optimistic gates for a problem, Here the lattice is formed using the quatum states and time calculation is based on the change of the state between two qubits ,Best suited for nano electronics ![Quantum Cellular Automata]() Wikipedia - Basic PDF1 - Just to get the keywords PDF2 - Nice Explanation and an easily understandable application Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come QAUNTUM NEURAL NETWORK one line : Its really one of the hardest topic , To understand easily ,Normal Neural Network is doing parallel procss ,QNN is doing parallel of parallel processess ,In theory combination of various activation functions is possible in QNN ,In Normal NN more than one activation function reduce the performance and increase the complexity Quantum perceptrons info : Perceptron(layer) is the basic unit in Neural Network ,The quantum version of perceptron must satisfy both linear and non linear problems , Quantum Concepts is combination of linear(calculus of superposition) and nonlinear(State approximation using probability) ,To make a perceptron in quantum world ,Transformation(activation function) of non linearity to certain limit is needed ,which is carrying by phase estimation algorithm ![Quantum Perceptron 3]() PDF1 - Good Theory PDF2 - Good Explanation Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come QAUNTUM STATISTICAL DATA ANALYSIS one line : An under research concept ,It can be seen in multiple ways, one best way if you want to apply n derivative for a problem in current classical theory its difficult to compute as its serialization problem instead if you do parallelization of differentiation you must estimate via probability the value in all flows ,Quantum Probability Helps to achieve this ,as the loss calculation is very less . the other way comparatively booming is Quantum Bayesianism, its a solution to solve most of the uncertainity problem in statistics to combine time and space in highly advanced physical research QUANTUM PROGRAMMING LANGUAGES , TOOLs and SOFTWARES All info : All Programming languages ,softwares and tools in alphabetical order Software - Nice content of all Python library - A python library Matlab based python library - Matlab Python Library Quantum Tensor Network Github - Tensor Network Bayesforge - A Beautiful Amazon Web Service Enabled Framework for Quantum Alogorithms and Data Analytics Rigetti - A best tools repository to use quantum computer in real time Rigetti Forest - An API to connect Quantum Computer quil/pyQuil - A quantum instruction language to use forest framework Grove - Grove is a repository to showcase quantum Fourier transform, phase estimation, the quantum approximate optimization algorithm, and others developed using Forest QISKit - A IBM Kit to access quantum computer and mainly for quantum circuits IBM Bluemix Simulator - A Bluemix Simulator for Quantum Circuits Microsoft Quantum Development Kit - Microsoft Visual Studio Enbaled Kit for Quantum Circuit Creation Microsoft "Q#" - Microsoft Q Sharp a new Programming Language for Quantum Circuit Creation qiskit api python - An API to connect IBM Quantum Computer ,With the generated token its easy to connect ,but very limited utils ,Lot of new utils will come soon Cyclops Tensor Framework - A framework to do tensor network simulations Python ToolKit for chemistry and physics Quantum Algorithm simulations - A New Started Project for simulating molecule and solids Bayesian Based Quatum Projects Repository - A nice repository and the kickstarter of bayesforge Google Fermion Products - A newly launched product specifivally for chemistry simulation Tree Tensor Networks - Interesting Tensor Network in Incubator Deep Tensor Neural Network - Some useful information about Tensor Neural Network in Incubator Generative Tensorial Networks - A startup to apply machine learning via tensor network for drug discovery Google Bristlecone - A new Quantum Processor from Google , Aimed for Future Hardwares with full fledged AI support XANADU - A Light based Quantum Hardware(chips supports) and Software Company Started in Preparation Stage. Soon will be in market fathom computing - A new concept to train the ai in a processor using light and quantum based concepts. soon products will be launch Alibaba Quantum Computing Cloud Service - Cloud Service to access 11 Bit Quantum Computing Processor Atomistic Machine Learning Project - Seems something Interesting with Deep Tensor Network for Quantum Chemistry Applications circQ and Google Works - Google Top Efforts on Tools IBM Safe Cryptography on Cloud - IBM Started and Developing a Quantm Safe Cryptography to replace all our Certificate Authority via Cloud Google Tensor Network Open Source - Google Started the Most Scientist Preferred Way To Use a Quantum Computer Circuit. Tensor Flow Which Makes Easy to Design the Network and Will Leave the Work Effect Of Gates, Processor Preparation and also going to tell the beauty of Maths Google Tensor Network Github - Github Project of Google Tensor Network Quantum Tensorflow - Yet to come soon Quantum Spark - Yet to come soon Quatum Map Reduce - Yet to come soon Quantum Database - Yet to come soon Quantum Server - Yet to come soon Quantum Data Analytics - Yet to come soon QUANTUM HOT TOPICS Deep Quantum Learning why and what is deep learning? In one line , If you know deep learning you can get a good job :) ,Even a different platform undergraduated and graduated person done a master specialization in deep learning can work in this big sector :), Practically speaking machine learning (vector mathematics) , deep learning (vector space(Graphics) mathematics) and big data are the terms created by big companies to make a trend in the market ,but in science and research there is no word such that , Now a days if you ask a junior person working in this big companies ,what is deep learning ,you will get some reply as "doing linear regression with stochastic gradient for a unsupervised data using Convolutional Neural Network :)" ,They knows the words clearly and knows how to do programming using that on a bunch of "relative data" , If you ask them about the FCM , SVM and HMM etc algorithms ,they will simply say these are olden days algorithms , deep learning replaced all :), But actually they dont know from the birth to the till level and the effectiveness of algorithms and mathematics ,How many mathematical theorems in vector, spaces , tensors etc solved to find this "hiding the complexity technology", They did not played with real non relative data like medical images, astro images , geology images etc , finding a relation and features is really complex and looping over n number of images to do pattern matching is a giant work , Now a days the items mentioned as deep learning (= multiple hidden artifical neural network) is not suitable for that why quantum deep learning or deep quantum learning? In the mid of Artificial Neural Network Research people realised at the maximum extreme only certain mathematical operations possible to do with ANN and the aim of this ANN is to achieve parallel execution of many mathematical operations , In artificial Intelligence ,the world intelligence stands for mathematics ,how effective if a probem can be solvable is based on the mathematics logic applying on the problem , more the logic will give more performance(more intelligent), This goal open the gate for quantum artificial neural network, On applying the ideas behind the deep learning to quantum mechanics environment, its possible to apply complex mathematical equations to n number of non relational data to find more features and can improve the performance Quantum Machine Learning vs Deep Learning Its fun to discuss about this , In recent days most of the employees from Product Based Companies Like google,microsoft etc using the word deep learning ,What actually Deep Learning ? and is it a new inventions ? how to learn this ? Is it replacing machine learning ? these question come to the mind of junior research scholars and mid level employees The one answer to all questions is deep learning = parallel "for" loops ,No more than that ,Its an effective way of executing multiple tasks repeatly and to reduce the computation cost, But it introduce a big cap between mathematics and computerscience , How ? All classical algorithms based on serial processing ,Its depends on the feedback of the first loop ,On applying a serial classical algorithm in multiple clusters wont give a good result ,but some light weight parallel classical algorithms(Deep learning) doing the job in multiple clusters and its not suitable for complex problems, What is the solution for then? As in the title Quantum Machine Learning ,The advantage behind is deep learning is doing the batch processing simply on the data ,but quantum machine learning designed to do batch processing as per the algorithm The product companies realised this one and they started migrating to quantum machine learning and executing the classical algorithms on quantum concept gives better result than deep learning algorithms on classical computer and the target to merge both to give very wonderful result References Quora - Good Discussion Quora - The Bridge Discussion Pdf - Nice Discussion Google - Google Research Discussion Microsoft - Microsoft plan to merge both IBM - IBM plan to merge both IBM Project - IBM Project idea MIT and Google - Solutions for all questions QUANTUM MEETUPS Meetup 1 - Quantum Physics Meetup 2 - Quantum Computing London Meetup 3 - Quantum Computing New York Meetup 4 - Quantum Computing Canada Meetup 5 - Quantum Artificial Intelligence Texas Meetup 6 - Genarl Quantum Mechanics , Mathematics New York Meetup 7 - Quantum Computing Mountain View California Meetup 8 - Statistical Analysis New York Meetup 9 - Quantum Mechanics London UK Meetup 10 - Quantum Physics Sydney Australia Meetup 11 - Quantum Physics Berkeley CA Meetup 12 - Quantum Computing London UK Meetup 13 - Quantum Mechanics Carmichael CA Meetup 14 - Maths and Science Group Portland Meetup 15 - Quantum Physics Santa Monica, CA Meetup 16 - Quantum Mechanics London Meetup 17 - Quantum Computing London Meetup 18 - Quantum Meta Physics ,Kansas City , Missouri ,US Meetup 19 - Quantum Mechanics and Physics ,Boston ,Massachusetts ,US Meetup 20 - Quantum Physics and Mechanics ,San Francisco ,California Meetup 21 - Quantum Mechanics ,Langhorne, Pennsylvania Meetup 22 - Quantum Mechanics ,Portland QUANTUM BASED DEGREES Plenty of courses around the world and many Universities Launching it day by day ,Instead of covering only Quantum ML , Covering all Quantum Related topics gives more idea in the order below Available Courses Quantum Mechanics for Science and Engineers Online Standford university - Nice Preparatory Course edx - Quantum Mechanics for Everyone NPTEL 1 - Nice Series of Courses to understand basics and backbone of quantum mechanics NPTEL 2 NPTEL 3 NPTEL 4 NPTEL 5 Class Based Course UK Bristol Australia Australian National University Europe Maxs Planks University Quantum Physics Online MIT - Super Explanation and well basics NPTEL - Nice Series of Courses to understand basics and backbone of quantum Physics Class Based Course Europe University of Copenhagen Quantum Chemistry Online NPTEL 1 - Nice Series of Courses to understand basics and backbone of quantum Chemistry NPTEL 2 - Class Based Course Europe UGent Belgium Quantum Computing Online MIT - Super Explanation and well basics edx - Nice Explanation NPTEL - Nice Series of Courses to understand basics and backbone of quantum Computing Class Based Course Canada uwaterloo Singapore National University Singapore USA Berkley China Baidu Quantum Technology Class Based Course Canada uwaterloo Singapore National University Singapore Europe Munich Russia Skoltech Quantum Information Science External Links quantwiki Online MIT - Super Explanation and well basics edx - Nice Explanation NPTEL - Nice Series of Courses to understand basics and backbone of quantum information and computing Class Based Course USA MIT Standford University Joint Center for Quantum Information and Computer Science - University of Maryland Canada Perimeter Institute Singapore National University Singapore Europe ULB Belgium IQOQI Quantum Electronics Online MIT - Wonderful Course NPTEL - Nice Series of Courses to understand basics and backbone of quantum Electronics Class Based Course USA Texas Europe Zurich ICFO Asia Tata Institute Quantum Field Theory Online Standford university - Nice Preparatory Course edx - Some QFT Concepts available Class Based Course UK Imperial Europe Vrije Quantum Computer Science Class Based Course USA Oxford Joint Center for Quantum Information and Computer Science - University of Maryland Quantum Artificial Intelligence and Machine Learning External Links Quora 1 Quora 1 Artificial Agents Research for Quantum Designs Quantum Mathematics Class Based Course USA University of Notre CONSOLIDATED Quantum Research Papers scirate - Plenty of Quantum Research Papers Available Peter Wittek - Famous Researcher for the Quantum Machine Leanrning , Published a book in this topic [Murphy Yuezhen Niu] (https://scholar.google.com/citations?user=0wJPxfkAAAAJ&hl=en) - A good researcher published some nice articles Recent Quantum Updates forum ,pages and newsletter Quantum-Tech - A Beautiful Newsletter Page Publishing Amazing Links facebook Quantum Machine Learning - Running By me . Not that much good :). You can get some ideas Linkedlin Quantum Machine Learning - A nice page running by experts. Can get plenty of ideas FOSDEM 2019 Quantum Talks - A one day talk in fosdem 2019 with more than 10 research topics,tools and ideas FOSDEM 2020 Quantum Talks - Live talk in fosdem 2020 with plenty new research topics,tools and ideas License Dedicated Opensources ![Dedicated Opensources]() Source code of plenty of Algortihms in Image Processing , Data Mining ,etc in Matlab, Python ,Java and VC++ Scripts Good Explanations of Plenty of algorithms with flow chart etc Comparison Matrix of plenty of algorithms Is Quantum Machine Learning Will Reveal the Secret Maths behind Astrology? Awesome Machine Learning and Deep Learning Mathematics is online Published Basic Presentation of the series Quantum Machine Learning Contribution If you think this page might helpful. Please help for World Education Charity or kids who wants to learn

Ultimate-Data-Science-Toolkit---From-Python-Basics-to-GenerativeAI
github
LLM Vibe Score0.555
Human Vibe Score0.3470230117125603
bansalkanavMar 27, 2025

Ultimate-Data-Science-Toolkit---From-Python-Basics-to-GenerativeAI

Getting started with Machine Learning and Deep Learning Star this repo if you find it useful :star: Module 1 - Python Programming | Topic Name | What's Covered | | :---: | :---: | | Intro to Python | Applications and Features of Python, Hello World Program, Identifiers and Rules to define identifiers, Data Types (numeric, boolean, strings, list, tuple, set and dict), Comments, Input and Output, Operators - Arithmatic, Reltaional, Equality, Logical, Bitwise, Assignment, Ternary, Identity and Membership | | Data Structures in Python (Strings, List, Tuple, Set, Dictionary) | Strings - Creating a string, Indexing, Slicing, Split, Join, etc, List - Initialization, Indexing, Slicing, Sorting, Appending, etc, Tuple - Initialization, Indexing, Slicing, Count, Index, etc, Set - Initialization, Unordered Sequence, Set Opertaions, etc, Dictionary - Initialization, Updating, Keys, Values, Items, etc | | Control Statements (Conditionals and Loops) | Conditional Statements - Introducing Indentation, if statement, if...else statement, if..elif...else statement, Nested if else statement, Loops - while loops, while...else loop, Membership operator, for loop, for...else loop, Nested Loops, Break and Continue Statement, Why else? | | Functions and Modules | Functions - Introduction to Python Functions, Function Definition and Calling, Functions with Arguments/Parameters, Return Statement, Scope of a Variable, Global Variables, Modules - Introduction to Modules, Importing a Module, Aliasing, from...import statement, import everything, Some important modules - math, platform, random, webbrowser, etc | | Object Oriented Programming | Classes and Objects - Creating a class, Instantiating an Object, Constructor, Class Members - Variables and Mentods, Types of Variables - Instance, Static and Local Variables, Types of Methods - Instance, Class and Static Methods, Access Modifiers - Public, Private and Protected, Pillars of Object Oriented Programming - Inheritance, Polymorphism, Abstraction and Encapsulation, Setters and Getters, Inheritance vs Association | | Exception Handling | Errors vs Exception, Syntax and Indentation Errors, try...except block, Control Flow in try...except block, try with multiple except, finally block, try...except...else, Nested try...except...finally, User Defined Exception | | File Handling | Introduction to File Handling, Opening and Closing a File, File Object Properties, Read Data from Text Files, Write Data to Text Files, with statement, Renaming and Deleting Files | | Web API | Application Programming Interface, Indian Space Station API, API Request, Status Code, Query Parameters, Getting JSON from an API Request, Working with JSON - dump and load, Working with Twitter API | | Databases | Introduction to Databases, SQLite3 - Connecting Python with SQLite3, Performing CRUD Opertations, MySQL - Connecting Python with MySQL, Performing CRUD Opertations, MongoDB - Connecting Python with MongoDB, Performing CRUD Opertations, Object Relation Mapping - SQLAlchemy ORM, CRUD operations and Complex DB operations | | List Comprehension, Lambda, Filter, Map, Reduce) | List Comprehension, Anonymous Functions, Filter, Map, Reduce, Function Aliasing | | Problem Solving for Interviews | Swapping two numbers, Factorial of a number, Prime Number, Fibbonnacci Sequence, Armstrong Number, Palindrome Number, etc | Module 2 - Python for Data Analysis | Topic Name | What's Covered | | :---: | :---: | | Data Analytics Framework | Data Collection, Business Understanding, Exploratory Data Analysis, Data Preparation, Model Building, Model Evaluation, Deployment, Understanding Cross Industry Standard Process for Data Mining (CRISP-DM) and Microsoft's Team Data Science Process (TDSP) | | Numpy | Array Oriented Numerical Computations using Numpy, Creating a Numpy Array, Basic Operations on Numpy Array - Check Dimensions, Shape, Datatypes and ItemSize, Why Numpy, Various ways to create Numpy Array, Numpy arange() function, Numpy Random Module - rand(), randn(), randint(), uniform(), etc, Indexing and Slicing in Numpy Arrays, Applying Mathematical Operations on Numpy Array - add(), subtract(), multiply(), divide(), dot(), matmul(), sum(), log(), exp(), etc, Statistical Operations on Numpy Array - min(), max(), mean(), median(), var(), std(), corrcoef(), etc, Reshaping a Numpy Array, Miscellaneous Topics - Linspace, Sorting, Stacking, Concatenation, Append, Where and Numpy Broadcasting | | Pandas for Beginners | Pandas Data Structures - Series, Dataframe and Panel, Creating a Series, Data Access, Creating a Dataframe using Tuples and Dictionaries, DataFrame Attributes - columns, shape, dtypes, axes, values, etc, DataFrame Methods - head(), tail(), info(), describe(), Working with .csv and .xlsx - readcsv() and readexcel(), DataFrame to .csv and .xlsx - tocsv() and toexcel() | | Advance Pandas Operations | What's Covered | | Case Study - Pandas Manipulation | What's Covered | | Missing Value Treatment | What's Covered | | Visuallization Basics - Matplotlib and Seaborn | What's Covered | | Case Study - Covid19TimeSeries | What's Covered | | Plotly and Express | What's Covered | | Outliers - Coming Soon | What's Covered | Module 3 - Statistics for Data Analysis | Topic Name | What's Covered | | :---: | :---: | | Normal Distribution | What's Covered | | Central Limit Theorem | What's Covered | | Hypothesis Testing | What's Covered | | Chi Square Testing | What's Covered | | Performing Statistical Test | What's Covered | Module 4 - Machine Learning Data Preparation and Modelling with SKLearn Working with Text Data Working with Image Data Supervised ML Algorithms K - Nearest Neighbours Linear Regression Logistic Regression Gradient Descent Decision Trees Support Vector Machines Models with Feature Engineering Hyperparameter Tuning Ensembles Unsupervised ML Algorithms Clustering Principal Component Analysis Module 5 - MLOPs | Topic Name | What's Covered | | :---: | :---: | | Model Serialization and Deserialization | What's Covered | | Application Integration | What's Covered | | MLFlow - Experiment Tracking and Model Management | What's Covered | | Prefect - Orchestrate ML Pipeline | What's Covered | Module 6 - Case Studies | Topic Name | What's Covered | | :---: | :---: | | Car Price Prediction (Regression) | What's Covered | | Airline Sentiment Analysis (NLP - Classification) | What's Covered | | Adult Income Prediction (Classification) | What's Covered | | Web App Development + Serialization and Deserialization | What's Covered | | AWS Deployment | What's Covered | | Streamlit Heroku Deployment | What's Covered | | Customer Segmentation | What's Covered | | Web Scrapping | What's Covered | Module 7 - Deep Learning | Topic Name | What's Covered | | :---: | :---: | | Introduction to Deep Learning | What's Covered | | Training a Deep Neural Network + TensorFlow.Keras | What's Covered | | Convolutional Neural Network + TensorFlow.Keras | What's Covered | | Auto Encoders for Image Compression) | What's Covered | | Recurrent Neural Network (Coming Soon) | What's Covered |

eiten
github
LLM Vibe Score0.549
Human Vibe Score0.754375921646308
tradyticsMar 27, 2025

eiten

Eiten - Algorithmic Investing Strategies for Everyone Eiten is an open source toolkit by Tradytics that implements various statistical and algorithmic investing strategies such as Eigen Portfolios, Minimum Variance Portfolios, Maximum Sharpe Ratio Portfolios, and Genetic Algorithms based Portfolios. It allows you to build your own portfolios with your own set of stocks that can beat the market. The rigorous testing framework included in Eiten enables you to have confidence in your portfolios. If you are looking to discuss these tools in depth and talk about more tools that we are working on, please feel free to join our Discord channel where we have a bunch of more tools too. Files Description | Path | Description | :--- | :---------- | eiten | Main folder. | └ figures | Figures for this github repositories. | └ stocks | Folder to keep your stock lists that you want to use to create your portfolios. | └ strategies | A bunch of strategies implemented in python. | backtester.py | Backtesting module that both backtests and forward tests all portfolios. | data_loader.py | Module for loading data from yahoo finance. | portfolio_manager.py | Main file that takes in a bunch of arguments and generates several portfolios for you. | simulator.py | Simulator that uses historical returns and monte carlo to simulate future prices for the portfolios. | strategy_manager.py | Manages the strategies implemented in the 'strategies' folder. Required Packages You will need to install the following package to train and test the models. Scikit-learn Numpy Tqdm Yfinance Pandas Scipy You can install all packages using the following command. Please note that the script was written using python3. Build your portfolios Let us see how we can use all the strategies given in the toolkit to build our portfolios. The first thing you need to do is modify the stocks.txt file in the stocks folder and add the stocks of your choice. It is recommended to keep the list small i.e anywhere between 5 to 50 stocks should be fine. We have already put a small stocks list containing a bunch of tech stocks like AAPL, MSFT, TSLA etc. Let us build our portfolios now. This is the main command that you need to run. This command will use last 5 years of daily data excluding the last 90 days and build several portfolios for you. Based on those portfolios, it will then test them on the out of sample data of 90 days and show you the performance of each portfolio. Finally, it will also compare the performance with your choice of market index which is QQQ here. Let's dive into each of the parameters in detail. istest: The value determined if the program is going to keep some separate data for future testing. When this is enabled, the value of futurebars should be larger than 5. future_bars: These are the bars that the tool will exclude during portfolio building and will forward test the portfolios on the excluded set. This is also called out of sample data. datagranularityminutes: How much granular data do you want to use to build your portfolios. For long term portfolios, you should use daily data but for short term, you can use hourly or minute level data. The possible values here are 3600, 60, 30, 15, 5, 1. 3600 means daily. historytouse: Whether to use a specific number of historical bars or use everything that we receive from yahoo finance. For minute level data, we only receive up to one month of historical data. For daily, we receive 5 years worth of historical data. If you want to use all available data, the value should be all but if you want to use smaller history, you can set it to an integer value e.g 100 which will only use the last 100 bars to build the portfolios. applynoisefiltering: This uses random matrix theory to filter out the covariance matrix from randomness thus yielding better portfolios. A value of 1 will enable it and 0 will disable it. market_index: Which index do you want to use to compare your portfolios. This should mostly be SPY but since we analyzed tech stocks, we used QQQ. only_long: Whether to use long only portfolio or enable short selling as well. Long only portfolios have shown to have better performance using algorithmic techniques. eigenportfolionumber: Which eigen portfolio to use. Any value between 1-5 should work. The first eigen portfolio (1) represents the market portfolio and should act just like the underlying index such as SPY or QQQ. The second one is orthogonal and uncorrelated to the market and poses the greatest risk and reward. The following ones have reduced risk and reward. Read more on eigen-portfolios. stocksfilepath: File that contains the list of stocks that you want to use to build your portfolio. Some Portfolio Building Examples Here are a few examples for building different types of portfolios. Both long and short portfolios by analyzing last 90 days data and keeping the last 30 days as testing data. This will give us 60 days of portfolio construction data and 30 days of testing. Only long portfolio on 60 minute bars of the last 30 days. No future testing. Compare the results with SPY index instead of QQQ. Do not apply noise filtering on the covariance matrix. Use the first eigen portfolio (market portfolio) and compare with SQQQ, Portfolio Strategies Four different portfolio strategies are currently supported by the toolkit. Eigen Portfolios These portfolios are orthogonal and uncorrelated to the market in general thus yielding high reward and alpha. However, since they are uncorrelated to the market, they can also provide great risk. The first eigen portfolio is considered to be a market portfolio which is often ignored. The second one is uncorrelated to the others and provides the highest risk and reward. As we go down the numbering, the risk as well as the reward are reduced. Minimum Variance Portfolio (MVP) MVP tries to minimize the variance of the portfolio. These portfolios are lowest risk and reward. Maximum Sharpe Ratio Portfolio (MSR) MSR solves an optimization problem that tries to maximize the sharpe ratio of the portfolio. It uses past returns during the optimization process which means if past returns are not the same as future returns, the results can vary in future. Genetic Algorithm (GA) based Portfolio This is our own implementation of a GA based portfolio that again tries to maximize the sharpe ratio but in a slightly more robust way. This usually provides more robust portfolios than the others. When you run the command above, our tool will generate portfolios from all these strategies and give them to you. Let us look at some resulting portfolios. Resulting Portfolios For the purpose these results, we will use the 9 stocks in the stocks/stocks.txt file. When we run the above command, we first get the portfolio weights for all four strategies. For testing purposes, the above command used last five years of daily data up till April 29th. The remaining data for this year was used for forward testing i.e the portfolio strategies had no access to it when building the portfolios. What if my portfolio needs different stocks?: All you need to do is change the stocks in the stocks.txt file and run the tool again. Here is the final command again that we run in order to get our portfolios: Portfolio Weights We can see that the eigen portfolio is giving a large weight to TSLA while the others are dividing their weights more uniformly. An interesting phenomena happening here is the hedging with SQQQ that all the strategies have learned automatically. Every tool is assigning some positive weight to SQQQ while also assigning positive weights to other stocks which indicates that the strategies are automatically trying to hedge the portfolios from risk. Obviously this is not perfect, but just the fact that it's happening is fascinating. Let us look at the backtest results on the last five years prior to April 29, 2020. Backtest Results The backtests look pretty encouraging. The black dotted line is the market index i.e QQQ. Other lines are the strategies. Our custom genetic algorithm implementation seems to have the best backtest results because it's an advanced version of other strategies. The eigen portfolio that weighed TSLA the most have the most volatility but its profits are also very high. Finally, as expected, the MVP has the minimum variance and ultimately the least profits. However, since the variance is extremely low, it is a good portfolio for those who want to stay safe. The most interesting part comes next, let us look at the forward or future test results for these portfolios. Forward Test Results These results are from April 29th, 2020 to September 4th, 2020. The eigen portfolio performed the best but it also had a lot of volatility. Moreover, most of those returns are due to TSLA rocketing in the last few months. After that, our GA algorithm worked quite effectively as it beat the market index. Again, as expected, the MVP had the lowest risk and reward and slowly went up in 4-5 months. This shows the effectiveness and power of these algorithmic portfolio optimization strategies where we've developed different portfolios for different kinds of risk and reward profiles. Conclusion and Discussion We are happy to share this toolkit with the trading community and hope that people will like and contribute to it. As is the case with everything in trading, these strategies are not perfect but they are based on rigorous theory and some great empirical results. Please take care when trading with these strategies and always manage your risk. The above results were not cherry picked but the market has been highly bullish in the last few months which has led to the strong results shown above. We would love for the community to try out different strategies and share them with us. Special Thanks Special thanks to Scott Rome's blog. The eigen portfolios and minimum variance portfolio concepts came from his blog posts. The code for filtering eigen values of the covariance matrix was also mostly obtained from one of his posts. License A product by Tradytics Copyright (c) 2020-present, Tradytics.com

machine-learning-blackjack-solution
github
LLM Vibe Score0.42
Human Vibe Score0.022610872675250356
GregSommervilleMar 27, 2025

machine-learning-blackjack-solution

machine-learning-blackjack-solution Introduction A genetic algorithm is a type of artificial intelligence programming that uses ideas from evolution to solve complex problems. It works by creating a population of (initially random) candidate solutions, then repeatedly selecting pairs of candidates and combining their solutions using a process similar to genetic crossover. Sometimes candidate solutions even go through mutation, just to introduce new possibilities into the population. After a large number of generations, the best solution found up to that point is often the optimal, best solution possible. Genetic algorithms are particularly well-suited for combinatorial problems, where there are huge numbers of potential solutions to a problem. The evolutionary process they go through is, in essence, a search through a huge solution space. A solution space so large that you simply could never use a brute force approach. This project is a demonstration of using a genetic algorithm to find an optimal strategy for playing the casino game Blackjack. Please see this article for a story about how this program was used, and what the results were. The article describes some of the available settings, and shows how different values for those settings affect the final result. The source code is for a Windows application written in Cthat allows you to play with different settings like population size, selection style and mutation rate. Each generation's best solution is displayed, so you can watch the program literally evolve a solution. !blackjack strategy tester screenshot The property grid located at the upper left of the screen is where you adjust settings. There's an informational area below that, and the right side of the screen is the display area for the three tables that represent a strategy for playing Blackjack. The tall table on the left is for hard hands, the table in the upper right is for soft hands, and the table in the lower right is for pairs. We'll talk more about how to interpret this strategy in a bit. The columns along the tops of the three tables are for the dealer upcard. When you play Blackjack the dealer has one of his two cards initially turned face up, and the rank of that card has a big impact on recommended strategy. Notice that the upcard ranks don't include Jack, Queen or King. That's because those cards all count 10, so we group them and the Ten together and simplify the tables. To use the tables, first, determine if you have a pair, soft hand, or hard hand. Then look in the appropriate table, with the correct dealer upcard column. The cell in the table will be "H" when the correct strategy is to hit, "S" when the correct strategy is to stand, "D" for double-down, and (in the pairs table only) "P" for split. A Word About This "Optimal" Strategy Before we go any further, it needs to be stated that this problem of finding an optimal Blackjack strategy has already been solved. Back in the 1960s, a mathematician named Edward O. Thorp authored a book called Beat the Dealer, which included charts showing the optimal "Basic" strategy. That strategy looks like this: !optimal blackjack strategy So we're solving a problem that has already been solved, but that's actually good. That means we can compare our results to the known best solution. For example, if our result strategy tells us to do anything but stand when holding a pair of Tens, Jacks, Queens or Kings, we know there's a problem. There's one other thing to get out of the way before we go any further, and that's the idea of nondeterministic code. That means that if we run the same code twice in a row, we're likely to get two different results. That's something that happens with genetic algorithms due to their inherent randomness. There's no guarantee you'll find the absolute optimal solution, but it is assured that you will find an optimal or near-optimal solution. It's something that isn't typical when writing code, so it takes some adjustment for most programmers. Genetic Algorithms Now let's talk about the details of a genetic algorithm. Fitness Scores First of all, we need a way to evaluate candidates so we can compare them to each other. That means a numeric fitness score, which in this case is quite simple: you simulate playing a certain number of hands using the strategy, and then count the number of chips you have at the end. The big question is, how many hands should we test with? The challenge of trying to test a strategy is that due to the innate randomness of Blackjack, you could use the same strategy ten times and get ten completely different results. Obviously, the more hands you play, the more the randomness gets smoothed out, and the quality of the underlying strategy starts to emerge. If you doubt this, just think about flipping a coin. If you only flip it five times, there's certainly a possibility that it'll come up heads all five times (in fact, that happens just over 3% of the time). However, if you flip it 500 times, there's no way it's going to end up all heads - the odds of it happening are 0.5500, which works out to be roughly once every 3 x 10150 times you try it. After some testing and analysis, it was determined that a minimum of 100,000 hands per test is needed for a reasonable level of accuracy. There's still variance even at that number, but in order to cut the variance in half, you'd need to bump the number of hands to 500,000. One reason this accuracy is important is that in the later generations, the differences between candidates are very small. Evolution has caused the main parts of the strategy to converge on a particular approach, and towards the end all it's doing is refining the minor details. In those cases it's important to accurately determine the difference between two similar candidates. Representation Representation is simply the idea that we need to use a data structure for a candidate solution that can be combined via crossover, and possibly mutated. In this case, that's also quite simple because the way that human beings represent a Blackjack strategy is to use three tables, as we've seen. Representing those in code with three two-dimensional arrays is the obvious approach. Each cell in those three tables will have "Hit", "Stand", "Double-Down", or (only for pairs) "Split". By the way, since there are 160 cells in the hard hands table, and 80 cells in the soft hands table, and 100 cells in the pairs table, we can calculate exactly how many possible distinct strategies there are for Blackjack: 4100 x 380 x 3160 = 5 x 10174 possible Blackjack strategies That's a big number, which is obviously impossible to search using brute force. Genetic algorithms (GAs) are extremely helpful when trying to find an optimal solution from a very large set of possible solutions like this. Blackjack Rules and Strategies The rules of Blackjack are fairly simple. The dealer and the player both are dealt two cards. The player sees both of their cards (they are usually dealt face up), and one of the dealer's cards is dealt face up. Each card has a value - for cards between 2 and 10, the value is the same as the card's rank (so an Eight of Spades counts as 8, for example). All face cards count as 10, and an Ace can either be 1 or 11 (it counts as 11 only when that does not result in a hand that exceeds 21). The suit of a card does not matter. After the cards are dealt, if the player has Blackjack (a total of 21) and the dealer does not, the player is immediately paid 1.5 times their original bet, and a new hand is dealt. If the player has 21 and the dealer does also, then it's a tie and the player gets their original bet back, and a new hand is dealt. If the player wasn't dealt a Blackjack, then play continues with the player deciding whether to Stand (not get any more cards), Hit (receive an additional card), Double-down (place an additional bet, and receive one and only one more card), or, in the case of holding a pair, splitting the hand, which means placing an additional bet and receiving two new cards, so the end result is that the player is now playing two (or, in the case of multiple splits, more than two) hands simultaneously. If the player hits or double-downs and has a resulting hand that exceeds 21, then they lose and play continues with the next hand. If not, then the dealer draws until their hand totals at least 17. If the dealer exceeds 21 at this point, the player receives a payment equal to twice their original bet. If the dealer doesn't exceed 21, then the hands are compared and the player with the highest total that doesn't exceed 21 wins. Because of these rules, certain effective strategies emerge. One common strategy is that if you hold a hard hand with a value of 20, 19 or 18, you should Stand, since you avoid busting by going over 21, and you have a nice hand total that might win in a showdown with the dealer. Another common strategy is to split a pair of Aces, since Aces are so powerful (due to the fact that count as 11 or 1, you can often Hit a hand with a soft Ace with no risk of busting). Likewise, splitting a pair of 8s is a good idea because with a hard total of 16, it's likely you will bust if you take a Hit (since so many cards count as 10). As a human being, all it takes is a little knowledge about the rules in order to construct a strategy. The GA program doesn't have that advantage, and operates completely without any pre-programmed knowledge of Blackjack. It simply uses the relative fitness scores and the mechanism of evolution to find the solution. GA Settings There are many variables or settings for a GA. You can adjust population size, how parent candidates are selected, how the resulting children may be mutated, and several other items. The following sections describe some of these settings: Setting: Selection Style Once we've solved representation and have a fitness function, the next step is to select two candidates for crossover during the process of building a new generation. There are three common styles for selection, and this program supports all of them. First, you can choose Roulette Wheel selection. It's named for a Roulette wheel because you can imagine each candidate's fitness score being a wedge in a pie chart, with a size proportionate to its relative fitness compared to the other candidates. (Of course, this assumes that all fitness scores are positive, which we will talk about shortly). The main benefit of Roulette Wheel selection is that selection is fitness-proportionate. Imagine if you had only three candidates, with fitness scores of 1, 3, and 8. The relative selection probabilities for those candidates will be 1/12, 3/12, and 8/12. The downside of Roulette Wheel selection is that it tends to be somewhat slow in terms of processing. The selection process is done by iterating through the candidates until a particular condition is matched - in other words, O(N) performance. Another potential problem with Roulette Wheel selection is that there may be situations where fitness scores vary widely, to such an extent that only certain candidates have any reasonable chance of being selected. This happens frequently in early generations, since the majority of candidates are mostly random. Although this might sound like a positive (since you ultimately want to select candidates with high fitness scores), it also results in a loss of genetic diversity. In other words, even though a particular candidate may have a low fitness score in an early generation, it may contain elements that are needed to find the ultimate solution in later generations. Ranked Selection is the solution to this problem. Instead of using raw fitness scores during the selection process, the candidates are sorted by fitness, with the worst candidate receiving a score of 0, the second worse receiving 1, and so forth, all the way to the best candidate, which has a score equal to the population size - 1. Ranked Selection is quite slow, since it combines the O(N) performance of Roulette Wheel, with the additional requirement that the candidates be sorted before selection. However, there may be circumstances where it performs better than other selection approaches. Finally, the fastest selection method of all is called Tournament Selection. This method simply selects N random candidates from the current generation, and then uses the one with the best fitness score. A tournament size of 2 means two random candidates are selected, and the best of those two is used. If you have a large tournament size (like 10), then 10 different candidates will be selected, with the best of those being the ultimate selection. That obviously tilts the balance between randomness and quality. Tournament selection works well in most cases, but it does require some experimentation to find the best tourney size. Setting: Elitism Elitism is a technique that helps ensure that the best candidates are always maintained. Since all selection methods are random to some degree, it is possible to completely lose the best candidates from one generation to another. By using Elitism, we automatically advance a certain percentage of the best candidates to the next generation. Elitism does have a negative impact on performance since all of the candidates must be sorted by fitness score. Typically Elitism is done before filling the rest of a new generation with new candidates created by crossover. Crossover Details Once two candidate solutions have been selected, the next step in building a new generation is to combine those two into a single new candidate, hopefully using the best of both parent strategies. There are a number of ways to do crossover, but the method used in this program is quite straightforward - the two fitness scores are compared, and crossover happens in a relatively proportionate way. If one candidate has a fitness of 10, and the other has a fitness of 5, then the one with fitness 10 contributes twice as much to the child as the parent with a fitness of 5. Since the fitness scores in this program are based on how much the strategy would win over thousands of hands, almost all fitness scores will be negative. (This is obviously because the rules are set up so the house always wins.) This makes it difficult to calculate relative fitnesses (how do you compare a positive number with a negative, and find relative proportions?), and also causes problems with selection methods like Roulette Wheel or Ranked. To solve this, we find the lowest fitness score of the generation and add that value to each candidate. This results in an adjusted fitness score of 0 for the very worse candidate, so it never gets selected. Mutation As has been mentioned a few times, maintaining genetic diversity in our population of candidate solutions is a good thing. It helps the GA ultimately find the very best solution, by occasionally altering a candidate in a positive direction. There are two settings for mutation. MutationRate controls what percentage of new candidates have mutation done on them. MutationImpact controls what percentage of their strategy is randomized. Population Size Population size has a significant impact on performance. The smaller the population size, the faster the GA will execute. On the other hand, if the size is too low the population may not have enough genetic diversity to find the ultimate solution. During testing, it looks like 700 to 1000 is a good balance between speed and correctness. Performance Notes This program consumes a lot of processing power. Running tests of hundreds of thousands of hands of Blackjack for hundreds or thousands of candidates consumes a lot of time. It's really imperative to write the code so that it works as efficiently as possible. If your CPU isn't consistently at or above 95% usage, there's still room for improvement. Multi-threading is a natural fit for genetic algorithms because we often want to perform the same action on each candidate. The best example of this is when we calculate fitness scores. This is often an operation that takes quite a bit of time. In our case, we're dealing out 100,000 hands, and each hand has to be played until the end. If we're single-threading that code, it's going to take a long time. Multi-threading is really the way to go. Luckily, there's a ridiculously simple way to efficiently use all of your processors for an operation like this. This code loops over all of the candidates in the currentGeneration list, calls the fitness function and sets the fitness property for each: Regardless of the number of items in the list or the number of processors on your machine, the code will efficiently run the code in a multi-threaded manner, and continue only when all of the threads are complete. One of the side effects of making this code multi-threaded is that all of the code relating to evaluating a candidate must be thread-safe, including any Singleton objects. When making code thread-safe, pay attention that you don't accidentally introduce code that will slow your program down unintentionally, because sometimes it can be quite subtle. Random numbers are central to how genetic algorithms work, so it's critical that they can be used correctly from a multithreaded environment. That means that each random number generator must be separate from the others, and it also means that each must produce a distinct series of random numbers. Random number generators use seed values which are usually time-based, like the number of milliseconds the computer has been turned on. Starting with that seed, subsequent calls will return a series of numbers that look random, but really aren't. If you start with the same seed, you get the same sequence. And that's a problem because if you create multiple random number generator objects in a loop using the default time-based seed, several of them will have the same time-based initial seed value, which will result in the same sequence of "random" numbers. That's a bug, because it can reduce the true randomness of the program a great deal, and that's vital to a genetic algorithm. There are a couple of ways to solve this problem. First, you can make the random object truly a singleton, and restrict access to it by using a Clock statement. The makes all access serialized for any random number need, which reduces performance. Another approach is to make the variable static per thread. By declaring the variable as static and also marking it with the [ThreadStatic] attribute, the .NET runtime allocates one static variable per thread. That eliminates the locking/serialization, but also has performance issues. The approach used in this application is to use a non-default seed value. In this case we call Guid.NewGuid().GetHashCode(), which generates a new, unique GUID, then gets an integer hashcode value that should be unique, depending on how GetHashCode is implemented. While multithreading really helps performance, there are also other things we can do to improve performance. For example, when dealing with large populations, the hundreds or thousands of objects that will be generated each generation can quickly turn into a huge problem related to garbage collection. In the end, the easiest way to solve that is to look through the code and find objects being allocate inside a loop. It's better to declare the variable outside of the loop, and then clear it in the loop, rather than reallocate it. In a program like this one where you could be looping hundreds of thousands of times, this can result in a very significant performance boost. For example, in an early version of this code, a Deck object was created for each hand. Since there are hundreds of candidate solutions running hundreds of thousands of trial hands, this was a huge inefficiency. The code was changed to allocate one deck per test sequence. The deck was shuffled as needed, so it never needs to be reallocated. Beyond the cards in the deck, another object type that was repeatedly created and destroyed were the candidate strategies. To mitigate this problem, a StrategyPool class was created that handles allocation and deallocation. This means that strategy objects are reused, rather than dynamically created when needed. The pool class has to be thread-safe, so it does serialize access to its methods via a Clock statement, but overall using the pool approach produced a good performance increase. Finally, a subtle form of object allocation is conversion. In an early version of the code, a utility card function used Convert.ToInt32(rankEnum). Obviously, the easiest way to convert from an enum to an int is simply to cast it, like (int)rankEnum. But it's hard to know exactly what the difference is between that approach, int.Parse(), int.TryParse(), or Convert.ToInt32(), since they can all be used and are roughly equivalent. Perhaps the compiler was boxing the enum value before passing it to Convert.ToInt32(), because the profiler identified this as a function that had large amounts of thread contention waiting - and the problem got much, much worse as the generations passed. By rewriting the conversion to use a simple cast, the program performance increased threefold (3x). Contributing Please read CONTRIBUTING.md for details on our code of conduct, and the process for submitting pull requests to us. Author Greg Sommerville - Initial work* License This project is licensed under the Apache 2.0 License - see the LICENSE.md file for details

aima-java
github
LLM Vibe Score0.521
Human Vibe Score0.06620214044837505
aimacodeMar 25, 2025

aima-java

AIMA3e-Java (JDK 8+) Java implementation of algorithms from Russell and Norvig's Artificial Intelligence - A Modern Approach 3rd Edition. You can use this in conjunction with a course on AI, or for study on your own. We're looking for solid contributors to help. Getting Started Links Overview of Project Interested in Contributing Setting up your own workspace Comments on architecture and design Demo Applications that can be run from your browser (unfortunately not up to date) Javadoc for the aima-core project (outdated) Download the latest official (but outdated) version = 1.9.1 (Dec 18 2016) Latest Maven Information (for integration as a third party library) Index of Implemented Algorithms |Figure|Page|Name (in 3rd edition)|Code | -------- |:--------:| :-----| :----- | |2|34|Environment|Environment| |2.1|35|Agent|Agent| |2.3|36|Table-Driven-Vacuum-Agent|TableDrivenVacuumAgent| |2.7|47|Table-Driven-Agent|TableDrivenAgentProgram| |2.8|48|Reflex-Vacuum-Agent|ReflexVacuumAgent| |2.10|49|Simple-Reflex-Agent|SimpleReflexAgentProgram| |2.12|51|Model-Based-Reflex-Agent|ModelBasedReflexAgentProgram| |3|66|Problem|Problem| |3.1|67|Simple-Problem-Solving-Agent|SimpleProblemSolvingAgent| |3.2|68|Romania|SimplifiedRoadMapOfRomania| |3.7|77|Tree-Search|TreeSearch| |3.7|77|Graph-Search|GraphSearch| |3.10|79|Node|Node| |3.11|82|Breadth-First-Search|BreadthFirstSearch| |3.14|84|Uniform-Cost-Search|UniformCostSearch| |3|85|Depth-first Search|DepthFirstSearch| |3.17|88|Depth-Limited-Search|DepthLimitedSearch| |3.18|89|Iterative-Deepening-Search|IterativeDeepeningSearch| |3|90|Bidirectional search|BidirectionalSearch| |3|92|Best-First search|BestFirstSearch| |3|92|Greedy best-First search|GreedyBestFirstSearch| |3|93|A\* Search|AStarSearch| |3.26|99|Recursive-Best-First-Search |RecursiveBestFirstSearch| |4.2|122|Hill-Climbing|HillClimbingSearch| |4.5|126|Simulated-Annealing|SimulatedAnnealingSearch| |4.8|129|Genetic-Algorithm|GeneticAlgorithm| |4.11|136|And-Or-Graph-Search|AndOrSearch| |4|147|Online search problem|OnlineSearchProblem| |4.21|150|Online-DFS-Agent|OnlineDFSAgent| |4.24|152|LRTA\*-Agent|LRTAStarAgent| |5.3|166|Minimax-Decision|MinimaxSearch| |5.7|170|Alpha-Beta-Search|AlphaBetaSearch| |6|202|CSP|CSP| |6.1|204|Map CSP|MapCSP| |6.3|209|AC-3|AC3Strategy| |6.5|215|Backtracking-Search|AbstractBacktrackingSolver| |6.8|221|Min-Conflicts|MinConflictsSolver| |6.11|224|Tree-CSP-Solver|TreeCspSolver| |7|235|Knowledge Base|KnowledgeBase| |7.1|236|KB-Agent|KBAgent| |7.7|244|Propositional-Logic-Sentence|Sentence| |7.10|248|TT-Entails|TTEntails| |7|253|Convert-to-CNF|ConvertToCNF| |7.12|255|PL-Resolution|PLResolution| |7.15|258|PL-FC-Entails?|PLFCEntails| |7.17|261|DPLL-Satisfiable?|DPLLSatisfiable| |7.18|263|WalkSAT|WalkSAT| |7.20|270|Hybrid-Wumpus-Agent|HybridWumpusAgent| |7.22|272|SATPlan|SATPlan| |9|323|Subst|SubstVisitor| |9.1|328|Unify|Unifier| |9.3|332|FOL-FC-Ask|FOLFCAsk| |9.6|338|FOL-BC-Ask|FOLBCAsk| |9|345|CNF|CNFConverter| |9|347|Resolution|FOLTFMResolution| |9|354|Demodulation|Demodulation| |9|354|Paramodulation|Paramodulation| |9|345|Subsumption|SubsumptionElimination| |10.9|383|Graphplan|GraphPlan| |11.5|409|Hierarchical-Search|HierarchicalSearchAlgorithm| |11.8|414|Angelic-Search|---| |13.1|484|DT-Agent|DT-Agent| |13|484|Probability-Model|ProbabilityModel| |13|487|Probability-Distribution|ProbabilityDistribution| |13|490|Full-Joint-Distribution|FullJointDistributionModel| |14|510|Bayesian Network|BayesianNetwork| |14.9|525|Enumeration-Ask|EnumerationAsk| |14.11|528|Elimination-Ask|EliminationAsk| |14.13|531|Prior-Sample|PriorSample| |14.14|533|Rejection-Sampling|RejectionSampling| |14.15|534|Likelihood-Weighting|LikelihoodWeighting| |14.16|537|GIBBS-Ask|GibbsAsk| |15.4|576|Forward-Backward|ForwardBackward| |15|578|Hidden Markov Model|HiddenMarkovModel| |15.6|580|Fixed-Lag-Smoothing|FixedLagSmoothing| |15|590|Dynamic Bayesian Network|DynamicBayesianNetwork| |15.17|598|Particle-Filtering|ParticleFiltering| |16.9|632|Information-Gathering-Agent|InformationGatheringAgent| |17|647|Markov Decision Process|MarkovDecisionProcess| |17.4|653|Value-Iteration|ValueIteration| |17.7|657|Policy-Iteration|PolicyIteration| |17.9|663|POMDP-Value-Iteration|POMDPValueIteration| |18.5|702|Decision-Tree-Learning|DecisionTreeLearner| |18.8|710|Cross-Validation-Wrapper|CrossValidation| |18.11|717|Decision-List-Learning|DecisionListLearner| |18.24|734|Back-Prop-Learning|BackPropLearning| |18.34|751|AdaBoost|AdaBoostLearner| |19.2|771|Current-Best-Learning|CurrentBestLearning| |19.3|773|Version-Space-Learning|VersionSpaceLearning| |19.8|786|Minimal-Consistent-Det|MinimalConsistentDet| |19.12|793|FOIL|FOIL| |21.2|834|Passive-ADP-Agent|PassiveADPAgent| |21.4|837|Passive-TD-Agent|PassiveTDAgent| |21.8|844|Q-Learning-Agent|QLearningAgent| |22.1|871|HITS|HITS| |23.5|894|CYK-Parse|CYK| |25.9|982|Monte-Carlo-Localization|MonteCarloLocalization| Index of implemented notebooks |Chapter No|Name |Status (in 3rd edition)|Status (in 4th edition) | -------- |:--------:| :-----| :----- | |3| Solving Problems by Searching| In Progress| Not started| |6| Constraint Satisfaction Problems |In Progress|---| |12| Knowledge Representation|Done|---| |13| Quantifying Uncertainty |Done | --- | |14| Probabilistic Reasoning|In Progress| ---| Before starting to work on a new notebook: Open a new issue with the following heading: Notebook: Chapter Name - Version . Check that the issue is not assigned to anyone. Mention a topics list of what you will be implementing in the notebook for that particular chapter. You can iteratively refine the list once you start working. Start a discussion on what can go in that particular notebook. "---" indicates algorithms yet to be implemented. Index of data structures Here is a table of the data structures yet to be implemented. |Fig|Page|Name (in book)|Code| | -------- |:--------:| :-----| :----- | |9.8|341|Append|---| |10.1|369|AIR-CARGO-TRANSPORT-PROBLEM|---| |10.2|370|SPARE-TIRE-PROBLEM|---| |10.3|371|BLOCKS-WORLD |---| |10.7|380|HAVE-CAKE-AND-EAT-CAKE-TOO-PROBLEM|---| |11.1|402|JOB-SHOP-SCHEDULING-PROBLEM|---| |11.4|407|REFINEMENT-HIGH-LEVEL-ACTIONS|---| |23.6|895|SENTENCE-TREE|---| |29.1|1062|POWERS-OF-2|---|

AI-PhD-S24
github
LLM Vibe Score0.472
Human Vibe Score0.0922477795435268
rphilipzhangMar 25, 2025

AI-PhD-S24

Artificial Intelligence for Business Research (Spring 2024) Scribed Lecture Notes Class Recordings (You need to apply for access.) Teaching Team Instructor*: Renyu (Philip) Zhang, Associate Professor, Department of Decisions, Operations and Technology, CUHK Business School, philipzhang@cuhk.edu.hk, @911 Cheng Yu Tung Building. Teaching Assistant*: Leo Cao, Full-time TA, Department of Decisions, Operations and Technology, CUHK Business School, yinglyucao@cuhk.edu.hk. Please be noted that Leo will help with any issues related to the logistics, but not the content, of this course. Tutorial Instructor*: Qiansiqi Hu, MSBA Student, Department of Decisions, Operations and Technology, CUHK Business School, 1155208353@link.cuhk.edu.hk. BS in ECE, Shanghai Jiaotong University Michigan Institute. Basic Information Website: https://github.com/rphilipzhang/AI-PhD-S24 Time: Tuesday, 12:30pm-3:15pm, from Jan 9, 2024 to Apr 16, 2024, except for Feb 13 (Chinese New Year) and Mar 5 (Final Project Discussion) Location: Cheng Yu Tung Building (CYT) LT5 About Welcome to the mono-repo of the PhD course AI for Business Research (DSME 6635) at CUHK Business School in Spring 2024. You may download the Syllabus of this course first. The purpose of this course is to learn the following: Have a basic understanding of the fundamental concepts/methods in machine learning (ML) and artificial intelligence (AI) that are used (or potentially useful) in business research. Understand how business researchers have utilized ML/AI and what managerial questions have been addressed by ML/AI in the recent decade. Nurture a taste of what the state-of-the-art AI/ML technologies can do in the ML/AI community and, potentially, in your own research field. We will meet each Tuesday at 12:30pm in Cheng Yu Tung Building (CYT) LT5 (please pay attention to this room change). Please ask for my approval if you need to join us via the following Zoom links: Zoom link, Meeting ID 996 4239 3764, Passcode 386119. Most of the code in this course will be distributed through the Google CoLab cloud computing environment to avoid the incompatibility and version control issues on your local individual computer. On the other hand, you can always download the Jupyter Notebook from CoLab and run it your own computer. The CoLab files of this course can be found at this folder. The Google Sheet to sign up for groups and group tasks can be found here. The overleaf template for scribing the lecture notes of this course can be found here. If you have any feedback on this course, please directly contact Philip at philipzhang@cuhk.edu.hk and we will try our best to address it. Brief Schedule Subject to modifications. All classes start at 12:30pm and end at 3:15pm. |Session|Date |Topic|Key Words| |:-------:|:-------------:|:----:|:-:| |1|1.09|AI/ML in a Nutshell|Course Intro, ML Models, Model Evaluations| |2|1.16|Intro to DL|DL Intro, Neural Nets, Computational Issues in DL| |3|1.23|Prediction and Traditional NLP|Prediction in Biz Research, Pre-processing| |4|1.30|NLP (II): Traditional NLP|$N$-gram, NLP Performance Evaluations, Naïve Bayes| |5|2.06|NLP (III): Word2Vec|CBOW, Skip Gram| |6|2.20|NLP (IV): RNN|Glove, Language Model Evaluation, RNN| |7|2.27|NLP (V): Seq2Seq|LSTM, Seq2Seq, Attention Mechanism| |7.5|3.05|NLP (V.V): Transformer|The Bitter Lesson, Attention is All You Need| |8|3.12|NLP (VI): Pre-training|Computational Tricks in DL, BERT, GPT| |9|3.19|NLP (VII): LLM|Emergent Abilities, Chain-of-Thought, In-context Learning, GenAI in Business Research| |10|3.26|CV (I): Image Classification|CNN, AlexNet, ResNet, ViT| |11|4.02|CV (II): Image Segmentation and Video Analysis|R-CNN, YOLO, 3D-CNN| |12|4.09|Unsupervised Learning (I): Clustering & Topic Modeling|GMM, EM Algorithm, LDA| |13|4.16|Unsupervised Learning (II): Diffusion Models|VAE, DDPM, LDM, DiT| Important Dates All problem sets are due at 12:30pm right before class. |Date| Time|Event|Note| |:--:|:-:|:---:|:--:| |1.10| 11:59pm|Group Sign-Ups|Each group has at most two students.| |1.12| 7:00pm-9:00pm|Python Tutorial|Given by Qiansiqi Hu, Python Tutorial CoLab| |1.19| 7:00pm-9:00pm|PyTorch Tutorial|Given by Qiansiqi Hu, PyTorch Tutorial CoLab| |3.05|9:00am-6:00pm|Final Project Discussion|Please schedule a meeting with Philip.| |3.12| 12:30pm|Final Project Proposal|1-page maximum| |4.30| 11:59pm|Scribed Lecture Notes|Overleaf link| |5.12|11:59pm|Project Paper, Slides, and Code|Paper page limit: 10| Useful Resources Find more on the Syllabus. Books: ESL, Deep Learning, Dive into Deep Learning, ML Fairness, Applied Causal Inference Powered by ML and AI Courses: ML Intro by Andrew Ng, DL Intro by Andrew Ng, NLP (CS224N) by Chris Manning, CV (CS231N) by Fei-Fei Li, Deep Unsupervised Learning by Pieter Abbeel, DLR by Sergey Levine, DL Theory by Matus Telgarsky, LLM by Danqi Chen, Generative AI by Andrew Ng, Machine Learning and Big Data by Melissa Dell and Matthew Harding, Digital Economics and the Economics of AI by Martin Beraja, Chiara Farronato, Avi Goldfarb, and Catherine Tucker Detailed Schedule The following schedule is tentative and subject to changes. Session 1. Artificial Intelligence and Machine Learning in a Nutshell (Jan/09/2024) Keywords: Course Introduction, Machine Learning Basics, Bias-Variance Trade-off, Cross Validation, $k$-Nearest Neighbors, Decision Tree, Ensemble Methods Slides: Course Introduction, Machine Learning Basics CoLab Notebook Demos: k-Nearest Neighbors, Decision Tree Homework: Problem Set 1: Bias-Variance Trade-Off Online Python Tutorial: Python Tutorial CoLab, 7:00pm-9:00pm, Jan/12/2024 (Friday), given by Qiansiqi Hu, 1155208353@link.cuhk.edu.hk. Zoom Link, Meeting ID: 923 4642 4433, Pass code: 178146 References: The Elements of Statistical Learning (2nd Edition), 2009, by Trevor Hastie, Robert Tibshirani, Jerome Friedman, https://hastie.su.domains/ElemStatLearn/. Probabilistic Machine Learning: An Introduction, 2022, by Kevin Murphy, https://probml.github.io/pml-book/book1.html. Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106. Athey, Susan, and Guido W. Imbens. 2019. Machine learning methods that economists should know about. Annual Review of Economics 11: 685-725. Hofman, Jake M., et al. 2021. Integrating explanation and prediction in computational social science. Nature 595.7866: 181-188. Bastani, Hamsa, Dennis Zhang, and Heng Zhang. 2022. Applied machine learning in operations management. Innovative Technology at the Interface of Finance and Operations. Springer: 189-222. Kelly, Brian, and Dacheng Xiu. 2023. Financial machine learning, SSRN, https://ssrn.com/abstract=4501707. The Bitter Lesson, by Rich Sutton, which develops so far the most critical insight of AI: "The biggest lesson that can be read from 70 years of AI research is that general methods that leverage computation are ultimately the most effective, and by a large margin." Session 2. Introduction to Deep Learning (Jan/16/2024) Keywords: Random Forests, eXtreme Gradient Boosting Trees, Deep Learning Basics, Neural Nets Models, Computational Issues of Deep Learning Slides: Machine Learning Basics, Deep Learning Basics CoLab Notebook Demos: Random Forest, Extreme Gradient Boosting Tree, Gradient Descent, Chain Rule Presentation: By Xinyu Li and Qingyu Xu. Gu, Shihao, Brian Kelly, and Dacheng Xiu. 2020. Empirical asset pricing via machine learning. Review of Financial Studies 33: 2223-2273. Link to the paper. Homework: Problem Set 2: Implementing Neural Nets Online PyTorch Tutorial: PyTorch Tutorial CoLab, 7:00pm-9:00pm, Jan/19/2024 (Friday), given by Qiansiqi Hu, 1155208353@link.cuhk.edu.hk. Zoom Link, Meeting ID: 923 4642 4433, Pass code: 178146 References: Deep Learning, 2016, by Ian Goodfellow, Yoshua Bengio and Aaron Courville, https://www.deeplearningbook.org/. Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. Probabilistic Machine Learning: Advanced Topics, 2023, by Kevin Murphy, https://probml.github.io/pml-book/book2.html. Deep Learning with PyTorch, 2020, by Eli Stevens, Luca Antiga, and Thomas Viehmann. Gu, Shihao, Brian Kelly, and Dacheng Xiu. 2020. Empirical asset pricing with machine learning. Review of Financial Studies 33: 2223-2273. Session 3. DL Basics, Predictions in Business Research, and Traditonal NLP (Jan/23/2024) Keywords: Optimization and Computational Issues of Deep Learning, Prediction Problems in Business Research, Pre-processing and Word Representations in Traditional Natural Language Processing Slides: Deep Learning Basics, Prediction Problems in Business Research, NLP(I): Pre-processing and Word Representations.pdf) CoLab Notebook Demos: He Initialization, Dropout, Micrograd, NLP Pre-processing Presentation: By Letian Kong and Liheng Tan. Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106. Link to the paper. Homework: Problem Set 2: Implementing Neural Nets, due at 12:30pm, Jan/30/2024 (Tuesday). References: Kleinberg, Jon, Jens Ludwig, Sendhil Mullainathan, and Ziad Obermeyer. 2015. Prediction policy problems. American Economic Review 105(5): 491-495. Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106. Kleinberg, Jon, Himabindu Lakkaraju, Jure Leskovec, Jens Ludwig, and Sendhil Mullainathan. 2018. Human decisions and machine predictions. Quarterly Journal of Economics 133(1): 237-293. Bajari, Patrick, Denis Nekipelov, Stephen P. Ryan, and Miaoyu Yang. 2015. Machine learning methods for demand estimation. American Economic Review, 105(5): 481-485. Farias, Vivek F., and Andrew A. Li. 2019. Learning preferences with side information. Management Science 65(7): 3131-3149. Cui, Ruomeng, Santiago Gallino, Antonio Moreno, and Dennis J. Zhang. 2018. The operational value of social media information. Production and Operations Management, 27(10): 1749-1769. Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574. Chapter 2, Introduction to Information Retrieval, 2008, Cambridge University Press, by Christopher D. Manning, Prabhakar Raghavan and Hinrich Schutze, https://nlp.stanford.edu/IR-book/information-retrieval-book.html. Chapter 2, Speech and Language Processing (3rd ed. draft), 2023, by Dan Jurafsky and James H. Martin, https://web.stanford.edu/~jurafsky/slp3/. Parameter Initialization and Batch Normalization (in Chinese) GPU Comparisons-vs-NVIDIA-H100-(PCIe)-vs-NVIDIA-RTX-6000-Ada/624vs632vs640) GitHub Repo for Micrograd, by Andrej Karpathy. Hand Written Notes Session 4. Traditonal NLP (Jan/30/2024) Keywords: Pre-processing and Word Representations in NLP, N-Gram, Naïve Bayes, Language Model Evaluation, Traditional NLP Applied to Business/Econ Research Slides: NLP(I): Pre-processing and Word Representations.pdf), NLP(II): N-Gram, Naïve Bayes, and Language Model Evaluation.pdf) CoLab Notebook Demos: NLP Pre-processing, N-Gram, Naïve Bayes Presentation: By Zhi Li and Boya Peng. Hansen, Stephen, Michael McMahon, and Andrea Prat. 2018. Transparency and deliberation within the FOMC: A computational linguistics approach. Quarterly Journal of Economics, 133(2): 801-870. Link to the paper. Homework: Problem Set 3: Implementing Traditional NLP Techniques, due at 12:30pm, Feb/6/2024 (Tuesday). References: Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574. Hansen, Stephen, Michael McMahon, and Andrea Prat. 2018. Transparency and deliberation within the FOMC: A computational linguistics approach. Quarterly Journal of Economics, 133(2): 801-870. Chapters 2, 12, & 13, Introduction to Information Retrieval, 2008, Cambridge University Press, by Christopher D. Manning, Prabhakar Raghavan and Hinrich Schutze, https://nlp.stanford.edu/IR-book/information-retrieval-book.html. Chapter 2, 3 & 4, Speech and Language Processing (3rd ed. draft), 2023, by Dan Jurafsky and James H. Martin, https://web.stanford.edu/~jurafsky/slp3/. Natural Language Tool Kit (NLTK) Documentation Hand Written Notes Session 5. Deep-Learning-Based NLP: Word2Vec (Feb/06/2024) Keywords: Traditional NLP Applied to Business/Econ Research, Word2Vec: Continuous Bag of Words and Skip-Gram Slides: NLP(II): N-Gram, Naïve Bayes, and Language Model Evaluation.pdf), NLP(III): Word2Vec.pdf) CoLab Notebook Demos: Word2Vec: CBOW, Word2Vec: Skip-Gram Presentation: By Xinyu Xu and Shu Zhang. Timoshenko, Artem, and John R. Hauser. 2019. Identifying customer needs from user-generated content. Marketing Science, 38(1): 1-20. Link to the paper. Homework: No homework this week. Probably you should think about your final project when enjoying your Lunar New Year Holiday. References: Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574. Tetlock, Paul. 2007. Giving content to investor sentiment: The role of media in the stock market. Journal of Finance, 62(3): 1139-1168. Baker, Scott, Nicholas Bloom, and Steven Davis, 2016. Measuring economic policy uncertainty. Quarterly Journal of Economics, 131(4): 1593-1636. Gentzkow, Matthew, and Jesse Shapiro. 2010. What drives media slant? Evidence from US daily newspapers. Econometrica, 78(1): 35-71. Timoshenko, Artem, and John R. Hauser. 2019. Identifying customer needs from user-generated content. Marketing Science, 38(1): 1-20. Mikolov, Tomas, Kai Chen, Greg Corrado, and Jeff Dean. 2013. Efficient estimation of word representations in vector space. ArXiv Preprint, arXiv:1301.3781. Mikolov, Tomas, Ilya Sutskever, Kai Chen, Greg Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. Advances in Neural Information Processing Systems (NeurIPS) 26. Parts I - II, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Word Embeddings Trained on Google News Corpus Hand Written Notes Session 6. Deep-Learning-Based NLP: RNN and Seq2Seq (Feb/20/2024) Keywords: Word2Vec: GloVe, Word Embedding and Language Model Evaluations, Word2Vec and RNN Applied to Business/Econ Research, RNN Slides: Guest Lecture Announcement, NLP(III): Word2Vec.pdf), NLP(IV): RNN & Seq2Seq.pdf) CoLab Notebook Demos: Word2Vec: CBOW, Word2Vec: Skip-Gram Presentation: By Qiyu Dai and Yifan Ren. Huang, Allen H., Hui Wang, and Yi Yang. 2023. FinBERT: A large language model for extracting information from financial text. Contemporary Accounting Research, 40(2): 806-841. Link to the paper. Link to GitHub Repo. Homework: Problem Set 4 - Word2Vec & LSTM for Sentiment Analysis References: Ash, Elliot, and Stephen Hansen. 2023. Text algorithms in economics. Annual Review of Economics, 15: 659-688. Associated GitHub with Code Demonstrations. Li, Kai, Feng Mai, Rui Shen, and Xinyan Yan. 2021. Measuring corporate culture using machine learning. Review of Financial Studies, 34(7): 3265-3315. Chen, Fanglin, Xiao Liu, Davide Proserpio, and Isamar Troncoso. 2022. Product2Vec: Leveraging representation learning to model consumer product choice in large assortments. Available at SSRN 3519358. Pennington, Jeffrey, Richard Socher, and Christopher Manning. 2014. Glove: Global vectors for word representation. Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP) (pp. 1532-1543). Parts 2 and 5, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Chapters 9 and 10, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. RNN and LSTM Visualizations Hand Written Notes Session 7. Deep-Learning-Based NLP: Attention and Transformer (Feb/27/2024) Keywords: RNN and its Applications to Business/Econ Research, LSTM, Seq2Seq, Attention Mechanism Slides: Final Project, NLP(IV): RNN & Seq2Seq.pdf), NLP(V): Attention & Transformer.pdf) CoLab Notebook Demos: RNN & LSTM, Attention Mechanism Presentation: By Qinghe Gui and Chaoyuan Jiang. Zhang, Mengxia and Lan Luo. 2023. Can consumer-posted photos serve as a leading indicator of restaurant survival? Evidence from Yelp. Management Science 69(1): 25-50. Link to the paper. Homework: Problem Set 4 - Word2Vec & LSTM for Sentiment Analysis References: Qi, Meng, Yuanyuan Shi, Yongzhi Qi, Chenxin Ma, Rong Yuan, Di Wu, Zuo-Jun (Max) Shen. 2023. A Practical End-to-End Inventory Management Model with Deep Learning. Management Science, 69(2): 759-773. Sarzynska-Wawer, Justyna, Aleksander Wawer, Aleksandra Pawlak, Julia Szymanowska, Izabela Stefaniak, Michal Jarkiewicz, and Lukasz Okruszek. 2021. Detecting formal thought disorder by deep contextualized word representations. Psychiatry Research, 304, 114135. Hansen, Stephen, Peter J. Lambert, Nicholas Bloom, Steven J. Davis, Raffaella Sadun, and Bledi Taska. 2023. Remote work across jobs, companies, and space (No. w31007). National Bureau of Economic Research. Sutskever, Ilya, Oriol Vinyals, and Quoc V. Le. 2014. Sequence to sequence learning with neural networks. Advances in neural information processing systems, 27. Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. ICLR Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... and Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30. Parts 5, 6, and 8, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Chapters 9, 10, and 11, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. RNN and LSTM Visualizations PyTorch's Tutorial of Seq2Seq for Machine Translation Illustrated Transformer Transformer from Scratch, with the Code on GitHub Hand Written Notes Session 7.5. Deep-Learning-Based NLP: Attention is All You Need (Mar/05/2024) Keywords: Bitter Lesson: Power of Computation in AI, Attention Mechanism, Transformer Slides: The Bitter Lesson, NLP(V): Attention & Transformer.pdf) CoLab Notebook Demos: Attention Mechanism, Transformer Homework: One-page Proposal for Your Final Project References: The Bitter Lesson, by Rich Sutton Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. ICLR Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... and Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30. Part 8, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Chapter 11, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. Illustrated Transformer Transformer from Scratch, with the Code on GitHub Andrej Karpathy's Lecture to Build Transformers Hand Written Notes Session 8. Deep-Learning-Based NLP: Pretraining (Mar/12/2024) Keywords: Computations in AI, BERT (Bidirectional Encoder Representations from Transformers), GPT (Generative Pretrained Transformers) Slides: Guest Lecture by Dr. Liubo Li on Deep Learning Computation, Pretraining.pdf) CoLab Notebook Demos: Crafting Intelligence: The Art of Deep Learning Modeling, BERT API @ Hugging Face Presentation: By Zhankun Chen and Yiyi Zhao. Noy, Shakked and Whitney Zhang. 2023. Experimental evidence on the productivity effects of generative artificial intelligence. Science, 381: 187-192. Link to the Paper Homework: Problem Set 5 - Sentiment Analysis with Hugging Face, due at 12:30pm, March 26, Tuesday. References: Devlin, Jacob, Ming-Wei Chang, Kenton Lee, Kristina Toutanova. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. ArXiv preprint arXiv:1810.04805. GitHub Repo Radford, Alec, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. Improving language understanding by generative pre-training, (GPT-1) PDF link, GitHub Repo Radford, Alec, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI blog, 1(8), 9. (GPT-2) PDF Link, GitHub Repo Brown, Tom, et al. 2020. Language models are few-shot learners. Advances in neural information processing systems, 33, 1877-1901. (GPT-3) GitHub Repo Huang, Allen H., Hui Wang, and Yi Yang. 2023. FinBERT: A large language model for extracting information from financial text. Contemporary Accounting Research, 40(2): 806-841. GitHub Repo Part 9, Lecture Notes and Slides for CS 224N: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto. Link to CS 224N Part 2 & 4, Slides for COS 597G: Understanding Large Language Models, by Danqi Chen. Link to COS 597G A Visual Guide to BERT, How GPT-3 Works Andrej Karpathy's Lecture to Build GPT-2 (124M) from Scratch Hand Written Notes Session 9. Deep-Learning-Based NLP: Large Language Models (Mar/19/2024) Keywords: Large Language Models, Generative AI, Emergent Ababilities, Instruction Fine-Tuning (IFT), Reinforcement Learning with Human Feedback (RLHF), In-Context Learning, Chain-of-Thought (CoT) Slides: What's Next, Pretraining.pdf), Large Language Models.pdf) CoLab Notebook Demos: BERT API @ Hugging Face Presentation: By Jia Liu. Liu, Liu, Dzyabura, Daria, Mizik, Natalie. 2020. Visual listening in: Extracting brand image portrayed on social media. Marketing Science, 39(4): 669-686. Link to the Paper Homework: Problem Set 5 - Sentiment Analysis with Hugging Face, due at 12:30pm, March 26, Tuesday (soft-deadline). References: Wei, Jason, et al. 2021. Finetuned language models are zero-shot learners. ArXiv preprint arXiv:2109.01652, link to the paper. Wei, Jason, et al. 2022. Emergent abilities of large language models. ArXiv preprint arXiv:2206.07682, link to the paper. Ouyang, Long, et al. 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35, 27730-27744. Wei, Jason, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35, 24824-24837. Kaplan, Jared. 2020. Scaling laws for neural language models. ArXiv preprint arXiv:2001.08361, link to the paper. Hoffmann, Jordan, et al. 2022. Training compute-optimal large language models. ArXiv preprint arXiv:2203.15556, link to the paper. Shinn, Noah, et al. 2023. Reflexion: Language agents with verbal reinforcement learning. ArXiv preprint arXiv:2303.11366, link to the paper. Reisenbichler, Martin, Thomas Reutterer, David A. Schweidel, and Daniel Dan. 2022. Frontiers: Supporting content marketing with natural language generation. Marketing Science, 41(3): 441-452. Romera-Paredes, B., Barekatain, M., Novikov, A. et al. 2023. Mathematical discoveries from program search with large language models. Nature, link to the paper. Part 10, Lecture Notes and Slides for CS224N: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto. Link to CS 224N COS 597G: Understanding Large Language Models, by Danqi Chen. Link to COS 597G Andrej Karpathy's 1-hour Talk on LLM CS224n, Hugging Face Tutorial Session 10. Deep-Learning-Based CV: Image Classification (Mar/26/2024) Keywords: Large Language Models Applications, Convolution Neural Nets (CNN), LeNet, AlexNet, VGG, ResNet, ViT Slides: What's Next, Large Language Models.pdf), Image Classification.pdf) CoLab Notebook Demos: CNN, LeNet, & AlexNet, VGG, ResNet, ViT Presentation: By Yingxin Lin and Zeshen Ye. Netzer, Oded, Alain Lemaire, and Michal Herzenstein. 2019. When words sweat: Identifying signals for loan default in the text of loan applications. Journal of Marketing Research, 56(6): 960-980. Link to the Paper Homework: Problem Set 6 - AlexNet and ResNet, due at 12:30pm, April 9, Tuesday. References: Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton. 2012. Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems, 25. He, Kaiming, Xiangyu Zhang, Shaoqing Ren and Jian Sun. 2016. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition, 770-778. Dosovitskiy, Alexey, et al. 2020. An image is worth 16x16 words: Transformers for image recognition at scale. ArXiv preprint, arXiv:2010.11929, link to the paper, link to the GitHub repo. Jean, Neal, Marshall Burke, Michael Xie, Matthew W. Davis, David B. Lobell, and Stefand Ermon. 2016. Combining satellite imagery and machine learning to predict poverty. Science, 353(6301), 790-794. Zhang, Mengxia and Lan Luo. 2023. Can consumer-posted photos serve as a leading indicator of restaurant survival? Evidence from Yelp. Management Science 69(1): 25-50. Course Notes (Lectures 5 & 6) for CS231n: Deep Learning for Computer Vision, by Fei-Fei Li, Ruohan Gao, & Yunzhu Li. Link to CS231n. Chapters 7 and 8, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola. Link to the book. Fine-Tune ViT for Image Classification with Hugging Face 🤗 Transformers Hugging Face 🤗 ViT CoLab Tutorial Session 11. Deep-Learning-Based CV (II): Object Detection & Video Analysis (Apr/2/2024) Keywords: Image Processing Applications, Localization, R-CNNs, YOLOs, Semantic Segmentation, 3D CNN, Video Analysis Applications Slides: What's Next, Image Classification.pdf), Object Detection and Video Analysis.pdf) CoLab Notebook Demos: Data Augmentation, Faster R-CNN & YOLO v5 Presentation: By Qinlu Hu and Yilin Shi. Yang, Jeremy, Juanjuan Zhang, and Yuhan Zhang. 2023. Engagement that sells: Influencer video advertising on TikTok. Available at SSRN Link to the Paper Homework: Problem Set 6 - AlexNet and ResNet, due at 12:30pm, April 9, Tuesday. References: Girshick, R., Donahue, J., Darrell, T. and Malik, J., 2014. Rich feature hierarchies for accurate object detection and semantic segmentation. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 580-587). Redmon, Joseph, Santosh Divvala, Ross Girshick, and Ali Farhadi. 2016. You only look once: Unified, real-time object detection. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 779-788). Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R. and Fei-Fei, L., 2014. Large-scale video classification with convolutional neural networks. Proceedings of the IEEE conference on Computer Vision and Pattern Recognition (pp. 1725-1732). Glaeser, Edward L., Scott D. Kominers, Michael Luca, and Nikhil Naik. 2018. Big data and big cities: The promises and limitations of improved measures of urban life. Economic Inquiry, 56(1): 114-137. Zhang, S., Xu, K. and Srinivasan, K., 2023. Frontiers: Unmasking Social Compliance Behavior During the Pandemic. Marketing Science, 42(3), pp.440-450. Course Notes (Lectures 10 & 11) for CS231n: Deep Learning for Computer Vision, by Fei-Fei Li, Ruohan Gao, & Yunzhu Li. Link to CS231n. Chapter 14, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola. Link to the book. Hand Written Notes Session 12. Unsupervised Learning: Clustering, Topic Modeling & VAE (Apr/9/2024) Keywords: K-Means, Gaussian Mixture Models, EM-Algorithm, Latent Dirichlet Allocation, Variational Auto-Encoder Slides: What's Next, Clustering, Topic Modeling & VAE.pdf) CoLab Notebook Demos: K-Means, LDA, VAE Homework: Problem Set 7 - Unsupervised Learning (EM & LDA), due at 12:30pm, April 23, Tuesday. References: Blei, David M., Ng, Andrew Y., and Jordan, Michael I. 2003. Latent Dirichlet allocation. Journal of Machine Learning Research, 3(Jan): 993-1022. Kingma, D.P. and Welling, M., 2013. Auto-encoding Variational Bayes. arXiv preprint arXiv:1312.6114. Kingma, D.P. and Welling, M., 2019. An introduction to variational autoencoders. Foundations and Trends® in Machine Learning, 12(4), pp.307-392. Bandiera, O., Prat, A., Hansen, S., & Sadun, R. 2020. CEO behavior and firm performance. Journal of Political Economy, 128(4), 1325-1369. Liu, Jia and Olivier Toubia. 2018. A semantic approach for estimating consumer content preferences from online search queries. Marketing Science, 37(6): 930-952. Mueller, Hannes, and Christopher Rauh. 2018. Reading between the lines: Prediction of political violence using newspaper text. American Political Science Review, 112(2): 358-375. Tian, Z., Dew, R. and Iyengar, R., 2023. Mega or Micro? Influencer Selection Using Follower Elasticity. Journal of Marketing Research. Chapters 8.5 and 14, The Elements of Statistical Learning (2nd Edition), 2009, by Trevor Hastie, Robert Tibshirani, Jerome Friedman, Link to Book. Course Notes (Lectures 1 & 4) for CS294-158-SP24: Deep Unsupervised Learning, taught by Pieter Abbeel, Wilson Yan, Kevin Frans, Philipp Wu. Link to CS294-158-SP24. Hand Written Notes Session 13. Unsupervised Learning: Diffusion Models (Apr/16/2024) Keywords: VAE, Denoised Diffusion Probabilistic Models, Latent Diffusion Models, CLIP, Imagen, Diffusion Transformers Slides: Clustering, Topic Modeling & VAE.pdf), Diffusion Models.pdf), Course Summary CoLab Notebook Demos: VAE, DDPM, DiT Homework: Problem Set 7 - Unsupervised Learning (EM & LDA), due at 12:30pm, April 23, Tuesday. References: Kingma, D.P. and Welling, M., 2013. Auto-encoding Variational Bayes. arXiv preprint arXiv:1312.6114. Kingma, D.P. and Welling, M., 2019. An introduction to variational autoencoders. Foundations and Trends® in Machine Learning, 12(4), pp.307-392. Ho, J., Jain, A. and Abbeel, P., 2020. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33, 6840-6851. Chan, S.H., 2024. Tutorial on Diffusion Models for Imaging and Vision. arXiv preprint arXiv:2403.18103. Peebles, W. and Xie, S., 2023. Scalable diffusion models with transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 4195-4205. Link to GitHub Repo. Tian, Z., Dew, R. and Iyengar, R., 2023. Mega or Micro? Influencer Selection Using Follower Elasticity. Journal of Marketing Research. Ludwig, J. and Mullainathan, S., 2024. Machine learning as a tool for hypothesis generation. Quarterly Journal of Economics, 139(2), 751-827. Burnap, A., Hauser, J.R. and Timoshenko, A., 2023. Product aesthetic design: A machine learning augmentation. Marketing Science, 42(6), 1029-1056. Course Notes (Lecture 6) for CS294-158-SP24: Deep Unsupervised Learning, taught by Pieter Abbeel, Wilson Yan, Kevin Frans, Philipp Wu. Link to CS294-158-SP24. CVPR 2022 Tutorial: Denoising Diffusion-based Generative Modeling: Foundations and Applications, by Karsten Kreis, Ruiqi Gao, and Arash Vahdat Link to the Tutorial Lilian Weng (OpenAI)'s Blog on Diffusion Models Lilian Weng (OpenAI)'s Blog on Diffusion Models for Video Generation Hugging Face Diffusers 🤗 Library Hand Written Notes

How-to-learn-Deep-Learning
github
LLM Vibe Score0.524
Human Vibe Score0.1392403398579415
emilwallnerMar 23, 2025

How-to-learn-Deep-Learning

Approach A practical, top-down approach, starting with high-level frameworks with a focus on Deep Learning. UPDATED VERSION: 👉 Check out my 60-page guide, No ML Degree, on how to land a machine learning job without a degree. Getting started [2 months] There are three main goals to get up to speed with deep learning: 1) Get familiar to the tools you will be working with, e.g. Python, the command line and Jupyter notebooks 2) Get used to the workflow, everything from finding the data to deploying a trained model 3) Building a deep learning mindset, an intuition for how deep learning models behave and how to improve them Spend a week on codecademy.com and learn the python syntax, command line and git. If you don't have any previous programming experience, it's good to spend a few months learning how to program. Otherwise, it's easy to become overwhelmed. Spend one to two weeks using Pandas and Scikit-learn on Kaggle problems using Jupyter Notebook on Colab, e.g. Titanic, House prices, and Iris. This gives you an overview of the machine learning mindset and workflow. Spend one month implementing models on cloud GPUs. Start with FastAI and PyTorch. The FastAI community is the go-to place for people wanting to apply deep learning and share the state of the art techniques. Once you have done this, you will know how to add value with ML. Portfolio [3 - 12 months] Think of your portfolio as evidence to a potential employer that you can provide value for them. When you are looking for your first job, there are four main roles you can apply for Machine Learning Engineering, Applied Machine Learning Researcher / Residencies, Machine Learning Research Scientist, and Software Engineering. A lot of the work related to machine learning is pure software engineering roles (category 4), e.g. scaling infrastructure, but that's out of scope for this article. It's easiest to get a foot in the door if you aim for Machine Learning Engineering roles. There are a magnitude more ML engineering roles compared to category 2 & 3 roles, they require little to no theory, and they are less competitive. Most employers prefer scaling and leveraging stable implementations, often ~1 year old, instead of allocating scarce resources to implement SOTA papers, which are often time-consuming and seldom work well in practice. Once you can cover your bills and have a few years of experience, you are in a better position to learn theory and advance to category 2 & 3 roles. This is especially true if you are self-taught, you often have an edge against an average university graduate. In general, graduates have weak practical skills and strong theory skills. Context You'll have a mix of 3 - 10 technical and non-technical people looking at your portfolio, regardless of their background, you want to spark the following reactions: the applicant has experience tackling our type of problems, the applicant's work is easy to understand and well organized, and the work was without a doubt 100% made by the applicant. Most ML learners end up with the same portfolio as everyone else. Portfolio items include things as MOOC participation, dog/cat classifiers, and implementations on toy datasets such as the titanic and iris datasets. They often indicate that you actively avoid real-world problem-solving, and prefer being in your comfort zone by copy-pasting from tutorials. These portfolio items often signal negative value instead of signaling that you are a high-quality candidate. A unique portfolio item implies that you have tackled a unique problem without a solution, and thus have to engage in the type of problem-solving an employee does daily. A good starting point is to look for portfolio ideas on active Kaggle competitions, and machine learning consulting projects, and demo versions of common production pipelines. Here's a Twitter thread on how to come up with portfolio ideas. Here are rough guidelines to self-assess the strength of your portfolio: Machine learning engineering: Even though ML engineering roles are the most strategic entry point, they are still highly competitive. In general, there are ~50 software engineering roles for every ML role. From the self-learners I know, 2/3 fail to get a foot in the door and end up taking software engineering roles instead. You are ready to look for a job when you have two high-quality projects that are well-documented, have unique datasets, and are relevant to a specific industry, say banking or insurance. Project Type | Base score | -------------| -----------| Common project | -1 p || Unique project | 10 p | Multiplier Type | Factor -----------------|----------------- Strong documentation | 5x 5000-word article | 5x Kaggle Medal | 10x Employer relevancy | 20x Hireable: 5,250 p Competative: 15,000 p Applied research / research assistant/ residencies: For most companies, the risk of pursuing cutting edge research is often too high, thus only the biggest companies tend to need this skillset. There are smaller research organizations that hire for these positions, but these positions tend to be poorly advertised and have a bias for people in their existing community. Many of these roles don't require a Ph.D., which makes them available to most people with a Bachelor's or Master's degrees, or self-learners with one year of focussed study. Given the status, scarcity, and requirements for these positions, they are the most competitive ML positions. Positions at well-known companies tend to get more than a thousand applicants per position. Daily, these roles require that you understand and can implement SOTA papers, thus that's what they will be looking for in your portfolio. Projects type | Base score --------------| ----------- Common project | -10 p Unique project | 1 p SOTA paper implementation | 20 p Multiplier type | Factor ----------------| --------------- Strong documentation | 5x 5000-word article | 5x SOTA performance | 5x Employer relevancy | 20x Hireable: 52,500 p Competitive: 150,000 p Research Scientist: Research scientist roles require a Ph.D. or equivalent experience. While the former category requires the ability to implement SOTA papers, this category requires you to come up with research ideas. The mainstream research community measure the quality of research ideas by their impact, here is a list of the venues and their impact. To have a competitive portfolio, you need two published papers in the top venues in an area that's relevant to your potential employer. Project type | Base score -------------| ---------------- Common project | -100 p An unpublished paper | 5 p ICML/ICLR/NeurIPS publication | 500p All other publications | 50 p Multiplier type | Factor ------------------| ------------------ First author paper | 10x Employer relevancy | 20x Hireable: 20,000 p Competitive roles and elite PhD positions: 200,000 p Examples: My first portfolio item (after 2 months of learning): Code | Write-up My second portfolio item (after 4 months of learning): Code | Write-up Dylan Djian's first portfolio item: Code | Write-up Dylan Djian's second portfolio item: Code | Write-up Reiichiro Nakano's first portfolio item: Code | Write-up Reiichiro Nakano's second portfolio item: Write-up Most recruiters will spend 10-20 seconds on each of your portfolio items. Unless they can understand the value in that time frame, the value of the project is close to zero. Thus, writing and documentation are key. Here's another thread on how to write about portfolio items. The last key point is relevancy. It's more fun to make a wide range of projects, but if you want to optimize for breaking into the industry, you want to do all projects in one niche, thus making your skillset super relevant for a specific pool of employers. Further Inspiration: FastAI student projects Stanford NLP student projects Stanford CNN student projects Theory 101 [4 months] Learning how to read papers is critical if you want to get into research, and a brilliant asset as an ML engineer. There are three key areas to feel comfortable reading papers: 1) Understanding the details of the most frequent algorithms, gradient descent, linear regression, and MLPs, etc 2) Learning how to translate the most frequent math notations into code 3) Learn the basics of algebra, calculus, statistics, and machine learning For the first week, spend it on 3Blue1Brown's Essence of linear algebra, the Essence of Calculus, and StatQuests' the Basics (of statistics) and Machine Learning. Use a spaced repetition app like Anki and memorize all the key concepts. Use images as much as possible, they are easier to memorize. Spend one month recoding the core concepts in python numpy, including least squares, gradient descent, linear regression, and a vanilla neural network. This will help you reduce a lot of cognitive load down the line. Learning that notations are compact logic and how to translate it into code will make you feel less anxious about the theory. I believe the best deep learning theory curriculum is the Deep Learning Book by Ian Goodfellow and Yoshua Bengio and Aaron Courville. I use it as a curriculum, and the use online courses and internet resources to learn the details about each concept. Spend three months on part 1 of the Deep learning book. Use lectures and videos to understand the concepts, Khan academy type exercises to master each concept, and Anki flashcards to remember them long-term. Key Books: Deep Learning Book by Ian Goodfellow and Yoshua Bengio and Aaron Courville. Deep Learning for Coders with fastai and PyTorch: AI Applications Without a PhD by Jeremy Howard and Sylvain. Gugger. Deep Learning with Python by François Chollet. Neural Networks and Deep Learning by Michael Nielsen. Grokking Deep Learning by Andrew W. Trask. Forums FastAI Keras Slack Distill Slack Pytorch Twitter Other good learning strategies: Emil Wallner S. Zayd Enam Catherine Olsson Greg Brockman V2 Greg Brockman V1 Andrew Ng Amid Fish Spinning Up by OpenAI Confession as an AI researcher YC Threads: One and Two If you have suggestions/questions create an issue or ping me on Twitter. UPDATED VERSION: 👉 Check out my 60-page guide, No ML Degree, on how to land a machine learning job without a degree. Language versions: Korean | English

deep-rts
github
LLM Vibe Score0.447
Human Vibe Score0.06348640915593705
cairMar 20, 2025

deep-rts

Description DeepRTS is a high-performance Real-TIme strategy game for Reinforcement Learning research. It is written in C++ for performance, but provides an python interface to better interface with machine-learning toolkits. Deep RTS can process the game with over 6 000 000 steps per second and 2 000 000 steps when rendering graphics. In comparison to other solutions, such as StarCraft, this is over 15 000% faster simulation time running on Intel i7-8700k with Nvidia RTX 2080 TI. The aim of Deep RTS is to bring a more affordable and sustainable solution to RTS AI research by reducing computation time. It is recommended to use the master-branch for the newest (and usually best) version of the environment. I am greatful for any input in regards to improving the environment. Please use the following citation when using this in your work! Dependencies Python >= 3.9.1 Installation Method 1 (From Git Repo) Method 2 (Clone & Build) Available maps Scenarios Deep RTS features scenarios which is pre-built mini-games. These mini-games is well suited to train agents on specific tasks, or to test algorithms in different problem setups. The benefits of using scenarios is that you can trivially design reward functions using criterias that each outputs a reward/punishment signal depending on completion of the task. Examples of tasks are to: collect 1000 gold do 100 damage take 1000 damage defeat 5 enemies Deep RTS currently implements the following scenarios Minimal Example In-Game Footage 10x10 - 2 Player - free-for-all 15x15 - 2 Player - free-for-all 21x21 - 2 Player - free-for-all 31x31 - 2 Player - free-for-all 31x31 - 4 Player - free-for-all 31x3 - 6 Player - free-for-all

dcai-lab
github
LLM Vibe Score0.541
Human Vibe Score0.3372420543528328
dcai-courseMar 8, 2025

dcai-lab

Lab assignments for Introduction to Data-Centric AI This repository contains the lab assignments for the Introduction to Data-Centric AI class. Contributions are most welcome! If you have ideas for improving the labs, please open an issue or submit a pull request. If you're looking for the 2023 version of the labs, check out the 2023 branch. [Lab 1: Data-Centric AI vs. Model-Centric AI][lab-1] The [first lab assignment][lab-1] walks you through an ML task of building a text classifier, and illustrates the power (and often simplicity) of data-centric approaches. [lab-1]: datacentricmodel_centric/Lab%20-%20Data-Centric%20AI%20vs%20Model-Centric%20AI.ipynb [Lab 2: Label Errors][lab-2] [This lab][lab-2] guides you through writing your own implementation of automatic label error identification using Confident Learning, the technique taught in [today’s lecture][lec-2]. [lab-2]: label_errors/Lab%20-%20Label%20Errors.ipynb [lec-2]: https://dcai.csail.mit.edu/lectures/label-errors/ [Lab 3: Dataset Creation and Curation][lab-3] [This lab assignment][lab-3] is to analyze an already collected dataset labeled by multiple annotators. [lab-3]: dataset_curation/Lab%20-%20Dataset%20Curation.ipynb [Lab 4: Data-centric Evaluation of ML Models][lab-4] [This lab assignment][lab-4] is to try improving the performance of a given model solely by improving its training data via some of the various strategies covered here. [lab-4]: datacentricevaluation/Lab%20-%20Data-Centric%20Evaluation.ipynb [Lab 5: Class Imbalance, Outliers, and Distribution Shift][lab-5] [The lab assignment][lab-5] for this lecture is to implement and compare different methods for identifying outliers. For this lab, we've focused on anomaly detection. You are given a clean training dataset consisting of many pictures of dogs, and an evaluation dataset that contains outliers (non-dogs). Your task is to implement and compare various methods for detecting these outliers. You may implement some of the ideas presented in [today's lecture][lec-5], or you can look up other outlier detection algorithms in the linked references or online. [lab-5]: outliers/Lab%20-%20Outliers.ipynb [lec-5]: https://dcai.csail.mit.edu/lectures/imbalance-outliers-shift/ [Lab 6: Growing or Compressing Datasets][lab-6] [This lab][lab-6] guides you through an implementation of active learning. [lab-6]: growing_datasets/Lab%20-%20Growing%20Datasets.ipynb [Lab 7: Interpretability in Data-Centric ML][lab-7] [This lab][lab-7] guides you through finding issues in a dataset’s features by applying interpretability techniques. [lab-7]: interpretable_features/Lab%20-%20Interpretable%20Features.ipynb [Lab 8: Encoding Human Priors: Data Augmentation and Prompt Engineering][lab-8] [This lab] guides you through prompt engineering, crafting inputs for large language models (LLMs). With these large pre-trained models, even small amounts of data can make them very useful. This lab is also [available on Colab][lab-8-colab]. [lab-8]: promptengineering/LabPrompt_Engineering.ipynb [lab-8-colab]: https://colab.research.google.com/drive/1cipH-u6Jz0EH-6Cd9MPYgY4K0sJZwRJq [Lab 9: Data Privacy and Security][lab-9] The [lab assignment][lab-9] for this lecture is to implement a membership inference attack. You are given a trained machine learning model, available as a black-box prediction function. Your task is to devise a method to determine whether or not a given data point was in the training set of this model. You may implement some of the ideas presented in [today’s lecture][lec-9], or you can look up other membership inference attack algorithms. [lab-9]: membership_inference/Lab%20-%20Membership%20Inference.ipynb [lec-9]: https://dcai.csail.mit.edu/lectures/data-privacy-security/ License Copyright (c) by the instructors of Introduction to Data-Centric AI (dcai.csail.mit.edu). dcai-lab is free software: you can redistribute it and/or modify it under the terms of the GNU Affero General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. dcai-lab is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See GNU Affero General Public LICENSE for details.

Awesome-Ai-Tools
github
LLM Vibe Score0.385
Human Vibe Score0.0020930582944730723
aliammari1Feb 21, 2025

Awesome-Ai-Tools

Awesome-Ai-Tools This repo contains AI tools that will help you achieve your goals. The tools are categorized into different sections based on their functionality. Contents Awesome-Ai-Tools Contents Productivity Time Management Task Management Email Management Creativity Art Music Writing Communication Writing Personality Analysis Translation Data Science Machine Learning Data Analysis Data Visualization Natural Language Processing Text Classification Named Entity Recognition Computer Vision Image Classification Object Detection Robotics Robot Simulation Robot Control Miscellaneous Language Models Generative Models Productivity If you're looking to boost your productivity, there are a number of AI tools that can help. Time Management RescueTime - RescueTime is an AI-powered time tracking tool that helps you understand how you're spending your time on your computer. It can help you identify areas where you're wasting time and make adjustments to your workflow to be more productive. Focus@Will - Focus@Will is an AI-powered music service that helps you stay focused and productive while you work. It uses neuroscience to create music that is scientifically optimized to help you concentrate. Clockify - Clockify is an AI-powered time tracking tool that helps you track your time across different projects and tasks. It can help you identify areas where you're spending too much time and make adjustments to your workflow to be more productive. Trello - Trello is an AI-powered task management tool that helps you stay organized and on top of your to-do list. It can help you prioritize tasks, set deadlines, and even collaborate with others on projects. Motion - Motion is an AI-powered calendar and task management tool that automatically schedules your tasks and meetings for optimal productivity. Reclaim.ai - Reclaim is an intelligent calendar assistant that helps you protect your time by automatically scheduling meetings and tasks. Task Management Todoist - Todoist is an AI-powered task management tool that helps you stay organized and on top of your to-do list. It can help you prioritize tasks, set deadlines, and even suggest tasks based on your previous activity. Asana - Asana is an AI-powered task management tool that helps you stay organized and on top of your to-do list. It can help you prioritize tasks, set deadlines, and even collaborate with others on projects. Notion - Notion is an AI-powered productivity tool that can help you manage tasks, take notes, and collaborate with others on projects. It can also be used to create wikis, databases, and other types of content. Taskade - Taskade is an AI-powered productivity tool that can manage tasks and notes for individuals and teams. ClickUp - ClickUp is an AI-enhanced project management tool that helps teams organize work with automated task distributions and smart notifications. Monday.com - Monday.com uses AI to streamline workflow management and automate routine tasks. Email Management Boomerang - Boomerang is an AI-powered email management tool that helps you manage your inbox more efficiently. It can help you schedule emails to be sent later, remind you to follow up on emails, and even suggest responses to emails. SaneBox - SaneBox is an AI-powered email management tool that helps you manage your inbox more efficiently. It can help you prioritize emails, unsubscribe from unwanted emails, and even snooze emails to be dealt with later. Mailstrom - Mailstrom is an AI-powered email management tool that helps you clean up your inbox. It can help you quickly identify and delete unwanted emails, and even unsubscribe from newsletters and other types of email subscriptions. Creativity If you're looking to get more creative, there are a number of AI tools that can help. Art Artbreeder - Artbreeder is an AI-powered tool that allows you to create unique digital art by combining different images and styles. Runway ML - Runway is an AI-powered tool that allows users to edit and generate videos using natural language descriptions. Prisma - Prisma is an AI-powered tool that allows you to transform your photos into works of art using neural networks. Music AIVA - AIVA is an AI-powered music composition tool that can help you create original music for your projects. Writing monica - Monica is a chrome extension powered by ChatGPT API. It is designed to be your personal AI assistant for effortless chatting and copywriting. CopyAI - CopyAI is an AI-powered writing assistant that can help you generate high-quality marketing copy, product descriptions, and more. Grammarly - Grammarly is an AI-powered writing assistant that helps you catch grammar and spelling errors in your writing. It can also suggest improvements to your writing style to help you communicate more effectively. Jasper - Jasper is an AI writing assistant that helps create marketing copy, blog posts, and social media content. Rytr - Rytr is an AI writing tool that helps generate content in different tones and styles. Communication If you're looking to improve your communication skills, there are a number of AI tools that can help. Writing Linguix - Linguix is an AI-powered writing assistant that can help you improve your writing skills. It can catch grammar and spelling errors, suggest improvements to your writing style, and even help you avoid plagiarism. Hemingway Editor - Hemingway Editor is an AI-powered writing tool that helps you simplify your writing and make it more readable. It can help you identify complex sentences, passive voice, and other issues that can make your writing difficult to understand. Personality Analysis Crystal - Crystal is an AI-powered tool that helps you understand the personality of the people you're communicating with. It can provide insights into their communication style and suggest ways to communicate more effectively with them. IBM Watson Personality Insights - IBM Watson Personality Insights is a tool that uses natural language processing and machine learning algorithms to analyze text and provide insights into the personality traits of the author. Translation DeepL - DeepL is an AI-powered translation tool that provides high-quality translations in multiple languages. It uses neural network algorithms to provide more accurate translations than traditional translation tools. Google Translate - Google Translate is a free online translation tool that uses machine learning algorithms to provide translations in over 100 languages. Data Science If you're working with data, there are a number of AI tools that can help you analyze and make sense of it. Machine Learning DataRobot - DataRobot is an AI-powered platform that helps you build and deploy machine learning models. It can help you automate the process of building models and make predictions based on your data. TensorFlow - TensorFlow is an open-source machine learning framework developed by Google. It can help you build and train machine learning models for a variety of applications. PyTorch - PyTorch is another open-source machine learning framework that is popular among researchers and developers. It is known for its ease of use and flexibility. H2O.ai - H2O.ai is an open-source machine learning platform that allows you to build and deploy machine learning models at scale. PyTorch3d - Pytorch 3d is an open-source library for deep learning with 3d data. Auto-sklearn - Auto-sklearn is an automated machine learning toolkit that helps find the best machine learning pipeline for your dataset. Ludwig - Ludwig is a declarative machine learning framework that makes it easy to build and train models without writing code. Data Analysis Pandas - Pandas is an open-source data analysis library for Python. It can help you manipulate and analyze data in a variety of formats, including CSV, Excel, and SQL databases. RapidMiner - RapidMiner is an AI-powered data science platform that allows you to build and deploy predictive models without writing any code. Apache Spark - Apache Spark is an open-source big data processing framework that can help you analyze large datasets in a distributed computing environment. Data Visualization Tableau - Tableau is a data visualization tool that uses AI to help you explore and understand your data. It can help you identify patterns and trends in your data that might not be immediately obvious. Plotly - Plotly is an open-source data visualization library for Python. It can help you create interactive charts and graphs that can be embedded in web pages and other applications. D3.js - D3.js is a JavaScript library for data visualization that allows you to create dynamic and interactive visualizations using web standards like HTML, CSS, and SVG. Natural Language Processing If you're interested in natural language processing, there are a number of AI tools that can help you get started. Text Classification TextBlob - TextBlob is an open-source library for processing textual data in Python. It can help you perform tasks like sentiment analysis, part-of-speech tagging, and text classification. NLTK - NLTK (Natural Language Toolkit) is another open-source library for natural language processing in Python. It can help you perform tasks like tokenization, stemming, and named entity recognition. Amazon Comprehend - Amazon Comprehend is a natural language processing service that uses machine learning to analyze text and provide insights into the content and sentiment of the text. Named Entity Recognition spaCy - spaCy is an open-source library for advanced natural language processing in Python. It can help you build applications that can understand and analyze human language. One of its key features is named entity recognition, which can identify and classify entities like people, organizations, and locations. Google Cloud Natural Language API - Google Cloud Natural Language API is a natural language processing service that can analyze text and provide insights into the sentiment, entities, and syntax of the text. Computer Vision If you're interested in computer vision, there are a number of AI tools that can help you get started. Image Classification Clarifai - Clarifai is an AI-powered image recognition tool that can help you classify images based on their content. It can recognize objects, scenes, and even specific concepts like emotions and colors. Google Cloud Vision API - Google Cloud Vision API is a computer vision service that can analyze images and provide insights into the content of the images, including objects, faces, and text. Object Detection YOLO - YOLO (You Only Look Once) is an open-source object detection system that can detect objects in real-time video streams. It is known for its speed and accuracy. Amazon Rekognition - Amazon Rekognition is a computer vision service that can analyze images and videos and provide insights into the content of the media, including objects, faces, and text. Robotics If you're interested in robotics, there are a number of AI tools that can help you get started. Robot Simulation Gazebo - Gazebo is an open-source robot simulation tool that allows you to simulate robots in a virtual environment. It can help you test and debug your robot control algorithms before deploying them on a physical robot. Webots - Webots is another open-source robot simulation tool that allows you to simulate robots in a virtual environment. It supports a wide range of robots and sensors, and can be used for both research and education. Robot Control ROS - ROS (Robot Operating System) is an open-source framework for building robotics software. It can help you build and control robots using a variety of programming languages. Miscellaneous If you're looking for AI tools that don't fit into any of the above categories, here are a few to check out: Language Models GPT-3 - GPT-3 is an AI-powered language model developed by OpenAI. It can generate human-like text, answer questions, and even write code. BERT - BERT is a language model developed by Google AI. It is trained on a massive dataset of text and code, and can be used for a variety of tasks, including natural language understanding, question answering, and text classification. LLama 2 - LLama 2 models are a collection of pretrained and fine-tuned large language models developed and released by Meta AI . These models are built upon the success of LLama 1 and provide significant improvements, including a larger scale and more extensive context. Claude - Claude is an AI assistant developed by Anthropic that excels at analysis, writing, and coding tasks. PaLM 2 - PaLM 2 is Google's next-generation language model with improved multilingual, reasoning, and coding capabilities. Generative Models StyleGAN - StyleGAN is an AI-powered generative model that can create high-quality images of faces, animals, and other objects. It is known for its ability to create realistic and diverse images. Generative Pre-trained Transformer 3 (GPT-3) - GPT-3 is an AI-powered language model developed by OpenAI. It can generate human-like text, answer questions, and even write code.

ai50
github
LLM Vibe Score0.457
Human Vibe Score0.07953823122984799
nahueespinosaJan 17, 2025

ai50

My work on CS50’s Introduction to AI with Python https://cs50.harvard.edu/ai/ This course explores the concepts and algorithms at the foundation of modern artificial intelligence, diving into the ideas that give rise to technologies like game-playing engines, handwriting recognition, and machine translation. Through hands-on projects, students gain exposure to the theory behind graph search algorithms, classification, optimization, reinforcement learning, and other topics in artificial intelligence and machine learning as they incorporate them into their own Python programs. By course’s end, students emerge with experience in libraries for machine learning as well as knowledge of artificial intelligence principles that enable them to design intelligent systems of their own. Certificate: https://courses.edx.org/certificates/2ec5ff3f06b24bb595c21e3821591538 Notes I've taken some notes on key concepts and algorithms throughout the lectures for future reference. Lecture 0: Search Concepts Agent: entity that perceives its environment and acts upon that environment. State: a configuration of the agent and its environment. Actions: choices that can be made in a state. Transition model: a description of what state results from performing any applicable action in any state. Path cost: numerical cost associated with a given path. Evaluation function: function that estimates the expected utility of the game from a given state. Algorithms DFS (depth first search): search algorithm that always expands the deepest node in the frontier. BFS (breath first search): search algorithm that always expands the shallowest node in the frontier. Greedy best-first search: search algorithm that expands the node that is closest to the goal, as estimated by an heuristic function h(n). A\* search: search algorithm that expands node with lowest value of the "cost to reach node" plus the "estimated goal cost". Minimax: adversarial search algorithm. Projects Degrees Tic-Tac-Toe Lecture 1: Knowledge Concepts Sentence: an assertion about the world in a knowledge representation language. Knowledge base: a set of sentences known by a knowledge-based agent. Entailment: a entails b if in every model in which sentence a is true, sentence b is also true. Inference: the process of deriving new sentences from old ones. Conjunctive normal form: logical sentence that is a conjunction of clauses. First order logic: Propositional logic. Second order logic: Proposition logic with universal and existential quantification. Algorithms Model checking: enumerate all possible models and see if a proposition is true in every one of them. Conversion to CNF and Inference by resolution Projects Knights Minesweeper Lecture 2: Uncertainty Concepts Unconditional probability: degree of belief in a proposition in the absence of any other evidence. Conditional probability: degree of belief in a proposition given some evidence that has already been revealed. Random variable: a variable in probability theory with a domain of possible values it can take on. Independence: the knowledge that one event occurs does not affect the probability of the other event. Bayes' Rule: P(a) P(b|a) = P(b) P(a|b) Bayesian network: data structure that represents the dependencies among random variables. Markov assumption: the assumption that the current state depends on only a finite fixed number of previous states. Markov chain: a sequence of random variables where the distribution of each variable follows the Markov assumption. Hidden Markov Model: a Markov model for a system with hidden states that generate some observed event. Algorithms Inference by enumeration Sampling Likelihood weighting Projects Heredity PageRank Lecture 3: Optimization Concepts Optimization: choosing the best option from a set of options. Algorithms Local Search Hill climbing steepest-ascent: choose the highest-valued neighbor. stochastic: choose randomly from higher-valued neighbors. first-choice: choose the first higher-valued neighbor. random-restart: conduct hill climbing multiple times. local beam search: chooses the k highest-valued neighbors. Simulated annealing: early on, more likely to accept worse-valued neighbors than the current state. Linear programming Simplex Interior-Point Constraint satisfaction problems Arc consistency: to make X arc-consistent with respect to Y, removing elements from X's domain until every choice for X has a possible choice for Y Backtracking search Projects Crossword Lecture 4: Learning Concepts Supervised learning: given a data set of input-output pairs, learn a function to map inputs to outputs. Classification: supervised learning task of learning a function mapping an input point to a discrete category. Regression: supervised learning task of learning a function mapping and input point to a continuous value. Loss function: function that express how poorly our hypothesis performs (L1, L2). Overfitting: when a model fits too closely to a particular data set and therefore may fail to generalize to future data. Regularization: penalizing hypotheses that are more complex to favor simpler, more general hypotheses. Holdout cross-validation: splitting data into a training set and a test set, such that learning happens on the training set and is evaluated on the test set. k-fold cross-validation: splitting data into k sets, and experimenting k times, using each set as a test set once, and using remaining data as training set. Reinforcement learning: given a set of rewards or punishments, learn what actions to take in the future. Unsupervised learning: given input data without any additional feedback, learn patterns. Clustering: organizing a set of objects into groups in such a way that similar objects tend to be in the same group. Algorithms k-nearest-neighbor classification: given an input, chooses the most common class out of the k nearest data points to that input. Support Vector Machines (SVM) Markov decision process: model for decision-making, representing states, actions and their rewards. Q-learning: method for learning a function Q(s, a), estimate of the value of performing action a in state s. Greedy decision-making epsilon-greedy k-means clustering: clustering data based on repeatedly assigning points to clusters and updating those clusters' centers. Projects Shopping Nim Lecture 5: Neural Networks Concepts Artificial neural network: mathematical model for learning inspired by biological neural networks. Multilayer neural network: artificial neural network with an input layer, an output layer, and at least one hidden layer. Deep neural network: neural network with multiple hidden layer. Dropout: temporarily removing units - selected at random - from a neural network to prevent over-reliance on certain units. Image convolution: applying a filter that adds each pixel value of an image to its neighbors, weighted according to a kernel matrix. Pooling: reducing the size of an input by sampling from regions in the input. Convolutional neural network: neural networks that use convolution, usually for analyzing images. Recurrent neural network: neural network that generates output that feeds back into its own inputs. Algorithms Gradient descent: algorithm for minimizing loss when training neural network. Backpropagation: algorithm for training neural networks with hidden layers. Projects Traffic Lecture 6: Language Concepts Natural language processing n-gram: a continuous sequence of n items inside of a text. Tokenization: the task of splitting a sequence of characters into pieces (tokens). Text Categorization Bag-of-words model: represent text as an unordered collection of words. Information retrieval: the task of finding relevant documents in response to a user query. Topic modeling: models for discovering the topics for a set of documents. Term frequency: number of times a term appears in a document. Function words: words that have little meaning on their own, but are used to grammatically connect other words. Content words: words that carry meaning independently. Inverse document frequency: measure of how common or rare a word is across documents. Information extraction: the task of extracting knowledge from documents. WordNet: a lexical database of semantic relations between words. Word representation: looking for a way to represent the meaning of a word for further processing. one-hot: representation of meaning as a vector with a single 1, and with other values as 0. distribution: representation of meaning distributed across multiple values. Algorithms Markov model applied to language: generating the next word based on the previous words and a probability. Naive Bayes: based on the Bayes' Rule to calculate probability of a text being in a certain category, given it contains specific words. Assuming every word is independent of each other. Additive smoothing: adding a value a to each value in our distribution to smooth the data. Laplace smoothing: adding 1 to each value in our distribution (pretending we've seen each value one more time than we actually have). tf-idf: ranking of what words are important in a document by multiplying term frequency (TF) by inverse document frequency (IDF). Automated template generation: giving AI some terms and let it look into a corpus for patterns where those terms show up together. Then it can use those templates to extract new knowledge from the corpus. word2vec: model for generating word vectors. skip-gram architecture: neural network architecture for predicting context words given a target word. Projects Parser Questions

teach-AI-in-business
github
LLM Vibe Score0.443
Human Vibe Score0.018525334165293606
aenyneJan 9, 2025

teach-AI-in-business

Teaching AI in Business ![HitCount] I am collecting material for teaching AI-related issues to non-tech people. The links should provide for a general understanding of AI without going too deep into technical issues. Please contribute! Make this Issue your First Issue I am collecting material for teaching AI-related issues to non-tech people. The links should have provide for a general understanding of AI without going too deep into technical issues. Please contribute! Kindly use only those Resources with NO CODE NEW Check out also the AI Wiki NEW Online Videos & Courses | Link to Issue | Description | |---|---| | Top Trending Technologies | Youtube Channel to master top trending technologyies including artificial intelligence | | AI4All | AI 4 All is a resource for AI facilitators to bring AI to scholars and students | | Elements of AI | Elements of AI is a free open online course to teach AI principles | | Visual Introduction to Machine Learning | Visual introduction to Machine Learning is a beautiful website that gives a comprehensive introduction and easily understood first encounter with machine learning | | CS50's Introduction to Artificial Intelligence with Python | Learn to use machine learning in Python in this introductory course on artificial intelligence.| | Crash course for AI | This is a fun video series that introduces students and educators to Artificial Intelligence and also offers additional more advanced videos. Learn about the basics, neural networks, algorithms, and more. | Youtuber Channel Machine Learning Tutorial | Youtube Channel Turorial Teachable Machine for beginner | | Artificial Intelligence (AI) |Learn the fundamentals of Artificial Intelligence (AI), and apply them. Design intelligent agents to solve real-world problems including, search, games, machine learning, logic, and constraint satisfaction problems | | AI For Everyone by Andrew Ng | AI For Everyone is a course especially for people from a non-technical background to understand AI strategies | | How far is too far? The age of AI| This is a Youtube Orignals series by Robert Downey| | Fundamentals of Artificial Intelligence|This course is for absolute beginners with no technical knowledge.| | Bandit Algorithm (Online Machine Learning)|No requirement of technical knowledge, but a basic understending of Probability Ttheory would help| | An Executive's Guide to AI|This is an interactive guide to teaching business professionals how they might employ artificial intelligence in their business| | AI Business School|Series of videos that teach how AI may be incorporated in various business industries| | Artificial Intelligence Tutorial for Beginners | This video will provide you with a comprehensive and detailed knowledge of Artificial Intelligence concepts with hands-on examples. | | Indonesian Machine Learning Tutorial | Turorial Teachable Machine to train a computer for beginner | | Indonesian Youtube Playlist AI Tutorial | Youtube Playlist AI Tutorial For Beginner | | Artificial Intelligence Search Methods For Problem Solving By Prof. Deepak Khemani|These video lectures are for absolute beginners with no technical knowledge| | AI Basics Tutorial | This video starts from the very basics of AI and ML, and finally has a hands-on demo of the standard MNIST Dataset Number Detection model using Keras and Tensorflow.| | Simple brain.js Tutorial | This video explains a very simple javascript AI library called brain.js so you can easily run AI in the browser.| | Google AI| A complete kit for by google official for non-tech guy to start all over from basics, till advanced | | Microsoft AI for Beginners| A self-driven curriculum by Microsoft, which includes 24 lessons on AI. | Train Your Own AI | Link to Issue | Description | |---|---| | Teachable Machine | Use Teachable Machine to train a computer to recognize your own images, sounds, & poses | | eCraft2Learn | Resource and interactive space (Snap, a visual programming environment like Scratch) to learn how to create AI programs | | Google Quick Draw | Train an AI to guess from drawings| | Deepdream Generator| Merge Pictures to Deep Dreams using the Deepdream Generator| | Create ML|Quickly build and train Core ML models on your Mac with no code.| | What-If Tool|Visually probe the behavior of trained machine learning models, with minimal coding.| | Metaranx|Use and build artificial intelligence tools to analyze and make decisions about your data. Drag-and-drop. No code.| | obviously.ai|The total process of building ML algorithms, explaining results, and predicting outcomes in one single click.| Articles | By & Title | Description | |---|---| | Artificial Intelligence | Wikipedia Page of AI | | The Non-Technical AI Guide | One of the good blog post that could help AI more understandable for people without technical background | | LIAI | A detailed introduction to AI and neural networks | | Layman's Intro | A layman's introduction to AI | | AI and Machine Learning: A Nontechnical Overview | AI and Machine Learning: A Nontechnical Overview from OREILLY themselves is a guide to learn anyone everything they need to know about AI, focussed on non-tech people | | What business leaders need to know about artifical intelligence|Short article that summarizes the essential aspects of AI that business leaders need to understand| | How Will No-Code Impact the Future of Conversational AI | A humble explanation to the current state of converstational AI i.e.Chatbots and how it coul evolve with the current trend of no coding. | | Investopedia | Basic explanation of what AI is in a very basic and comprehensive way | | Packtpub | A non programmer’s guide to learning Machine learning | | Builtin | Artificial Intelligence.What is Artificial Intelligence? How Does AI Work? | | Future Of Life | Benefits & Risks of Artificial Intelligence | | NSDM India -Arpit | 100+ AI Tools For Non-Coders That Will Make Your Marketing Better. | | AI in Marketing for Startups & Non-technical Marketers | A practical guide for non-technical people | | Blog - Machine Learning MAstery | Blogs and Articles by Jason Browniee on ML | | AI Chatbots without programming| Chatbots are increasingly in demand among global businesses. This course will teach you how to build, analyze, deploy and monetize chatbots - with the help of IBM Watson and the power of AI.| Book Resources for Further Reading | Author | Book | Description & Notes | |---|---|---| | Ethem Alpaydin|Machine Learning: The New AI | Graph Theory with Applications to Engineering & Computer Science. A concise overview of machine learning—computer programs that learn from data—which underlies applications that include recommendation systems, face recognition, and driverless cars. | | Charu C. Aggarwal| Neural Networks and Deep Learning | This book covers both classical and modern models in deep learning. The primary focus is on the theory and algorithms of deep learning. The book is also rich in discussing different applications in order to give the practitioner a flavor of how neural architectures are designed for different types of problems. | | Hal Daumé III | A Course in Machine Learning | The purpose of this book is to provide a gentle and pedagogically organized introduction to the field. A second goal of this book is to provide a view of machine learning that focuses on ideas and models, not on math. | | Ian Goodfellow and Yoshua Bengio and Aaron Courville| Deep Learning | The book starts with a discussion on machine learning basics, including the applied mathematics and algorithms needed to effectively study deep learning from an academic perspective. There is no code covered in the book, making it perfect for a non-technical AI enthusiast. | | Peter Harrington|Machine Learning in Action| (Source: https://github.com/kerasking/book-1/blob/master/ML%20Machine%20Learning%20in%20Action.pdf) This book acts as a guide to walk newcomers through the techniques needed for machine learning as well as the concepts behind the practices.| | Jeff Heaton| Artificial Intelligence for Humans |This book helps its readers get an overview and understanding of AI algorithms. It is meant to teach AI for those who don’t have an extensive mathematical background. The readers need to have only a basic knowledge of computer programming and college algebra.| | John D. Kelleher, Brian Mac Namee and Aoife D'Arcy|Fundamentals of Machine Learning for Predictive Data Analytics: Algorithms, Worked Examples, and Case Studies (The MIT Press)|This book covers all the fundamentals of machine learning, diving into the theory of the subject and using practical applications, working examples, and case studies to drive the knowledge home.| | Deepak Khemani| [A First Course in Artificial Intelligence] | It is an introductory course on Artificial Intelligence, a knowledge-based approach using agents all across and detailed, well-structured algorithms with proofs. This book mainly follows a bottom-up approach exploring the basic strategies needed problem-solving on the intelligence part. | | Maxim Lapan | Deep Reinforcement Learning Hands-On - Second Edition | Deep Reinforcement Learning Hands-On, Second Edition is an updated and expanded version of the bestselling guide to the very latest reinforcement learning (RL) tools and techniques. It provides you with an introduction to the fundamentals of RL, along with the hands-on ability to code intelligent learning agents to perform a range of practical tasks. | | Tom M Mitchell | Machine Learning | This book covers the field of machine learning, which is the study of algorithms that allow computer programs to automatically improve through experience. The book is intended to support upper level undergraduate and introductory level graduate courses in machine learning. | | John Paul Mueller and Luca Massaron|Machine Learning For Dummies|This book aims to get readers familiar with the basic concepts and theories of machine learning and how it applies to the real world. And "Dummies" here refers to absolute beginners with no technical background.The book introduces a little coding in Python and R used to teach machines to find patterns and analyze results. From those small tasks and patterns, we can extrapolate how machine learning is useful in daily lives through web searches, internet ads, email filters, fraud detection, and so on. With this book, you can take a small step into the realm of machine learning and we can learn some basic coding in Pyton and R (if interested)| | Michael Nielsen| Neural Networks and Deep Learning |Introduction to the core principles of Neural Networks and Deep Learning in AI| | Simon Rogers and Mark Girolami| A Course in Machine Learning |A First Course in Machine Learning by Simon Rogers and Mark Girolami is the best introductory book for ML currently available. It combines rigor and precision with accessibility, starts from a detailed explanation of the basic foundations of Bayesian analysis in the simplest of settings, and goes all the way to the frontiers of the subject such as infinite mixture models, GPs, and MCMC.| |Peter Norvig| Paradigm of Artificial Intelligence Programming |Paradigms of AI Programming is the first text to teach advanced Common Lisp techniques in the context of building major AI systems. By reconstructing authentic, complex AI programs using state-of-the-art Common Lisp, the book teaches students and professionals how to build and debug robust practical programs, while demonstrating superior programming style and important AI concepts.| | Stuart Russel & Peter Norvig | Artificial Intelligence: A Modern Approach, 3rd Edition | This is the prescribed text book for my Introduction to AI university course. It starts off explaining all the basics and definitions of what AI is, before launching into agents, algorithms, and how to apply them. Russel is from the University of California at Berkeley. Norvig is from Google.| | Richard S. Sutton and Andrew G. Barto| Reinforcement Learning: An Introduction |Reinforcement learning, one of the most active research areas in artificial intelligence, is a computational approach to learning whereby an agent tries to maximize the total amount of reward it receives while interacting with a complex, uncertain environment.| | Alex Smola and S.V.N. Vishwanathan | Introduction to Machine Learning | Provides the reader with an overview of the vast applications of ML, including some basic tools of statistics and probability theory. Also includes discussions on sophisticated ideas and concepts. | | Shai Shalev-Shwartz and Shai Ben-David | Understanding Machine Learning From Theory to Algorithms |The primary goal of this book is to provide a rigorous, yet easy to follow, introduction to the main concepts underlying machine learning. | | Chandra S.S.V | Artificial Intelligence and Machine Learning | This book is primarily intended for undergraduate and postgraduate students of computer science and engineering. This textbook covers the gap between the difficult contexts of Artificial Intelligence and Machine Learning. It provides the most number of case studies and worked-out examples. In addition to Artificial Intelligence and Machine Learning, it also covers various types of learning like reinforced, supervised, unsupervised and statistical learning. It features well-explained algorithms and pseudo-codes for each topic which makes this book very useful for students. | | Oliver Theobald|Machine Learning For Absolute Beginners: A Plain English Introduction|This is an absolute beginners ML guide.No mathematical background is needed, nor coding experience — this is the most basic introduction to the topic for anyone interested in machine learning.“Plain” language is highly valued here to prevent beginners from being overwhelmed by technical jargon. Clear, accessible explanations and visual examples accompany the various algorithms to make sure things are easy to follow.| | Tom Taulli | Artificial Intelligence Basics: A Non-Technical Introduction | This book equips you with a fundamental grasp of Artificial Intelligence and its impact. It provides a non-technical introduction to important concepts such as Machine Learning, Deep Learning, Natural Language Processing, Robotics and more. Further the author expands on the questions surrounding the future impact of AI on aspects that include societal trends, ethics, governments, company structures and daily life. | |Cornelius Weber, Mark Elshaw, N. Michael Mayer| Reinforcement Learning |Learning is a very important aspect. This book is on reinforcement learning which involves performing actions to achieve a goal. The first 11 chapters of this book describe and extend the scope of reinforcement learning.| |John D. Kelleher, Brian Mac Namee, Aoife D'arcy| Algorithms, Worked Examples, and Case Studies | A comprehensive introduction to the most important machine learning approaches used in predictive data analytics, covering both theoretical concepts and practical applications. |

ai-learning-roadmap
github
LLM Vibe Score0.442
Human Vibe Score0.035708035270567436
gopala-krNov 30, 2024

ai-learning-roadmap

Lists of all AI related learning materials and practical tools to get started with AI apps Design Thinking – An Introduction Stanford's virtual Crash Course in Design Thinking Amazon Web Services Learning Material AWS AI Session– The session provides an overview of all Amazon AI technology offerings (Lex, Polly, Rekognition, ML, and Deep Learning AMI) Self-Paced Labs AWS self-paced labs provide hands-on practice in a live AWS environment with AWS services and real-world cloud scenarios. Follow step-by-step instructions to learn a service, practice a use case, or prepare for AWS Certification. Introductory Lab Introduction to AWS Lambda Lex Introduction to Amazon Lex Amazon Lex Webinar Amazon Lex: AWS conversational interface (chat bot) Documentation Polly Introduction to Amazon Polly Amazon Polly Webinar - Amazon Polly – AWS Text To Speech (TTS) service Documentation What is Amazon Polly? Developer Resources Rekognition Introduction to Amazon Rekognition Amazon Rekognition - Deep Learning-Based Image Analysis Webinar Amazon Rekognition – AWS image recognition service Documentation – What is Amazon Rekognition? Machine Learning Machine Learning Session 1 – Empowering Developers to Build Smart Applications Session 2 - Predicting Customer Churn with Amazon Machine Learning AWS Machine Learning – End to end, managed service for creating and testing ML models and then deploying those models into production Documentation What is Amazon Machine Learning? Developer Resources AWS Deep Learning AMI – Amazon Machine Image (AMI) optimized for deep learning efforts Recommended Additional Resources Take your skills to the next level with fundamental, advanced, and expert level labs. Creating Amazon EC2 Instances with Microsoft Windows Building Your First Amazon Virtual Private Cloud (VPC) Working with AWS CodeCommit on Windows Working with Amazon DynamoDB Google Cloud - Learning Material Below is the learning material that will help you learn about Google Cloud. Network Networking 101 – 43 mins The codelab provides common cloud developer experience as follows: Set up your lab environment and learn how to work with your GCP environment. Use of common open source tools to explore your network around the world. Deploy a common use case: use of HTTP Load Balancing and Managed Instance Groups to host a scalable, multi-region web server. Testing and monitoring your network and instances. Cleanup. Developing Solutions for Google Cloud Platform – 8 hours Infrastructure Build a Slack Bot with Node.js on Kubernotes – 43 mins Creating a Virtual Machine – 10 mins Getting Started with App Engine (Python) – 13 mins Data Introduction to Google Cloud Data Prep – 7 mins Create a Managed MySQL database with Cloud SQL – 19 mins Upload Objects to Cloud Storage – 11 mins AI, Big Data & Machine Learning Introduction to Google Cloud Machine Learning – 1 hour Machine Learning APIs by Example – 30 min Google Cloud Platform Big Data and Machine Learning Fundamentals Additional AI Materials Auto-awesome: Advanced Data Science on Google Cloud Platform – 45 min Run a Big Data Text Processing Pipeline in Cloud Dataflow – 21 min Image Classification Using Cloud ML Engine & Datalab – 58 min Structured Data Regression Using Cloud ML Engine & Datalab – 58 min (Optional) Deep Learning & Tensorflow Tensorflow and Deep Learning Tutorial – 2:35 hours Deep Learning Course – advanced users only Additional Reference Material Big Data & Machine Learning @ Google Cloud Next '17 - A collection of 49 videos IBM Watson Learning Material (Contributions are welcome in this space) [IBM Watson Overview]() [IBM Watson Cognitive APIs]() [IBM Watson Knowledge Studio]() Visual Studio UCI datasets Microsoft Chat Bots Learning Material Skills Prerequisite Git and Github NodeJS VS Code IDE Training Paths If you have the above Prerequisite skills, then take Advanced Training Path else take Novice Training Path. Prerequisite Tutorials Git and Github Node.js Node.js Tutorials for Beginners Node.js Tutorial in VS Code Introduction To Visual Studio Code Novice Training Path Environment Set Up Download and Install Git Set up GitHub Account_ Download and Install NodeJS Download and Install IDE - Visual Studio Code Download and Install the Bot Framework Emulator Git clone the Bot Education project - git clone Set Up Azure Free Trial Account Cognitive Services (Defining Intelligence) Read Cognitive Services ADS Education Deck – git clone Review the guide for Understanding Natural language with LUIS Complete the NLP (LUIS) Training Lab from the installed Bot Education project – \bot-education\Student-Resources\Labs\CognitiveServices\Lab_SetupLanguageModel.md Bot Framework (Building Chat Bots) Read Bot Framework ADS Education Deck from downloaded - (Your Path)\bot-extras Review Bot Framework documentation (Core Concepts, Bot Builder for NodeJS, and Bot Intelligence) - Setup local environment and run emulator from the installed Bot Education project – \bot-education\Student-Resources\Labs\Node\Lab1_SetupCheckModel.md Review and test in the emulator the “bot-hello” from \bot-education\Student-Resources\BOTs\Node\bot-hello Advanced Training Path Environment Set Up Download and Install Git Set up GitHub Account_ Download and Install NodeJS Download and Install IDE - Visual Studio Code Download and Install the Bot Framework Emulator Git clone the Bot Education project - git clone Set Up Azure Free Trial Account Git clone the Bot Builder Samples – git clone Cognitive Services (Defining Intelligence) Read Cognitive Services ADS Education Deck – git clone Review the guide for Understanding Natural language with LUIS Bot Framework (Building Chat Bots) Read Bot Framework ADS Education Deck from downloaded - (Your Path)\bot-extras Review Bot Framework documentation (Core Concepts, Bot Builder for NodeJS, and Bot Intelligence) - Setup local environment and run emulator from the installed Bot Education project – \bot-education\Student-Resources\Labs\Node\Lab1_SetupCheckModel.md Cognitive Services (Defining Intelligence) - Labs Complete the NLP (LUIS) Training Lab from the installed BOT Education project \bot-education\Student-Resources\Labs\CognitiveServices\Lab_SetupLanguageModel.md Review, Deploy and run the LUIS BOT sample Bot Framework (Building Chat Bots) – Labs Setup local environment and run emulator from the installed Bot Education project \bot-education\Student-Resources\Labs\Node\Lab1_SetupCheckModel.md Review and test in the emulator the “bot-hello” from \bot-education\Student-Resources\BOTs\Node\bot-hello Review and test in the emulator the “bot-recognizers” from \bot-education\Student-Resources\BOTs\Node\bot-recognizers Lecture Videos Source Berkeley Lecture TitleLecturerSemester Lecture 1 Introduction Dan Klein Fall 2012 Lecture 2 Uninformed Search Dan Klein Fall 2012 Lecture 3 Informed Search Dan Klein Fall 2012 Lecture 4 Constraint Satisfaction Problems I Dan Klein Fall 2012 Lecture 5 Constraint Satisfaction Problems II Dan Klein Fall 2012 Lecture 6 Adversarial Search Dan Klein Fall 2012 Lecture 7 Expectimax and Utilities Dan Klein Fall 2012 Lecture 8 Markov Decision Processes I Dan Klein Fall 2012 Lecture 9 Markov Decision Processes II Dan Klein Fall 2012 Lecture 10 Reinforcement Learning I Dan Klein Fall 2012 Lecture 11 Reinforcement Learning II Dan Klein Fall 2012 Lecture 12 Probability Pieter Abbeel Spring 2014 Lecture 13 Markov Models Pieter Abbeel Spring 2014 Lecture 14 Hidden Markov Models Dan Klein Fall 2013 Lecture 15 Applications of HMMs / Speech Pieter Abbeel Spring 2014 Lecture 16 Bayes' Nets: Representation Pieter Abbeel Spring 2014 Lecture 17 Bayes' Nets: Independence Pieter Abbeel Spring 2014 Lecture 18 Bayes' Nets: Inference Pieter Abbeel Spring 2014 Lecture 19 Bayes' Nets: Sampling Pieter Abbeel Fall 2013 Lecture 20 Decision Diagrams / Value of Perfect Information Pieter Abbeel Spring 2014 Lecture 21 Machine Learning: Naive Bayes Nicholas Hay Spring 2014 Lecture 22 Machine Learning: Perceptrons Pieter Abbeel Spring 2014 Lecture 23 Machine Learning: Kernels and Clustering Pieter Abbeel Spring 2014 Lecture 24 Advanced Applications: NLP, Games, and Robotic Cars Pieter Abbeel Spring 2014 Lecture 25 Advanced Applications: Computer Vision and Robotics Pieter Abbeel Spring 2014 Additionally, there are additional Step-By-Step videos which supplement the lecture's materials. These videos are listed below: Lecture TitleLecturerNotes SBS-1 DFS and BFS Pieter Abbeel Lec: Uninformed Search SBS-2 A* Search Pieter Abbeel Lec: Informed Search SBS-3 Alpha-Beta Pruning Pieter Abbeel Lec: Adversarial Search SBS-4 D-Separation Pieter Abbeel Lec: Bayes' Nets: Independence SBS-5 Elimination of One Variable Pieter Abbeel Lec: Bayes' Nets: Inference SBS-6 Variable Elimination Pieter Abbeel Lec: Bayes' Nets: Inference SBS-7 Sampling Pieter Abbeel Lec: Bayes' Nets: Sampling SBS-8 Gibbs' Sampling Michael Liang Lec: Bayes' Nets: Sampling --> SBS-8 Maximum Likelihood Pieter Abbeel Lec: Machine Learning: Naive Bayes SBS-9 Laplace Smoothing Pieter Abbeel Lec: Machine Learning: Naive Bayes SBS-10 Perceptrons Pieter Abbeel Lec: Machine Learning: Perceptrons Per-Semester Video Archive(Berkeley) The lecture videos from the most recent offerings are posted below. Spring 2014 Lecture Videos Fall 2013 Lecture Videos Spring 2013 Lecture Videos Fall 2012 Lecture Videos Spring 2014 Lecture TitleLecturerNotes Lecture 1 Introduction Pieter Abbeel Lecture 2 Uninformed Search Pieter Abbeel Lecture 3 Informed Search Pieter Abbeel Lecture 4 Constraint Satisfaction Problems I Pieter Abbeel Recording is a bit flaky, see Fall 2013 Lecture 4 for alternative Lecture 5 Constraint Satisfaction Problems II Pieter Abbeel Lecture 6 Adversarial Search Pieter Abbeel Lecture 7 Expectimax and Utilities Pieter Abbeel Lecture 8 Markov Decision Processes I Pieter Abbeel Lecture 9 Markov Decision Processes II Pieter Abbeel Lecture 10 Reinforcement Learning I Pieter Abbeel Lecture 11 Reinforcement Learning II Pieter Abbeel Lecture 12 Probability Pieter Abbeel Lecture 13 Markov Models Pieter Abbeel Lecture 14 Hidden Markov Models Pieter Abbeel Recording is a bit flaky, see Fall 2013 Lecture 18 for alternative Lecture 15 Applications of HMMs / Speech Pieter Abbeel Lecture 16 Bayes' Nets: Representation Pieter Abbeel Lecture 17 Bayes' Nets: Independence Pieter Abbeel Lecture 18 Bayes' Nets: Inference Pieter Abbeel Lecture 19 Bayes' Nets: Sampling Pieter Abbeel Unrecorded, see Fall 2013 Lecture 16 Lecture 20 Decision Diagrams / Value of Perfect Information Pieter Abbeel Lecture 21 Machine Learning: Naive Bayes Nicholas Hay Lecture 22 Machine Learning: Perceptrons Pieter Abbeel Lecture 23 Machine Learning: Kernels and Clustering Pieter Abbeel Lecture 24 Advanced Applications: NLP, Games, and Robotic Cars Pieter Abbeel Lecture 25 Advanced Applications: Computer Vision and Robotics Pieter Abbeel Lecture 26 Conclusion Pieter Abbeel Unrecorded Fall 2013 Lecture TitleLecturerNotes Lecture 1 Introduction Dan Klein Lecture 2 Uninformed Search Dan Klein Lecture 3 Informed Search Dan Klein Lecture 4 Constraint Satisfaction Problems I Dan Klein Lecture 5 Constraint Satisfaction Problems II Dan Klein Lecture 6 Adversarial Search Dan Klein Lecture 7 Expectimax and Utilities Dan Klein Lecture 8 Markov Decision Processes I Dan Klein Lecture 9 Markov Decision Processes II Dan Klein Lecture 10 Reinforcement Learning I Dan Klein Lecture 11 Reinforcement Learning II Dan Klein Lecture 12 Probability Pieter Abbeel Lecture 13 Bayes' Nets: Representation Pieter Abbeel Lecture 14 Bayes' Nets: Independence Dan Klein Lecture 15 Bayes' Nets: Inference Pieter Abbeel Lecture 16 Bayes' Nets: Sampling Pieter Abbeel Lecture 17 Decision Diagrams / Value of Perfect Information Pieter Abbeel Lecture 18 Hidden Markov Models Dan Klein Lecture 19 Applications of HMMs / Speech Dan Klein Lecture 20 Machine Learning: Naive Bayes Dan Klein Lecture 21 Machine Learning: Perceptrons Dan Klein Lecture 22 Machine Learning: Kernels and Clustering Pieter Abbeel Lecture 23 Machine Learning: Decision Trees and Neural Nets Pieter Abbeel Lecture 24 Advanced Applications: NLP and Robotic Cars Dan Klein Unrecorded, see Spring 2013 Lecture 24 Lecture 25 Advanced Applications: Computer Vision and Robotics Pieter Abbeel Lecture 26 Conclusion Dan Klein,Pieter Abbeel Unrecorded Spring 2013 Lecture TitleLecturerNotes Lecture 1 Introduction Pieter Abbeel Video Down Lecture 2 Uninformed Search Pieter Abbeel Lecture 3 Informed Search Pieter Abbeel Lecture 4 Constraint Satisfaction Problems I Pieter Abbeel Lecture 5 Constraint Satisfaction Problems II Pieter Abbeel Unrecorded, see Fall 2012 Lecture 5 Lecture 6 Adversarial Search Pieter Abbeel Lecture 7 Expectimax and Utilities Pieter Abbeel Lecture 8 Markov Decision Processes I Pieter Abbeel Lecture 9 Markov Decision Processes II Pieter Abbeel Lecture 10 Reinforcement Learning I Pieter Abbeel Lecture 11 Reinforcement Learning II Pieter Abbeel Lecture 12 Probability Pieter Abbeel Lecture 13 Bayes' Nets: Representation Pieter Abbeel Lecture 14 Bayes' Nets: Independence Pieter Abbeel Lecture 15 Bayes' Nets: Inference Pieter Abbeel Lecture 16 Bayes' Nets: Sampling Pieter Abbeel Lecture 17 Decision Diagrams / Value of Perfect Information Pieter Abbeel Lecture 18 Hidden Markov Models Pieter Abbeel Lecture 19 Applications of HMMs / Speech Pieter Abbeel Lecture 20 Machine Learning: Naive Bayes Pieter Abbeel Lecture 21 Machine Learning: Perceptrons I Nicholas Hay Lecture 22 Machine Learning: Perceptrons II Pieter Abbeel Lecture 23 Machine Learning: Kernels and Clustering Pieter Abbeel Lecture 24 Advanced Applications: NLP and Robotic Cars Pieter Abbeel Lecture 25 Advanced Applications: Computer Vision and Robotics Pieter Abbeel Lecture 26 Conclusion Pieter Abbeel Unrecorded Fall 2012 Lecture TitleLecturerNotes Lecture 1 Introduction Dan Klein Lecture 2 Uninformed Search Dan Klein Lecture 3 Informed Search Dan Klein Lecture 4 Constraint Satisfaction Problems I Dan Klein Lecture 5 Constraint Satisfaction Problems II Dan Klein Lecture 6 Adversarial Search Dan Klein Lecture 7 Expectimax and Utilities Dan Klein Lecture 8 Markov Decision Processes I Dan Klein Lecture 9 Markov Decision Processes II Dan Klein Lecture 10 Reinforcement Learning I Dan Klein Lecture 11 Reinforcement Learning II Dan Klein Lecture 12 Probability Pieter Abbeel Lecture 13 Bayes' Nets: Representation Pieter Abbeel Lecture 14 Bayes' Nets: Independence Pieter Abbeel Lecture 15 Bayes' Nets: Inference Pieter Abbeel Lecture 16 Bayes' Nets: Sampling Pieter Abbeel Lecture 17 Decision Diagrams / Value of Perfect Information Pieter Abbeel Lecture 18 Hidden Markov Models Pieter Abbeel Lecture 19 Applications of HMMs / Speech Dan Klein Lecture 20 Machine Learning: Naive Bayes Dan Klein Lecture 21 Machine Learning: Perceptrons Dan Klein Lecture 22 Machine Learning: Kernels and Clustering Dan Klein Lecture 23 Machine Learning: Decision Trees and Neural Nets Pieter Abbeel Lecture 24 Advanced Applications: Computer Vision and Robotics Pieter Abbeel Lecture 25 Advanced Applications: NLP and Robotic Cars Dan Klein,Pieter Abbeel Unrecorded Lecture 26 Conclusion Dan Klein,Pieter Abbeel Unrecorded Lecture Slides Here is the complete set of lecture slides, including videos, and videos of demos run in lecture: Slides [~3 GB]. The list below contains all the lecture powerpoint slides: Lecture 1: Introduction Lecture 2: Uninformed Search Lecture 3: Informed Search Lecture 4: CSPs I Lecture 5: CSPs II Lecture 6: Adversarial Search Lecture 7: Expectimax Search and Utilities Lecture 8: MDPs I Lecture 9: MDPs II Lecture 10: Reinforcement Learning I Lecture 11: Reinforcement Learning II Lecture 12: Probability Lecture 13: Markov Models Lecture 14: Hidden Markov Models Lecture 15: Particle Filters and Applications of HMMs Lecture 16: Bayes Nets I: Representation Lecture 17: Bayes Nets II: Independence Lecture 18: Bayes Nets III: Inference Lecture 19: Bayes Nets IV: Sampling Lecture 20: Decision Diagrams and VPI Lecture 21: Naive Bayes Lecture 22: Perceptron Lecture 23: Kernels and Clustering Lecture 24: Advanced Applications (NLP, Games, Cars) Lecture 25: Advanced Applications (Computer Vision and Robotics) Lecture 26: Conclusion The source files for all live in-lecture demos are being prepared from Berkeley AI for release Selected Research Papers Latest arxiv paper submissionson AI Peter Norvig-Teach Yourself Programming in Ten Years How to do Research At the MIT AI Lab A Roadmap towards Machine Intelligence Collaborative Filtering with Recurrent Neural Networks (2016) Wide & Deep Learning for Recommender Systems (2016) Deep Collaborative Filtering via Marginalized Denoising Auto-encoder (2015) Nonparametric bayesian multitask collaborative filtering (2013) Tensorflow: Large-scale machine learning on heterogeneous distributed systems https://infoscience.epfl.ch/record/82802/files/rr02-46.pdf Theano: A CPU and GPU math expression compiler. Caffe: Convolutional architecture for fast feature embedding Chainer: A powerful, flexible and intuitive framework of neural networks Large Scale Distributed Deep Networks Large-scale video classification with convolutional neural networks Efficient Estimation of Word Representations in Vector Space Grammar as a Foreign Language Going Deeper with Convolutions ON RECTIFIED LINEAR UNITS FOR SPEECH PROCESSING Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups. Multi-digit Number Recognition from Street View Imagery using Deep Convolutional Neural Networks google turning its lucrative web search over to AI machines Stanford Syllabus CS 20SI: Tensorflow for Deep Learning Research Crowd-Based Personalized Natural Language Explanations for Recommendations Comparative Study of Deep Learning Software Frameworks RedditML- What Are You Reading AI-Powered Social Bots(16 Jun 2017) The Many Tribes of Artificial Intelligence Source:https://medium.com/intuitionmachine/infographic-best-practices-in-training-deep-learning-networks-b8a3df1db53 The Deep Learning Roadmap Source:https://medium.com/intuitionmachine/the-deep-learning-roadmap-f0b4cac7009a Best Practices for Training Deep Learning Networks Source: https://medium.com/intuitionmachine/infographic-best-practices-in-training-deep-learning-networks-b8a3df1db53 ML/DL Cheatsheets Neural Network Architectures Source: http://www.asimovinstitute.org/neural-network-zoo/ Microsoft Azure Algorithm Flowchart Source: https://docs.microsoft.com/en-us/azure/machine-learning/machine-learning-algorithm-cheat-sheet SAS Algorithm Flowchart Source: http://blogs.sas.com/content/subconsciousmusings/2017/04/12/machine-learning-algorithm-use/ Algorithm Summary Source: http://machinelearningmastery.com/a-tour-of-machine-learning-algorithms/ Source: http://thinkbigdata.in/best-known-machine-learning-algorithms-infographic/ Algorithm Pro/Con Source: https://blog.dataiku.com/machine-learning-explained-algorithms-are-your-friend Python Algorithms Source: https://www.analyticsvidhya.com/blog/2015/09/full-cheatsheet-machine-learning-algorithms/ Python Basics Source: http://datasciencefree.com/python.pdf Source: https://www.datacamp.com/community/tutorials/python-data-science-cheat-sheet-basics#gs.0x1rxEA Numpy Source: https://www.dataquest.io/blog/numpy-cheat-sheet/ Source: http://datasciencefree.com/numpy.pdf Source: https://www.datacamp.com/community/blog/python-numpy-cheat-sheet#gs.Nw3V6CE Source: https://github.com/donnemartin/data-science-ipython-notebooks/blob/master/numpy/numpy.ipynb Pandas Source: http://datasciencefree.com/pandas.pdf Source: https://www.datacamp.com/community/blog/python-pandas-cheat-sheet#gs.S4P4T=U Source: https://github.com/donnemartin/data-science-ipython-notebooks/blob/master/pandas/pandas.ipynb Matplotlib Source: https://www.datacamp.com/community/blog/python-matplotlib-cheat-sheet Source: https://github.com/donnemartin/data-science-ipython-notebooks/blob/master/matplotlib/matplotlib.ipynb Scikit Learn Source: https://www.datacamp.com/community/blog/scikit-learn-cheat-sheet#gs.fZ2A1Jk Source: http://peekaboo-vision.blogspot.de/2013/01/machine-learning-cheat-sheet-for-scikit.html Source: https://github.com/rcompton/mlcheatsheet/blob/master/supervised_learning.ipynb Tensorflow Source: https://github.com/aymericdamien/TensorFlow-Examples/blob/master/notebooks/1Introduction/basicoperations.ipynb Pytorch Source: https://github.com/bfortuner/pytorch-cheatsheet Math Probability Source: http://www.wzchen.com/s/probability_cheatsheet.pdf Linear Algebra Source: https://minireference.com/static/tutorials/linearalgebrain4pages.pdf Statistics Source: http://web.mit.edu/~csvoss/Public/usabo/stats_handout.pdf Calculus Source: http://tutorial.math.lamar.edu/getfile.aspx?file=B,41,N

AI
github
LLM Vibe Score0.358
Human Vibe Score0.006489749001329033
MatousMarikOct 31, 2024

AI

AI I This repository contains practical tasks for the Artificial Intelligence 1 course, that is based on book by Russel and Norvig Artificial Intellignece: A Modern Approach, 4th Edition. Tasks are designed to review AI algorithms and use them to play games. Requirements All assignments will be written in python. Task were created for python 3.9 however there should not be any problems with backward compatibility. You can solve all assignments while working exclusively with python standard library, however for game visualizations you will need to install modul pygame. For installation you can use pip: python3 -m pip install -U pygame --user If you need more detailed, platform-specific instructions you can visit pygame-GettingStarted. Assignments In total there will be 5 programming assignments whose solutions will be submitted via ReCodEx. In each of them you will write an AI agent that plays suitable games for corresponding lecture topic. Moreover there will by partial assignments, in which you will need to implement algorithms, that will allow you to implement suitable agent functions, however your agent implementation can use any approach you like. | Game | Suggested Approach | | ---- | ------ | | Dino | rule-based agent | | Pac-Man | uniform-cost search | | Sokoban | A* with custom heuristics | | Cell Wars | minimax or Monte Carlo tree search | | Minesweeper | backtracking search for CSPs | Note that information provided in the early assignments is omitted in later ones.

coursera-practical-data-science-specialization
github
LLM Vibe Score0.465
Human Vibe Score0.0230635140825568
honghanhhOct 9, 2024

coursera-practical-data-science-specialization

Solutions on Practical Data Science Specialization Access all courses in the Coursera Practical Data Science Specialization Specialization offered by deeplearning.ai. This repo contains the SOLUTIONS of exercises/labs to achieve the badge. Course keynotes and solutions of related quizzes, assignments Practical Data Science Specialization on Coursera contains three courses: Course 1: Analyze Datasets and Train ML Models using AutoML Week 1: Artificial Intelligence (AI) mimics human behavior. Machine Learning (ML) is a subset of AI that uses statistical methods and algorithms that are able to learn from data without being explicitly programmed. Deep learning (DL) is a subset of machine learning that uses artificial neural networks to learn from data. AWS SageMaker --> [x] Practice Quiz: Week 1. [x] Graded External Tool: Register and visualize dataset. Week 2: Statistical Bias: Training data does not comprehensively represent the underlying problem space. Statistical Bias Causes: Activity Bias, Societal Bias, Selection Bias, Data Drift/Shift, ... Class Imbalance (CI) measures the imbalance in the number of members between different facet values. Detecting Statistical Bias by AWS SageMaker DataWrangler and AWS SageMaker Clarify. Feature Importance explains the features that make up the training data using a score. How useful or valuable the feature is relative to other features? SHAP (SHapley Additive exPlanations) --> [x] Practice Quiz: Week 2. [x] Graded External Tool: Detect data bias with Amazon SageMaker Clarify. Week 3: Data Prepreration includes Ingesting & Analyzing, Prepraring & Transforming, Training & Tuning, and Deploying & Managing. AutoML aims at automating the process of building a model. Model Hosting. --> [x] Practice Quiz: Week 3. [x] Graded External Tool: Train a model with Amazon SageMaker Autopilot. Week 4: Built-in Alogrithms in AWS SageMaker supports Classification, Regression, and Clustering problems. Text Analysis Evolution: Word2Vec (CBOW & Skip-gram), GloVe, FastText, Transformer, BlazingText, ELMo, GPT, BERT, ... --> [x] Practice Quiz: Week 4. [x] Graded External Tool: Train a text classifier using Amazon SageMaker BlazingText built-in algorithm. Course 2: Build, Train, and Deploy ML Pipelines using BERT Week 1 Feature Engineering involves converting raw data from one or more sources into meaningful features that can be used for training machine learning models. Feature Engineering Step includes feature selection, creation, and transformation. BERT is Transformer-based pretrained language models that sucessfully capture bidirectional contexts in word representation. Feature Store: centralized, reusable, discoverable. --> [x] Practice Quiz: Week 1. [x] Graded External Tool: Feature transformation with Amazon SageMaker processing job and Feature Store. Week 2 Learn how to train a customized Pretrained BERT and its variant models, debug, and profile with AWS SageMaker. --> [x] Practice Quiz: Week 2. [x] Graded External Tool: Train a review classifier with BERT and Amazon SageMaker. Week 3 MLOps builds on DevOps practices that encompass people, process, and technology. MLOps also includes considerations and practices that are really unique to machine learning workloads. --> [x] Practice Quiz: Week 3. [x] Graded External Tool: SageMaker pipelines to train a BERT-Based text classifier. Course 3: Optimize ML Models and Deploy Human-in-the-Loop Pipelines Week 1 Model Tuning aims to fit the model to the underlying data patterns in your training data and learn the best possible parameters for your model. Automatic Model Tuning includes grid search, random search, bayesian optimization, hyperband. Challenges: checkpointing, distribution training strategy. --> [x] Practice Quiz: Week 1. [x] Graded External Tool: Optimize models using Automatic Model Tuning. Week 2 [x] Practice Quiz: Week 2. [x] Graded External Tool: A/B testing, traffic shifting and autoscaling. Week 3 [x] Practice Quiz: Week 3. [x] Graded External Tool: Data labeling and human-in-the-loop pipelines with Amazon Augmented AI (A2I). Disclaimer The solutions here are ONLY FOR REFERENCE to guide you if you get stuck somewhere. Highly recommended to try out the quizzes and assignments yourselves first before referring to the solutions here. Feel free to discuss further with me on .

Best Programming Language For AI in 2024 | Intellipaat #Shorts #AI #Python
youtube
LLM Vibe Score0.371
Human Vibe Score0.61
IntellipaatAug 24, 2024

Best Programming Language For AI in 2024 | Intellipaat #Shorts #AI #Python

Curious about the Best Programming Language for AI in 2024? 🤖 In this #Shorts video, we explore the top language you should learn if you want to dive into the world of Artificial Intelligence. Whether you’re just starting out or looking to expand your skills, understanding the best tools for AI development is crucial. Watch to find out why Python continues to dominate the AI landscape and what makes it the go-to choice for developers. #BestProgrammingLanguageForAI #AI #Python #ArtificialIntelligence #ShortsVideo #ShortsFeed #ShortsFeedVideo #ShortsFeedViral #Intellipaat ✅ What makes Python the best programming language for AI in 2024? Python is considered the best programming language for AI in 2024 due to its simplicity, extensive libraries, and active community support. Its libraries like TensorFlow, PyTorch, and scikit-learn make it easier to implement complex algorithms and work with large datasets. Additionally, Python's readability and flexibility make it a favorite among developers working on AI projects, enabling rapid prototyping and development. ✅ Why is choosing the right programming language important for AI development? Choosing the right programming language is crucial for AI development because it impacts the efficiency and scalability of your projects. The right language should offer powerful tools, libraries, and frameworks that simplify AI tasks like data processing, machine learning, and natural language processing. Python, for instance, excels in these areas, making it the preferred choice for AI and ensuring that your projects are built on a solid, efficient foundation.

13 Best AI Tools For Startups & Entrepreneurs [2024]
youtube
LLM Vibe Score0.401
Human Vibe Score0.33
Business SolutionDec 15, 2023

13 Best AI Tools For Startups & Entrepreneurs [2024]

Here are the best AI tools for startups and entrepreneurs: Bubble ▶ Bubble free plan: https://businessolution.org/get/bubble/ Taskade ▶ Taskade free plan : https://businessolution.org/get/taskade/ Process Street ▶ Process Street free trial: https://businessolution.org/get/process-street/ CustomGPT ▶ Try CustomGPT for free: https://businessolution.org/get/customgpt/ MeetGeek ▶ MeetGeek free plan: https://businessolution.org/get/meetgeek-ai/ Mixo ▶ Try Mixo for free: https://businessolution.org/get/mixo/ Tidio ▶ Tidio free plan (+20% OFF): https://businessolution.org/get/tidio/ AdCreative.ai ▶ AdCreative.ai 25% OFF: https://businessolution.org/get/adcreative/ LeadFuze ▶ LeadFuze free trial: https://businessolution.org/get/leadfuze/ HubSpot ▶ HubSpot free plan: https://businessolution.org/get/hubspot/ ClickFunnels 2.0 ▶ ClickFunnels 2.0 free trial: https://businessolution.org/get/clickfunnels-2-0/ GoHire ▶ GoHire free trial: https://businessolution.org/get/gohire-2/ DeepBrain ▶ Try DeepBrain for free: https://businessolution.org/get/deepbrain/ Timestamps: 0:00 – AI Tools for Startups 0:17 – Bubble.io 2:26 – Taskade 4:35 – Process Street 6:20 – CustomGPT 7:44 – MeetGeek 8:31 – Mixo 9:09 – Tidio 10:15 – AdCreative.ai 11:34 – LeadFuze 12:51 – HubSpot 14:48 – ClickFunnels 2.0 16:10 – GoHire 17:25 – DeepBrain 👉‍ See all 17 AI tools for startups in this article: https://businessolution.org/ai-tools-for-startups/ In today's fast-paced and competitive business landscape, startups are constantly seeking innovative ways to gain a competitive edge and drive growth. Enter the realm of artificial intelligence (AI) tools for startups – a game-changing technology that holds the potential to revolutionize how new businesses operate, strategize, and scale. From automating repetitive tasks to unlocking valuable insights from data, AI tools offer startups an unprecedented opportunity to streamline operations, enhance decision-making, and deliver exceptional customer experiences. Imagine having access to intelligent algorithms that can analyze market trends, predict consumer behavior, and optimize resource allocation with unparalleled accuracy. These AI tools can empower startups to make data-driven decisions with confidence while freeing up valuable time and resources for creative problem-solving and strategic planning. By harnessing the power of AI technology, startups can navigate the complexities of today's business environment with agility, precision, and scalability like never before. Join us as we delve into the world of AI tools for startups and explore how this transformative technology is poised to reshape the entrepreneurial landscape in profound ways.

What is generative AI and how does it work? – The Turing Lectures with Mirella Lapata
youtube
LLM Vibe Score0.382
Human Vibe Score0.9
The Royal InstitutionOct 12, 2023

What is generative AI and how does it work? – The Turing Lectures with Mirella Lapata

How are technologies like ChatGPT created? And what does the future hold for AI language models? This talk was filmed at the Royal Institution on 29th September 2023, in collaboration with The Alan Turing Institute. Join this channel to get access to perks: https://www.youtube.com/channel/UCYeF244yNGuFefuFKqxIAXw/join Watch the Q&A with Mirella here: https://youtu.be/9i2x2HyeW-Y Generative AI refers to a type of artificial intelligence that involves creating new and original data or content. Unlike traditional AI models that rely on large datasets and algorithms to classify or predict outcomes, generative AI models are designed to learn the underlying patterns and structure of the data and generate novel outputs that mimic human creativity. ChatGPT is perhaps the most well-known example, but the field is far larger and more varied than text generation. Other applications of generative AI include image and video synthesis, speech generation, music composition, and virtual reality. In this lecture, Mirella Lapata will present an overview of this exciting—sometimes controversial—and rapidly evolving field. Mirella Lapata is professor of natural language processing in the School of Informatics at the University of Edinburgh. Her research focuses on getting computers to understand, reason with, and generate natural language. She is the first recipient (2009) of the British Computer Society and Information Retrieval Specialist Group (BCS/IRSG) Karen Sparck Jones award and a Fellow of the Royal Society of Edinburgh, the ACL, and Academia Europaea. 00:00 Intro 2:38 Generative AI isn’t new – so what’s changed? 8:43 How did we get to ChatGPT? 12:38 How are Large Language Models created? 22:48 How good can a LLM become? 26:57 Unexpected effects of scaling up LLMs 28:05 How can ChatGPT meet the needs of humans? 32:30 Chat GPT demo 38:07 Are Language Models always right or fair? 40:21 The impact of LLMs on society 42:54 Is AI going to kill us all? -- A very special thank you to our Patreon supporters who help make these videos happen, especially: modsiw, Anton Ragin, Edward Unthank, Robert L Winer, Andy Carpenter, William Hudson Don McLaughlin, efkinel lo, Martin Paull, Ben Wynne-Simmons, Ivo Danihelka, Kevin Winoto, Jonathan Killin, Stephan Giersche, William Billy Robillard, Jeffrey Schweitzer, Frances Dunne, jonas.app, Tim Karr, Alan Latteri, David Crowner, Matt Townsend, THOMAS N TAMADA, Andrew McGhee, Paul Brown, David Schick, Dave Ostler, Osian Gwyn Williams, David Lindo, Roger Baker, Rebecca Pan -- The Ri is on Twitter: http://twitter.com/ri_science and Facebook: http://www.facebook.com/royalinstitution and TikTok: https://www.tiktok.com/@ri_science Listen to the Ri podcast: https://podcasters.spotify.com/pod/show/ri-science-podcast Our editorial policy: https://www.rigb.org/editing-ri-talks-and-moderating-comments Subscribe for the latest science videos: http://bit.ly/RiNewsletter Product links on this page may be affiliate links which means it won't cost you any extra but we may earn a small commission if you decide to purchase through the link.

Best AI Tools for Accountants
youtube
LLM Vibe Score0.36
Human Vibe Score0.21
Miles EducationJun 28, 2023

Best AI Tools for Accountants

We’re pretty sure, these great AI tools will help you in the long run. Let’s have a look at their importance: VIC.AI: AI-powered Accounting Made Effortless! Their advanced algorithms are trained on vast invoice data, eliminating the need for templates or memorization. Accurate from day one, their Autopilot technology seamlessly integrates AI for streamlined invoicing.😇 Indy: AI-Powered Accounting Made Fast and Affordable! Freelancers, businesses, and entrepreneurs can tackle accounting tasks up to 20x faster than traditional software. Create income statements and financial statements in a fraction of the time, all at a lower cost than traditional accountants.🤔 Docyt: AI-Powered Accounting Automation for Faster Decision Making. Digitize financial data, automate workflows, and make faster decisions. Reduce costs and simplify bookkeeping and back-office tasks.😲 Blue Dot is an innovative market leader with a cutting-edge financial platform. Their all-in-one Tax Compliance Platform combines digitization, tax compliance, and automation to analyze employee spend data for VAT, Taxable Employee Benefits, and Corporate Income Tax.🥹 Comment below the names of the AI tools that you use for accounting.👇👇 #aitoolsforaccountants #aitools #ai #technology #upskill #cpa #uscpa #CPAexam #cpajobs #CPAscope #MilesEducation #workintheus #talent #fyp #explore #accounting #accountants #accountancy #cpacoursereview #jobopportunities #CPAplacement #CPAsalary #success #career

The 9 AI Skills You Need NOW to Stay Ahead of 97% of People
youtube
LLM Vibe Score0.289
Human Vibe Score0.91
AI UncoveredMay 14, 2023

The 9 AI Skills You Need NOW to Stay Ahead of 97% of People

The 9 AI Skills You Need NOW to Stay Ahead of 97% of People 🔒 Keep Your Digital Life Private: Stay Safe & Secure Online with NordVPN: https://nordvpn.com/safetyfirst Welcome to our latest educational video, "The 9 AI Skills You Need NOW to Stay Ahead of 97% of People." This video is designed for anyone eager to take a deep dive into the world of artificial intelligence and machine learning. Our goal is to provide you with the most essential AI skills needed to excel in this rapidly evolving field, keeping you ahead of the curve and well-positioned in the job market. In this comprehensive guide, we explore nine fundamental AI skills, ranging from understanding algorithms to deep learning, data science, natural language processing, computer vision, robotics, and more. We also provide practical tips on how to apply these skills in real-world scenarios, whether you're an AI enthusiast or a seasoned professional. AI is not just the future; it's here NOW. By acquiring these nine essential AI skills, you can position yourself among the top 3% of people who are ready to shape the future. Don't be left behind as AI transforms industries, from healthcare and finance to entertainment and transportation. This video is a must-watch for anyone interested in AI, machine learning, data analysis, robotics, or any related field. Whether you're just starting out, looking to upskill, or aiming to stay ahead in your career, these nine AI skills are your key to success. Remember to subscribe to our channel for more valuable content and hit the notification bell so you never miss an update. Join the conversation in the comments section - we'd love to hear your thoughts on AI and how you plan to incorporate these skills into your career or studies. So get ready, click play, and let's take a step towards the future together, learning the 9 AI skills you need NOW to stay ahead of 97% of people. Your AI journey starts here. Enjoy the video! #artificialintelligence #ai #airevolution Subscribe for more! Welcome to AI Uncovered, your ultimate destination for exploring the fascinating world of artificial intelligence! Our channel delves deep into the latest AI trends and technology, providing insights into cutting-edge AI tools, AI news, and breakthroughs in artificial general intelligence (AGI). We simplify complex concepts, making AI explained in a way that is accessible to everyone. At AI Uncovered, we're passionate about uncovering the most captivating stories in AI, including the marvels of ChatGPT and advancements by organizations like OpenAI. Our content spans a wide range of topics, from science news and AI innovations to in-depth discussions on the ethical implications of artificial intelligence. Our mission is to enlighten, inspire, and inform our audience about the rapidly evolving technology landscape. Whether you're a tech enthusiast, a professional seeking to stay ahead of AI trends, or someone curious about the future of artificial intelligence, AI Uncovered is the perfect place to expand your knowledge. Join us as we uncover the secrets behind AI tools and their potential to revolutionize our world. Subscribe to AI Uncovered and stay tuned for enlightening content that bridges the gap between AI novices and experts, covering AI news, AGI, ChatGPT, OpenAI, artificial intelligence, and more. Together, let's explore the limitless possibilities of technology and AI. Disclaimer: Some links included in this description might be affiliate links. If you purchase a product or service through the links that we provide, we may receive a small commission. There is no additional charge for you. Thank you for supporting AI Uncovered so we can continue to provide you with free, high-quality content.