Twitter author Profile picture
Apr 13 6 tweets 1 min read Twitter logo Read on Twitter
1/ A recent controversy at Google has sparked important questions about training ML models on the output of other models. Let's dive into the engineering, business, and legal aspects of this practice. Buckle up, folks! 🧵
2/ Engineering recipes for training algorithms on generated data are still evolving. Instances of using a competitor's model outputs to train your own are surfacing. Are these techniques fair game or should there be limits?
3/ Business-wise, data may not always make your business more defensible. Market leaders might spend resources gathering data, but if their product's data makes it easier for competitors to catch up, is that initial effort a strong defense?
4/ Legally, OpenAI’s terms of use forbid using their output to develop competing models. But what if someone scrapes text from ShareGPT that they didn't generate? And are such terms enforceable in light of antitrust & fair-use laws?
5/ Generative AI offers creative use cases for training models using others' outputs. As we explore these exciting trends, it's crucial to navigate the legal & ethical landscape responsibly. Stay tuned for more developments! #ResponsibleAI
💡 Follow me for more insights on AI, tech, and the ethical considerations surrounding them. Let's keep the conversation going! 🚀

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Twitter author

Twitter author Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @

Apr 13
1/ 🚀 #AutoGPT is trending today, but what's the hype all about? Let's unpack these AI game-changers, explore their potential, and recognize their limitations. Thread👇 #HypeVsReality
2/ 🎯 AutoGPTs are AI agents that can perform tasks autonomously, with little to no human intervention. They can even chain multiple GPT-4s together to work on different tasks simultaneously! However, they can get stuck & may need human help. #AutonomousAI #GPT4
3/ 🌟 Two main models dominate the AutoGPT landscape: BabyAGI by @yoheinakajima & AutoGPT by @SigGravitas. They're trending on GitHub and attracting devs worldwide. Don't forget @asimdotshrestha's AgentGPT, which runs in-browser!
Read 9 tweets
Apr 12
🚀 1/ Wanna take #AutoGPT to the next level? Say hello to @dataleapHQ the "Upwork for AI Agents" - a marketplace where you can hire AutoGPTs and other AI agents to get the job done. Buckle down for our vision paper! 🧵 dataleap.substack.com/p/ai-workforce…
@dataleapHQ 🙏 2/ Big shoutout to all the brilliant minds moving us closer to his vision:
@hwchase17, @LangChainAI & @gpt_index for the agent toolset , @yoheinakajima for sparking our imagination with #BabyAGI, @ShunyuYao12 for your ReAct thought breakthrough, and @SigGravitas for AutoGPT.
@dataleapHQ @hwchase17 @LangChainAI @gpt_index @yoheinakajima @ShunyuYao12 @SigGravitas 🤖 3/ The gig economy has traditionally been the domain of human freelancers, but at Dataleap we are looking to create a new era where AI agents and humans coexist, each leveraging their unique strengths. Sometimes they'll complement each other; other times, it's all AI.
Read 9 tweets
Apr 12
🧵1/ Just attended a 🔥 @LangChainAI webinar on AI agents, ft. some of the brightest minds in the space! Let's unpack the key takeaways & explore the cutting-edge work being done.

Guests:
@charles_irl
@ShunyuYao12
@mbusigin
@yoheinakajima
@hwchase17
@LangChainAI @charles_irl @ShunyuYao12 @mbusigin @yoheinakajima @hwchase17 🧠2/ Shunyu introduced the core idea of his #ReAct paper, which adds a "thought" between Action & Observation. The open question is, do we need a strict pattern of Thought-Action-Observation or should we just add thoughts as a special type of action, offering more flexibility?
@LangChainAI @charles_irl @ShunyuYao12 @mbusigin @yoheinakajima @hwchase17 🤖3/ Yohei built babyAGI, inspired by #HustleGPT. He used GPT-4 to create an autonomous Founder Bot, even penning a "scientific paper" about it. BabyAGI focuses on completing, generating, and prioritizing tasks. #BabyAGI
Read 16 tweets
Apr 12
🧵 1/ 🤖🧠 Stanford AI researchers have introduced a groundbreaking concept: Generative Agents, computer programs that simulate authentic human behavior using generative models. These agents display memory, introspection, and planning capabilities. Let's dive in. 👇
2/ 🎮 In the study, 25 Generative Agents were placed in a virtual sandbox-like world (think The Sims). They had unique backgrounds and interacted in a 2-day simulation. Examples of emergent behavior: one agent threw a party, another ran for mayor!
3/ 📊 Here's the kicker: actual humans role-playing the same 25 agents generated responses that were rated as less human-like than the chatbot-powered agents by an evaluation panel. Generative Agents are getting closer and closer to authentic human behavior. 😮
Read 10 tweets
Apr 12
🧵 1/ 📈 Yesterday we talked about how important chunking is when using vector databases like @pinecone, @weaviate_io or @trychroma. But what exactly are vector databases in the first place? Let's explore this game-changer! 👇
@pinecone @weaviate_io @trychroma 2/ 🤖 Machine Learning (ML) techniques can transform complex data into vector embeddings, describing data objects as numeric values in multiple dimensions. Vector databases index these embeddings for easy search & retrieval, finding similar values. 🧠
@pinecone @weaviate_io @trychroma 3/ 🔍 Vector databases excel at similarity search (vector search), allowing users to find related results without knowing specific keywords or metadata classifications. This provides accurate results while eliminating irrelevant ones that traditional search tech might return.
Read 8 tweets
Apr 12
🧵 1/ Vector stores & embeddings may be the talk of the town, but there's more to explore! Discover how to fine-tune LLaMA, an open-source language model, to make it sound like Homer Simpson! You can even apply this method to other characters! 🍩📺 Shout out to @bfirsh for this! Image
@bfirsh 2/ The process starts by using a dataset containing scripts from The Simpsons TV show (Seasons 1-12) obtained from Kaggle. With ~60k lines of dialog and 1.1M tokens, it's time to train LLaMA to reproduce the voice of the characters. 📚
@bfirsh 3/ To make LLaMA speak like a character, the dataset is parsed into scenes and prompts are generated. This is done by taking the previous lines in the scene, the character with the next line, and that line. The model is then prompted to complete the line in context. 🎭
Read 7 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us on Twitter!

:(