If there's anything you'd like to see, let me know!
Let's dive in:
High level overview of what's happening:
1️⃣ Load your audio file
2️⃣ Speech-to-text with Whisper
3️⃣ Split the transcript into chunks
4️⃣ Summarize the transcript
Let's dive into the code ⬇️
1. Open your audio file.
Whisper works on files with a maximum duration of approximately 20 minutes.
Longer file?
Split it up by using PyTube and handle each chunk seperately.
2. Call the Whisper API
Of course, you can use any speech-to-text API you prefer. I like Whisper for its accuracy and ease of use.
Any open-source alternatives with quality output?
3. Splitting the transcript into chunks
With the new OpenAI GPT 16k model, you can fit a large amount of context into one chunk. This is amazing for the model to 'understand' the full context and make connections.
Use some overlap in order for context to not be lost.
4. Prompting
Prompting is key. It determines your output more than anything else.
Be as concise as you can be. Instruct the model on how you want the output to look. You could include bullet points, main takeaways, follow-up actions and much more.
5. Initialize and run the summary chain
I'm using the refine summarization chain. This is great when you're working with large files.
It generates an initial summary based on the first chunk and updates it with the subsequent chunks.
6. Export the summary to a text file
Your meeting is summarized and you're ready to take action!
Once again, try to play around with the prompts you're using. This will greatly impact the resulting summarization.
7. Possible future implementations
You can go wild with this. You could use a Zapier integration to send the summarized meeting in an email, create appointments in your schedule and much more.
But did you know you can create AI-tools like ChatGPT yourself?
Unleash the potential by merging @LangChainAI and @streamlit.
Today's mission: craft a 4-hour workday blueprint based on @thedankoe's video.
Let me show you how 🧵
Before we dive in, this is day 1 of my LangChain Unchained series.
Each day, I'll implement a small LangChain project in Streamlit.
Follow @JorisTechTalk to stay up-to-date.
If there's anything you'd like to see, let me know!
Let's dive in:
A high-level overview:
1️⃣ Load YouTube transcript
2️⃣ Split transcript into chunks
3️⃣ Use summarization chain to create a strategy based on the content of the video.
4️⃣ Use a simple LLM Chain to create a detailed plan based on the strategy.
ChatGPT can give you a kick-start when learning new skills.
But I like to learn through YouTube videos.
With the power of @LangChainAI, you can generate a personalized YouTube study schedule based on a skill you'd like to learn.
Let me show you how: 🧵
#AI
Before we dive in, this is day 7 of my '7 days of LangChain'.
Every day, I've introduced you to a simple project that will guide you through the basics of LangChain.
Today's a longer one.
Follow @JorisTechTalk to stay up-to-date on my next series.
Let's dive in:
High level overview of what's happening:
1️⃣ Generate list of video id's from favorite YT channels
2️⃣ Load all transcripts
3️⃣ Split the transcript
4️⃣ Extract skills
5️⃣ Vectorize
6️⃣ Generate skillset
7️⃣ Find relevant videos.