Machine Learning

AcademicCodeConceptMachine LearningpaperSeries

Deep Dive: FineTune small GPT for SPAM, ScrapeGraphAI, Parallelizable LSTMs

Sebastian Raschka guides users in fine-tuning a small GPT model to classify SPAM messages with 96% accuracy. ScrapeGraphAI is a Python library that automates data extraction from websites using LLMs. And Sepp Hochreiter’s xLSTM architecture extends traditional LSTMs to compete with state-of-the-art Transformers. These innovations are making AI more accessible and efficient! 🚀🤖📚

Read More
CodeMachine LearningProjectVideosYouTube

Why Your RL Model Fails: Prioritized Replay and Actor-Critic in code (Part 2)

In this video, I break down the code behind designing an RL agent with an Actor-Critic architecture using a prioritized replay buffer! 🤖💻 Discover how to tackle sparse rewards, optimize training efficiency, and boost your model’s performance with practical tips and WandB tracking. If you want to go beyond theory and see how to implement these concepts in code, this is the video for you! Check it out and level up your RL skills today!

Read More
AcademicMachine LearningTechnologyVideosYouTube

Overcome Sparse Rewards in Reinforcement Learning

https://www.youtube.com/watch?v=iodNcqh3B-M
Struggling with sparse rewards in your RL projects? Discover how we used the ACER algorithm, reward shaping, and prioritized experience replay to boost our agent’s performance. 🚀 Watch my latest video to learn how to prevent policy collapse and achieve stable learning in complex environments. #ReinforcementLearning #AI #SparseRewards”

Read More
AcademicCodeConceptMachine LearningProjectVideos

Score Prediction from User Log with BERT

Discover latest project: predicting user scores from interaction logs with a quiz system! By analyzing user behavior and leveraging cutting-edge ML techniques like DistilBERT, curriculum learning, and LoRA, we’ve crafted a robust model to infer scores without needing correct answers or user choices. Dive into our innovative approach and see how we’re revolutionizing educational tools. #MachineLearning #AI #EdTech #Robotics #DistilBERT #CurriculumLearning #LoRA

Read More
AcademicCodeConceptMachine LearningSeries

Technow: torchtune, Boston Dynamics Atlas robot, Reka Core

Easily fine-tune large language models (LLMs) with torchtune, PyTorch’s new library. Torchtune offers modular and customizable fine-tuning recipes, supporting efficient workflows from data preparation to model evaluation. It integrates with popular tools like Hugging Face Hub and Weights & Biases and excels in memory efficiency, making it ideal for users with limited hardware. Designed for simplicity and extensibility, torchtune supports models like Llama 2 and Mistral, with future plans for models up to 70 billion parameters.

Read More
AcademicConceptMachine Learning

Deepdive: Many-Shot learning, AutoCrawler, Megalodon

Large language models (LLMs) are revolutionizing various fields, but their capabilities can be limited. This is where exciting new research comes in! By tackling challenges like limited learning from examples and information processing constraints, these advancements are making LLMs more adaptable, efficient, and powerful. Let’s dive into three key breakthroughs: Many-Shot In-Context Learning, AutoCrawler, and Megalodon.

Read More
AcademicCodeConceptMachine Learning

Deepdive: Custom Collate in DataLoader, Gemini cookbook, Build LLM from scratch

Optimize your large language model (LLM) training with PyTorch by using a custom collate function in DataLoader. This technique dynamically pads text sequences in each batch to match the longest sequence, reducing wasted computation from unnecessary padding. Learn how to implement a custom collate function using pad_sequence from torch.nn.utils.rnn and integrate it with DataLoader for efficient handling of variable-length sequences in your LLM training workflow.

Read More