Skip to content

A repository of technical articles on AI algorithms, model finetuning, AI agents, open-source libraries, and system design. Perfect for researchers and enthusiasts looking for in-depth insights.

License

Notifications You must be signed in to change notification settings

ksmooi/ai_power

Repository files navigation

[ Español ] [ 简体中文 ] [ 繁體中文 ] [ 日本語 ] [ 한국어 ]

AI Power

AI technology is advancing at lightning speed, with new algorithms and AI libraries emerging and evolving constantly. To empower more people to master the latest AI innovations and actively participate in open-source projects, I created AI Power. Join us in exploring the cutting-edge of AI technology and contribute to shaping the future!

Article Categories

Code-Driven Algorithm Learning

Title Description Keywords
Understanding Transformer Attentions An in-depth explanation of the attention mechanism in transformers, covering self-attention, multi-head attention, and their implementation in modern NLP models. Transformers, Self-Attention, MHA
The Vanilla Transformer Explained A comprehensive guide to the vanilla transformer model, detailing its architecture, components, and the forward pass process for sequence-to-sequence tasks. vanilla Transformer, Architecture, Sequence-to-Sequence
Inside CLIP An in-depth explanation of the CLIP model, covering its architecture, training process, and applications in linking images and text. CLIP, Architecture
Deep Dive into LLaVA A comprehensive guide to the implementation of the LLaVA model, exploring its architecture, components, and how it enhances language understanding tasks. LLaVA, Architecture, MultiModal
Deep Dive into Vision Transformer An in-depth explanation of the Vision Transformer (ViT) model, detailing its architecture, key components, and application in computer vision tasks. Vision Transformer, ViT, Architecture
AutoEncoder Explained An in-depth explanation of autoencoders, their architecture, types, and applications in data compression and feature learning. AutoEncoder, VAE, Architecture

Huggingface API

Title Description Keywords
Huggingface SBERT API - Part 1 This article introduces the Huggingface Sentence-BERT (SBERT) API, explaining its purpose, applications, and how to use it for embedding sentences and computing similarities. SBERT, Sentence Transformer, Embeddings
Huggingface SBERT API - Part 2 This continuation of the SBERT API guide covers advanced usage, fine-tuning models, and integrating SBERT with various applications. SBERT, Sentence Transformer, Embeddings
Huggingface Transformer Auto Class API - Part 1 This article provides an overview of the Huggingface Transformer Auto Class API, detailing its features, setup, and basic usage for different NLP tasks. Transformer, Auto Class, NLP, API
Huggingface Transformer Auto Class API - Part 2 A deeper dive into the Transformer Auto Class API, exploring custom configurations, model optimization, and use cases. Transformer, Auto Class, NLP, API
Huggingface Transformer Pipeline API This guide explains the Huggingface Transformer Pipeline API, showcasing its ease of use for various NLP tasks like text classification, named entity recognition, and text generation. Transformer Pipeline
Huggingface CLIP API This article introduces the Huggingface CLIP (Contrastive Language-Image Pre-training) API, explaining its purpose, applications, and how to use it for image and text embeddings. Huggingface, CLIP, API
Huggingface LLaVA Next API This guide details the Huggingface LLaVA Next API, outlining its features, setup, and usage for advanced language understanding tasks. LLaVA, MultiModal, API
Huggingface Vision Transformer (ViT) API This article provides an overview of the Huggingface Vision Transformer (ViT) API, explaining its usage for image classification and other vision tasks. ViT, Vision Transformer, Image Classification, API
Huggingface Diffusers API This article provides an overview of the Huggingface Diffusers API, explaining its functionality and usage for generating images from text descriptions. Diffusers API, Image Generation, Text-to-Image, Image-to-Image
Huggingface Diffusers Chained Pipeline This guide explains how to create chained pipelines using the Huggingface Diffusers API, showcasing how to combine multiple models for complex tasks. Diffusers API, Chained Pipeline
Huggingface Diffusers Pipeline API An in-depth look at the Huggingface Diffusers Pipeline API, detailing its features, setup, and applications for image and text generation. Diffusers Pipeline API

Tensor Operation

Title Description Keywords
Einops Einsum An introduction to the einsum function in the einops library, explaining its syntax, usage, and applications in tensor operations. einops, einsum, tensor operations
Einops Rearrange This guide details the rearrange function in the einops library, showcasing how to efficiently manipulate and transform tensor shapes. einops, einsum, tensor operations

Dataset

Title Description Keywords
Huggingface Datasets Loading This article covers how to load and preprocess datasets using the Huggingface Datasets library, including handling various data formats and sources. Huggingface, Datasets
Huggingface Datasets Main Classes A comprehensive guide to the main classes of the Huggingface Datasets library, explaining their functionalities and use cases. Huggingface, Datasets, Main Classes
Alpaca Self-Instruct Guide This guide provides a comprehensive overview of the Self-Instruct process using Alpaca, including step-by-step instructions and examples. Alpaca, Self-Instruct
Generating Datasets with Unstructured.io and GPT4 This article demonstrates how to use Unstructured.io and GPT-4 to process PDF files and generate datasets by extracting and organizing content. Unstructured.io, GPT-4
Generating Datasets with Table Transformer and GPT4 This article explains how to use Table Transformer and GPT-4 to generate datasets from PDF files by detecting and extracting table structures. Table Transformer, GPT-4

Model Training

Title Description Keywords
SFT: Model Architecture Tweaks This article discusses various tweaks and adjustments to model architecture for optimizing performance in supervised fine-tuning. Model Architecture, SFT
SFT: Training Strategy This article provides insights into effective training strategies for supervised fine-tuning, including tips and best practices. Training Strategy, SFT
SFT: Data Handling This article explains the techniques for handling and preparing data for supervised fine-tuning tasks. Data Handling, SFT
SFT: Loss Function This article explores different loss functions used in supervised fine-tuning and their impact on model performance. Loss Function, SFT
Huggingface Transformer Trainer API This guide explores how to fine-tune transformer models using the Huggingface Trainer API, covering setup, training, and evaluation processes. Huggingface, Transformer, Trainer API, SFT
Huggingface Evaluate API This article introduces the Huggingface Evaluate API, detailing its purpose, setup, and usage for evaluating machine learning models. Huggingface, Evaluate API, Metric
RLHF with PPO Overview An overview of Reinforcement Learning with Human Feedback (RLHF) using Proximal Policy Optimization (PPO) to train language models. RLHF, PPO, Language Models
Understanding DPO and ORPO This article explores Direct Preference Optimization (DPO) and Odds Ratio Preference Optimization (ORPO), detailing their methodologies, loss functions, and practical applications in fine-tuning language models to align with human preferences. RLHF, DPO, ORPO

Model Serving and Inference

Title Description Keywords
ONNX Model Optimization Techniques This article explores various techniques to optimize ONNX models for faster inference, including quantization, pruning, and hardware-specific optimizations. ONNX, Optimization, Quantization, Pruning
Understanding and Using trtexec with TensorRT This article provides an in-depth guide on how to use the trtexec tool in TensorRT, covering its various options and providing practical examples. TensorRT, trtexec
Introduction to TensorRT Capabilities This article explains the key capabilities of TensorRT, NVIDIA's high-performance deep learning inference library, with detailed C++ examples. TensorRT, Capabilities
Understanding the Inner Workings of TensorRT This article provides an in-depth explanation of how TensorRT operates, including details on object lifetimes, error handling, memory management, threading, and determinism, illustrated with C++ examples. TensorRT, Object Lifetimes, Error Handling
NvMultiObjectTracker Part 1: Introduction and Core Concepts This article introduces the key concepts, architecture, and workflow of NvMultiObjectTracker, a library for multi-object tracking in NVIDIA's DeepStream SDK. NvMultiObjectTracker, DeepStream, Multi-Object Tracking
Advanced Features and Applications of NvMultiObjectTracker This article explores advanced features like Re-Identification, Target Re-Association, Bounding-Box Unclipping, and Single-View 3D Tracking in NvMultiObjectTracker. NvMultiObjectTracker, Re-ID, 3D Tracking
Advanced Configuration of NvMultiObjectTracker This article discusses the advanced configuration parameters of NvMultiObjectTracker, focusing on how to optimize the tracker for specific use cases. DeepStream, NvMultiObjectTracker, Configuration

Our Goals

The goals of AI Power Source are as follows:

  • Effectively Understand AI Algorithms: Deepen understanding of various AI algorithms by reading code.
  • Quickly Learn AI Libraries: Accelerate mastery of various AI libraries using numerous example programs.
  • Analyze Code: Promote learning by analyzing the code of various AI frameworks and applications.
  • Learn Model Training Techniques: Quickly master the skills required for AI model training through abundant example programs and experience sharing.
  • MLOps Process Design: Learn and practice MLOps process design to improve the efficiency and reliability of model deployment and management.
  • System Architecture Design: Learn the system architecture design of AI applications through case studies, including software and cloud architecture design.

How to Contribute

If you wish to assist with the following tasks, we welcome you to join us:

  • Assist in Translating Existing Articles: Help translate and improve existing articles and educational materials.
  • Contribute New Articles: Contribute at least one new article each month, sharing your AI knowledge and experience.

Join Us

We welcome all individuals interested in AI technology to join our community, regardless of your experience level. Every contribution you make will play a significant role in promoting the popularization and development of AI technology.

Contact Us

If you have any questions or suggestions, please contact us via GitHub Issues.

Thank you for your participation and support!

About

A repository of technical articles on AI algorithms, model finetuning, AI agents, open-source libraries, and system design. Perfect for researchers and enthusiasts looking for in-depth insights.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published