基于 Tensorflow,仿 Scikit-Learn 设计的深度学习自然语言处理框架。支持 40 余种模型类,涵盖语言模型、文本分类、NER、MRC、知识蒸馏等各个领域
-
Updated
May 10, 2023 - Python
基于 Tensorflow,仿 Scikit-Learn 设计的深度学习自然语言处理框架。支持 40 余种模型类,涵盖语言模型、文本分类、NER、MRC、知识蒸馏等各个领域
Fine-Tuning Google's Vision Transformer LoRA technique. Two different LoRA adapters are tuned for separate classification (food and human actions). A simple Gradio interface is implemented to run the inference.
Final project for the UPC Postgrau in Artificial Intelligence & Deep Learning, showcasing a SELF-DRIVING demo built around lane detection and object detection models. Integrates LaneNet (with ENet backbone and BDD100K transforms), a custom Mask R-CNN segmentation pipeline, and Faster R-CNN with ResNet50-FPN.
This repository features an image sharpening pipeline using Knowledge Distillation. A high-capacity Restormer acts as the teacher model, while a lightweight Mini-UNet is trained as the student to mimic its performance.
AI-Farm is a distributed deep learning training framework that enables efficient model training across multiple machines. It provides a scalable infrastructure with real-time monitoring through a web admin panel, adaptive task distribution, and support for both CPU and GPU training.
A full-stack platform for designing reinforcement learning environments, running GPU-backed training, and monitoring agents in real-time.
Production-ready PyTorch framework for distributed deep learning training with Ray & Horovod backends. Optimized for computer vision and time series on Kubernetes clusters.
This repo documents my participation in the Kaggle red-teaming competition focused on probing OpenAI's newly released gpt-oss-20b model for previously undiscovered vulnerabilities and harmful behaviors. The goal is to identify, document, and report up to five distinct issues, contributing to the safety and alignment of open-source AI models.
PyTorch implementation of DDPM (Denoising Diffusion Probabilistic Models) for image generation. Includes U-Net with attention, DDIM sampling, EMA training, and CelebA dataset support.
Data processing, Machine Learning codes and Training scripts on Jean Zay High Performance Computing (HPC) remote GPUs
Autonomous AI research agents for single-GPU nanochat training — automated experiment design, execution, and analysis
Developed an end-to-end LLM pipeline that extracts Python code from GitHub, builds a high-quality dataset, fine-tunes CodeGen, and performs advanced code generation with DeepSeek. Demonstrates strong capabilities in LLM training, data engineering, and model optimization.
Public OpenClaw and Codex skills for reproducible CV experimentation and SOTA campaign management across Colab, Kaggle, browser automation, and GPU VMs.
Fine-Tuning Google's Vision Transformer LoRA technique. Two different LoRA adapters are tuned for Separate Classification (Food and Human actions). A simple Gradio interface is implemented to run the Inference.
Medical X-ray image classifier fine-tuning ResNet-18 with PyTorch.
EGM is a deep learning tool that learns your image editing style from raw/edited pairs and applies it to new images. It uses a Pix2Pix (conditional GAN) architecture.
1D CNN for customer complaint text classification using PyTorch.
Kaggle Competition for MAP Charting Student Math Mistunderstanding
Whatsapp Assistant using FastAPI, Google Gemini, WhisperAI, Twilio and Ngrok
Add a description, image, and links to the gpu-training topic page so that developers can more easily learn about it.
To associate your repository with the gpu-training topic, visit your repo's landing page and select "manage topics."