Hangbo Bao's Projects
A jekyll based resume template
Algorithm Course Lab
Software Dummy Display Adapter for Apple Silicon Macs to Have Custom HiDPI Resolutions.
最全中华古诗词数据库, 唐宋两朝近一万四千古诗人, 接近5.5万首唐诗加26万宋诗. 两宋时期1564位词人,21050首词。
Code to obtain the CNN / Daily Mail dataset (non-anonymized) for summarization
The implementation of DeBERTa
initial
instruction for the installation of the Nvidia driver + cuda + cudnn
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
An efficient implementation of the popular sequence models for text generation, summarization, and translation tasks.
Course from https://www.coursera.org/learn/machine-learning/home/welcome
Ongoing research training transformer language models at scale, including: BERT
MPNet: Masked and Permuted Pre-training for Language Understanding https://arxiv.org/pdf/2004.09297.pdf
SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
ProphetNet: Predicting Future N-gram for Sequence-to-Sequence Pre-training https://arxiv.org/pdf/2001.04063.pdf
Python爬虫代理IP池(proxy pool)
PyTorch image models, scripts, pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNet-V3/V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more
A simple crawler base C#
A latent text-to-image diffusion model
Tensorflow study code
Transformer with Untied Positional Encoding (TUPE). Code of paper "Rethinking the Positional Encoding in Language Pre-training".
UniLM AI - Unified "Language" Model Pre-training across Tasks, Languages, and Modalities
Official code Cross-Covariance Image Transformer (XCiT)