BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
CLIP (Contrastive Language-Image Pre-Training) is a neural network trained on a variety of (image, text) pairs.
Welcome to an open source implementation of OpenAI's CLIP (Contrastive Language-Image Pre-training).
Towards Robust Blind Face Restoration with Codebook Lookup Transformer (NeurIPS 2022)
A demo repository for UniMRCP plugin implementation with iflytek ASR & TTS API
A WebRTC, SIP and VoIP library for C# and .NET. Designed for real-time communications apps.