#python #computer_vision #convolutional_networks #embedding_vectors #embeddings #feature_extraction #feature_vector #image_processing #image_retrieval #machine_learning #milvus #pipeline #towhee #transformer #unstructured_data #video_processing #vision_transformer #vit
https://github.com/towhee-io/towhee
https://github.com/towhee-io/towhee
GitHub
GitHub - towhee-io/towhee: Towhee is a framework that is dedicated to making neural data processing pipelines simple and fast.
Towhee is a framework that is dedicated to making neural data processing pipelines simple and fast. - towhee-io/towhee
#jupyter_notebook #ade20k #cityscapes #coco #image_segmentation #instance_segmentation #oneformer #panoptic_segmentation #semantic_segmentation #transformer #universal_segmentation
https://github.com/SHI-Labs/OneFormer
https://github.com/SHI-Labs/OneFormer
GitHub
GitHub - SHI-Labs/OneFormer: [CVPR 2023] OneFormer: One Transformer to Rule Universal Image Segmentation
[CVPR 2023] OneFormer: One Transformer to Rule Universal Image Segmentation - SHI-Labs/OneFormer
#python #computer_vision #machine_learning #multimodal #natural_language_processing #pretrained_language_model #speech_processing #transformer #translation
https://github.com/microsoft/torchscale
https://github.com/microsoft/torchscale
GitHub
GitHub - microsoft/torchscale: Foundation Architecture for (M)LLMs
Foundation Architecture for (M)LLMs. Contribute to microsoft/torchscale development by creating an account on GitHub.
#python #bert #deep_learning #language_model #language_models #machine_learning #natural_language_processing #nlp #pytorch #transformer
https://github.com/extreme-bert/extreme-bert
https://github.com/extreme-bert/extreme-bert
GitHub
GitHub - extreme-bert/extreme-bert: ExtremeBERT is a toolkit that accelerates the pretraining of customized language models on…
ExtremeBERT is a toolkit that accelerates the pretraining of customized language models on customized datasets, described in the paper “ExtremeBERT: A Toolkit for Accelerating Pretraining of Custom...
#python #attention_mechanism #deep_learning #gpt #gpt_2 #gpt_3 #language_model #linear_attention #lstm #pytorch #rnn #rwkv #transformer #transformers
https://github.com/BlinkDL/RWKV-LM
https://github.com/BlinkDL/RWKV-LM
GitHub
GitHub - BlinkDL/RWKV-LM: RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like…
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it'...
#python #deep_learning #flax #jax #language_model #large_language_models #natural_language_processing #transformer
https://github.com/young-geng/EasyLM
https://github.com/young-geng/EasyLM
GitHub
GitHub - young-geng/EasyLM: Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning…
Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Flax. - young-geng/EasyLM
#python #bloom #deep_learning #gpt #inference #nlp #pytorch #transformer
https://github.com/huggingface/text-generation-inference
https://github.com/huggingface/text-generation-inference
GitHub
GitHub - huggingface/text-generation-inference: Large Language Model Text Generation Inference
Large Language Model Text Generation Inference. Contribute to huggingface/text-generation-inference development by creating an account on GitHub.
#python #deeplabv3 #image_segmentation #medical_image_segmentation #pspnet #pytorch #realtime_segmentation #retinal_vessel_segmentation #semantic_segmentation #swin_transformer #transformer #vessel_segmentation
MMSegmentation is an open-source semantic segmentation toolbox based on PyTorch, part of the OpenMMLab project. It offers a unified benchmark for various semantic segmentation methods, a modular design allowing easy customization, and support for multiple popular segmentation frameworks like PSPNet and DeepLabV3+. The toolbox is highly efficient and provides detailed tutorials, advanced guides, and support for numerous datasets and backbones. This makes it a powerful tool for researchers and developers to implement and develop new semantic segmentation methods efficiently. By using MMSegmentation, you can leverage its flexibility, extensive features, and community support to enhance your projects in image segmentation tasks.
https://github.com/open-mmlab/mmsegmentation
MMSegmentation is an open-source semantic segmentation toolbox based on PyTorch, part of the OpenMMLab project. It offers a unified benchmark for various semantic segmentation methods, a modular design allowing easy customization, and support for multiple popular segmentation frameworks like PSPNet and DeepLabV3+. The toolbox is highly efficient and provides detailed tutorials, advanced guides, and support for numerous datasets and backbones. This makes it a powerful tool for researchers and developers to implement and develop new semantic segmentation methods efficiently. By using MMSegmentation, you can leverage its flexibility, extensive features, and community support to enhance your projects in image segmentation tasks.
https://github.com/open-mmlab/mmsegmentation
GitHub
GitHub - open-mmlab/mmsegmentation: OpenMMLab Semantic Segmentation Toolbox and Benchmark.
OpenMMLab Semantic Segmentation Toolbox and Benchmark. - open-mmlab/mmsegmentation
#python #bert #deep_learning #flax #hacktoberfest #jax #language_model #language_models #machine_learning #model_hub #natural_language_processing #nlp #nlp_library #pretrained_models #python #pytorch #pytorch_transformers #seq2seq #speech_recognition #tensorflow #transformer
The Hugging Face Transformers library provides thousands of pretrained models for various tasks like text, image, and audio processing. These models can be used for tasks such as text classification, image detection, speech recognition, and more. The library supports popular deep learning frameworks like JAX, PyTorch, and TensorFlow, making it easy to switch between them.
The benefit to the user is that you can quickly download and use these pretrained models with just a few lines of code, saving time and computational resources. You can also fine-tune these models on your own datasets and share them with the community. Additionally, the library offers a simple `pipeline` API for immediate use on different inputs, making it user-friendly for both researchers and practitioners. This helps in reducing compute costs and carbon footprint while enabling high-performance results across various machine learning tasks.
https://github.com/huggingface/transformers
The Hugging Face Transformers library provides thousands of pretrained models for various tasks like text, image, and audio processing. These models can be used for tasks such as text classification, image detection, speech recognition, and more. The library supports popular deep learning frameworks like JAX, PyTorch, and TensorFlow, making it easy to switch between them.
The benefit to the user is that you can quickly download and use these pretrained models with just a few lines of code, saving time and computational resources. You can also fine-tune these models on your own datasets and share them with the community. Additionally, the library offers a simple `pipeline` API for immediate use on different inputs, making it user-friendly for both researchers and practitioners. This helps in reducing compute costs and carbon footprint while enabling high-performance results across various machine learning tasks.
https://github.com/huggingface/transformers
GitHub
GitHub - huggingface/transformers: 🤗 Transformers: the model-definition framework for state-of-the-art machine learning models…
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training. - GitHub - huggingface/t...
#python #amd #cuda #gpt #inference #inferentia #llama #llm #llm_serving #llmops #mlops #model_serving #pytorch #rocm #tpu #trainium #transformer #xpu
vLLM is a library that makes it easy, fast, and cheap to use large language models (LLMs). It is designed to be fast with features like efficient memory management, continuous batching, and optimized CUDA kernels. vLLM supports many popular models and can run on various hardware including NVIDIA GPUs, AMD CPUs and GPUs, and more. It also offers seamless integration with Hugging Face models and supports different decoding algorithms. This makes it flexible and easy to use for anyone needing to serve LLMs, whether for research or other applications. You can install vLLM easily with `pip install vllm` and find detailed documentation on their website.
https://github.com/vllm-project/vllm
vLLM is a library that makes it easy, fast, and cheap to use large language models (LLMs). It is designed to be fast with features like efficient memory management, continuous batching, and optimized CUDA kernels. vLLM supports many popular models and can run on various hardware including NVIDIA GPUs, AMD CPUs and GPUs, and more. It also offers seamless integration with Hugging Face models and supports different decoding algorithms. This makes it flexible and easy to use for anyone needing to serve LLMs, whether for research or other applications. You can install vLLM easily with `pip install vllm` and find detailed documentation on their website.
https://github.com/vllm-project/vllm
GitHub
GitHub - vllm-project/vllm: A high-throughput and memory-efficient inference and serving engine for LLMs
A high-throughput and memory-efficient inference and serving engine for LLMs - vllm-project/vllm
❤1