#python #cv #deep_learning #machine_learning #multi_modal #nlp #science #speech
https://github.com/modelscope/modelscope
https://github.com/modelscope/modelscope
GitHub
GitHub - modelscope/modelscope: ModelScope: bring the notion of Model-as-a-Service to life.
ModelScope: bring the notion of Model-as-a-Service to life. - modelscope/modelscope
#python #bloom #deep_learning #gpt #inference #nlp #pytorch #transformer
https://github.com/huggingface/text-generation-inference
https://github.com/huggingface/text-generation-inference
GitHub
GitHub - huggingface/text-generation-inference: Large Language Model Text Generation Inference
Large Language Model Text Generation Inference. Contribute to huggingface/text-generation-inference development by creating an account on GitHub.
#python #document_ai #document_image_analysis #document_layout_analysis #document_parser #document_understanding #layoutlm #nlp #ocr #publaynet #pubtabnet #pytorch #table_detection #table_recognition #tensorflow
https://github.com/deepdoctection/deepdoctection
https://github.com/deepdoctection/deepdoctection
GitHub
GitHub - deepdoctection/deepdoctection: A Repo For Document AI
A Repo For Document AI. Contribute to deepdoctection/deepdoctection development by creating an account on GitHub.
#python #active_learning #ai #annotation_tool #developer_tools #gpt_4 #human_in_the_loop #langchain #llm #machine_learning #mlops #natural_language_processing #nlp #rlhf #text_annotation #text_labeling #weak_supervision #weakly_supervised_learning
https://github.com/argilla-io/argilla
https://github.com/argilla-io/argilla
GitHub
GitHub - argilla-io/argilla: Argilla is a collaboration tool for AI engineers and domain experts to build high-quality datasets
Argilla is a collaboration tool for AI engineers and domain experts to build high-quality datasets - argilla-io/argilla
❤1
#python #embeddings #information_retrieval #language_model #large_language_models #llm #machine_learning #nearest_neighbor_search #neural_search #nlp #search #search_engine #semantic_search #sentence_embeddings #similarity_search #transformers #txtai #vector_database #vector_search #vector_search_engine
https://github.com/neuml/txtai
https://github.com/neuml/txtai
GitHub
GitHub - neuml/txtai: 💡 All-in-one AI framework for semantic search, LLM orchestration and language model workflows
💡 All-in-one AI framework for semantic search, LLM orchestration and language model workflows - neuml/txtai
#jupyter_notebook #computer_vision #gpt #huggingface_transformers #llm #machinelearning #nlp_machine_learning #rag
https://github.com/katanaml/sparrow
https://github.com/katanaml/sparrow
GitHub
GitHub - katanaml/sparrow: Structured data extraction and instruction calling with ML, LLM and Vision LLM
Structured data extraction and instruction calling with ML, LLM and Vision LLM - katanaml/sparrow
#python #beit #beit_3 #bitnet #deepnet #document_ai #foundation_models #kosmos #kosmos_1 #layoutlm #layoutxlm #llm #minilm #mllm #multimodal #nlp #pre_trained_model #textdiffuser #trocr #unilm #xlm_e
Microsoft is developing advanced AI models through large-scale self-supervised pre-training across various tasks, languages, and modalities. These models, such as Foundation Transformers (Magneto) and Kosmos-2.5, are designed to be highly generalizable and capable of handling multiple tasks like language understanding, vision, speech, and multimodal interactions. The benefit to users includes state-of-the-art performance in document AI, speech recognition, machine translation, and more, making these models highly versatile and efficient for a wide range of applications. Additionally, tools like TorchScale and Aggressive Decoding enhance stability, efficiency, and speed in model training and deployment.
https://github.com/microsoft/unilm
Microsoft is developing advanced AI models through large-scale self-supervised pre-training across various tasks, languages, and modalities. These models, such as Foundation Transformers (Magneto) and Kosmos-2.5, are designed to be highly generalizable and capable of handling multiple tasks like language understanding, vision, speech, and multimodal interactions. The benefit to users includes state-of-the-art performance in document AI, speech recognition, machine translation, and more, making these models highly versatile and efficient for a wide range of applications. Additionally, tools like TorchScale and Aggressive Decoding enhance stability, efficiency, and speed in model training and deployment.
https://github.com/microsoft/unilm
GitHub
GitHub - microsoft/unilm: Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities - microsoft/unilm
#python #chinese #clip #computer_vision #contrastive_loss #coreml_models #deep_learning #image_text_retrieval #multi_modal #multi_modal_learning #nlp #pretrained_models #pytorch #transformers #vision_and_language_pre_training #vision_language
This project is about a Chinese version of the CLIP (Contrastive Language-Image Pretraining) model, trained on a large dataset of Chinese text and images. Here’s what you need to know This model helps you quickly perform tasks like calculating text and image features, cross-modal retrieval (finding images based on text or vice versa), and zero-shot image classification (classifying images without any labeled examples).
- **Ease of Use** The model has been tested on various datasets and shows strong performance in zero-shot image classification and cross-modal retrieval tasks.
- **Resources**: The project includes pre-trained models, training and testing codes, and detailed tutorials on how to use the model for different tasks.
Overall, this project makes it easy to work with Chinese text and images using advanced AI techniques, saving you time and effort.
https://github.com/OFA-Sys/Chinese-CLIP
This project is about a Chinese version of the CLIP (Contrastive Language-Image Pretraining) model, trained on a large dataset of Chinese text and images. Here’s what you need to know This model helps you quickly perform tasks like calculating text and image features, cross-modal retrieval (finding images based on text or vice versa), and zero-shot image classification (classifying images without any labeled examples).
- **Ease of Use** The model has been tested on various datasets and shows strong performance in zero-shot image classification and cross-modal retrieval tasks.
- **Resources**: The project includes pre-trained models, training and testing codes, and detailed tutorials on how to use the model for different tasks.
Overall, this project makes it easy to work with Chinese text and images using advanced AI techniques, saving you time and effort.
https://github.com/OFA-Sys/Chinese-CLIP
GitHub
GitHub - OFA-Sys/Chinese-CLIP: Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation. - OFA-Sys/Chinese-CLIP
#python #bert #deep_learning #flax #hacktoberfest #jax #language_model #language_models #machine_learning #model_hub #natural_language_processing #nlp #nlp_library #pretrained_models #python #pytorch #pytorch_transformers #seq2seq #speech_recognition #tensorflow #transformer
The Hugging Face Transformers library provides thousands of pretrained models for various tasks like text, image, and audio processing. These models can be used for tasks such as text classification, image detection, speech recognition, and more. The library supports popular deep learning frameworks like JAX, PyTorch, and TensorFlow, making it easy to switch between them.
The benefit to the user is that you can quickly download and use these pretrained models with just a few lines of code, saving time and computational resources. You can also fine-tune these models on your own datasets and share them with the community. Additionally, the library offers a simple `pipeline` API for immediate use on different inputs, making it user-friendly for both researchers and practitioners. This helps in reducing compute costs and carbon footprint while enabling high-performance results across various machine learning tasks.
https://github.com/huggingface/transformers
The Hugging Face Transformers library provides thousands of pretrained models for various tasks like text, image, and audio processing. These models can be used for tasks such as text classification, image detection, speech recognition, and more. The library supports popular deep learning frameworks like JAX, PyTorch, and TensorFlow, making it easy to switch between them.
The benefit to the user is that you can quickly download and use these pretrained models with just a few lines of code, saving time and computational resources. You can also fine-tune these models on your own datasets and share them with the community. Additionally, the library offers a simple `pipeline` API for immediate use on different inputs, making it user-friendly for both researchers and practitioners. This helps in reducing compute costs and carbon footprint while enabling high-performance results across various machine learning tasks.
https://github.com/huggingface/transformers
GitHub
GitHub - huggingface/transformers: 🤗 Transformers: the model-definition framework for state-of-the-art machine learning models…
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training. - GitHub - huggingface/t...
#jupyter_notebook #computer_vision #deep_learning #drug_discovery #forecasting #large_language_models #mxnet #nlp #paddlepaddle #pytorch #recommender_systems #speech_recognition #speech_synthesis #tensorflow #tensorflow2 #translation
This repository provides top-quality deep learning examples that are easy to train and deploy on NVIDIA GPUs. It includes a wide range of models for computer vision, natural language processing, recommender systems, speech to text, and more. These examples are updated monthly and come in Docker containers with the latest NVIDIA software, ensuring the best performance. The models support multiple GPUs and nodes, and some are optimized for Tensor Cores, which can significantly speed up training. This makes it easier for users to achieve high accuracy and performance in their deep learning projects.
https://github.com/NVIDIA/DeepLearningExamples
This repository provides top-quality deep learning examples that are easy to train and deploy on NVIDIA GPUs. It includes a wide range of models for computer vision, natural language processing, recommender systems, speech to text, and more. These examples are updated monthly and come in Docker containers with the latest NVIDIA software, ensuring the best performance. The models support multiple GPUs and nodes, and some are optimized for Tensor Cores, which can significantly speed up training. This makes it easier for users to achieve high accuracy and performance in their deep learning projects.
https://github.com/NVIDIA/DeepLearningExamples
GitHub
GitHub - NVIDIA/DeepLearningExamples: State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with…
State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterprise-grade infrastructure. - NVIDIA/DeepLearningExamples
#python #agent #agents #ai_search #chatbot #chatgpt #data_pipelines #deep_learning #document_parser #document_understanding #genai #graph #graphrag #llm #nlp #pdf_to_text #preprocessing #rag #retrieval_augmented_generation #table_structure_recognition #text2sql
RAGFlow is an open-source tool that helps businesses answer questions accurately using large language models and deep document understanding. It extracts information from various complex data formats, such as Word documents, Excel files, and web pages, and provides grounded citations to support its answers. You can try a demo online or set it up on your own server using Docker. The setup is relatively straightforward, requiring a few steps like cloning the repository, building the Docker image, and configuring the system settings. RAGFlow offers key features like template-based chunking, reduced hallucinations, and compatibility with multiple data sources, making it a powerful tool for truthful question-answering capabilities. This benefits users by providing reliable and explainable answers, streamlining their workflow, and supporting integration with their business systems.
https://github.com/infiniflow/ragflow
RAGFlow is an open-source tool that helps businesses answer questions accurately using large language models and deep document understanding. It extracts information from various complex data formats, such as Word documents, Excel files, and web pages, and provides grounded citations to support its answers. You can try a demo online or set it up on your own server using Docker. The setup is relatively straightforward, requiring a few steps like cloning the repository, building the Docker image, and configuring the system settings. RAGFlow offers key features like template-based chunking, reduced hallucinations, and compatibility with multiple data sources, making it a powerful tool for truthful question-answering capabilities. This benefits users by providing reliable and explainable answers, streamlining their workflow, and supporting integration with their business systems.
https://github.com/infiniflow/ragflow
GitHub
GitHub - infiniflow/ragflow: RAGFlow is a leading open-source Retrieval-Augmented Generation (RAG) engine that fuses cutting-edge…
RAGFlow is a leading open-source Retrieval-Augmented Generation (RAG) engine that fuses cutting-edge RAG with Agent capabilities to create a superior context layer for LLMs - infiniflow/ragflow
#python #emnlp2024 #knowledge_curation #large_language_models #naacl #nlp #report_generation #retrieval_augmented_generation
STORM is a system that helps you write articles like those on Wikipedia by using internet searches. Here’s how it benefits you STORM conducts internet research, collects references, and generates an outline for your topic.
- **Collaborative Feature** You can install STORM using `pip install knowledge-storm` and customize it according to your needs.
- **User-Friendly**: Over 70,000 people have used STORM, and it helps experienced Wikipedia editors in their pre-writing stage.
This system makes researching and writing articles much easier and more efficient.
https://github.com/stanford-oval/storm
STORM is a system that helps you write articles like those on Wikipedia by using internet searches. Here’s how it benefits you STORM conducts internet research, collects references, and generates an outline for your topic.
- **Collaborative Feature** You can install STORM using `pip install knowledge-storm` and customize it according to your needs.
- **User-Friendly**: Over 70,000 people have used STORM, and it helps experienced Wikipedia editors in their pre-writing stage.
This system makes researching and writing articles much easier and more efficient.
https://github.com/stanford-oval/storm
GitHub
GitHub - stanford-oval/storm: An LLM-powered knowledge curation system that researches a topic and generates a full-length report…
An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations. - stanford-oval/storm
#mdx #deep_learning #hacktoberfest #nlp #transformers
The Hugging Face course teaches you how to use Transformers for natural language processing tasks. You'll learn about the Hugging Face ecosystem, including tools like Transformers, Datasets, Tokenizers, and Accelerate, as well as the Hugging Face Hub. This free course helps you understand how to fine-tune models and share your results. It's beneficial because it provides hands-on experience with popular AI libraries and allows you to build and showcase your own projects on the Hugging Face platform.
https://github.com/huggingface/course
The Hugging Face course teaches you how to use Transformers for natural language processing tasks. You'll learn about the Hugging Face ecosystem, including tools like Transformers, Datasets, Tokenizers, and Accelerate, as well as the Hugging Face Hub. This free course helps you understand how to fine-tune models and share your results. It's beneficial because it provides hands-on experience with popular AI libraries and allows you to build and showcase your own projects on the Hugging Face platform.
https://github.com/huggingface/course
GitHub
GitHub - huggingface/course: The Hugging Face course on Transformers
The Hugging Face course on Transformers. Contribute to huggingface/course development by creating an account on GitHub.
❤1
#python #ai #artificial_intelligence #cython #data_science #deep_learning #entity_linking #machine_learning #named_entity_recognition #natural_language_processing #neural_network #neural_networks #nlp #nlp_library #python #spacy #text_classification #tokenization
spaCy is a powerful tool for understanding and processing human language. It helps computers analyze text by breaking it into parts like words, sentences, and entities (like names or places). This makes it useful for tasks such as identifying who is doing what in a sentence or finding specific information from large texts. Using spaCy can save time and improve accuracy compared to manual analysis. It supports many languages and integrates well with advanced models like BERT, making it ideal for real-world applications.
https://github.com/explosion/spaCy
spaCy is a powerful tool for understanding and processing human language. It helps computers analyze text by breaking it into parts like words, sentences, and entities (like names or places). This makes it useful for tasks such as identifying who is doing what in a sentence or finding specific information from large texts. Using spaCy can save time and improve accuracy compared to manual analysis. It supports many languages and integrates well with advanced models like BERT, making it ideal for real-world applications.
https://github.com/explosion/spaCy
GitHub
GitHub - explosion/spaCy: 💫 Industrial-strength Natural Language Processing (NLP) in Python
💫 Industrial-strength Natural Language Processing (NLP) in Python - explosion/spaCy
#python #bot #bot_framework #botkit #bots #chatbot #chatbots #chatbots_framework #conversation_driven_development #conversational_agents #conversational_ai #conversational_bots #machine_learning #machine_learning_library #mitie #natural_language_processing #nlp #nlu #rasa #spacy #wit
Rasa is an open-source framework that helps build advanced chatbots. It allows developers to create contextual assistants that can have layered conversations, making interactions more natural. Rasa supports integration with various platforms like Facebook Messenger, Slack, and Google Home Actions. This flexibility and customization capability make it a popular choice for businesses to automate customer support and enhance user experience. By using Rasa, users can create intelligent chatbots that understand and respond to user inputs effectively, improving communication and engagement.
https://github.com/RasaHQ/rasa
Rasa is an open-source framework that helps build advanced chatbots. It allows developers to create contextual assistants that can have layered conversations, making interactions more natural. Rasa supports integration with various platforms like Facebook Messenger, Slack, and Google Home Actions. This flexibility and customization capability make it a popular choice for businesses to automate customer support and enhance user experience. By using Rasa, users can create intelligent chatbots that understand and respond to user inputs effectively, improving communication and engagement.
https://github.com/RasaHQ/rasa
GitHub
GitHub - RasaHQ/rasa: 💬 Open source machine learning framework to automate text- and voice-based conversations: NLU, dialogue…
💬 Open source machine learning framework to automate text- and voice-based conversations: NLU, dialogue management, connect to Slack, Facebook, and more - Create chatbots and voice assistants - R...