opencv/dldt
Deep Learning Deployment Toolkit
Language: C++
#deep_learning #inference #inference_engine #performance
Stars: 97 Issues: 2 Forks: 13
https://github.com/opencv/dldt
Deep Learning Deployment Toolkit
Language: C++
#deep_learning #inference #inference_engine #performance
Stars: 97 Issues: 2 Forks: 13
https://github.com/opencv/dldt
GitHub
GitHub - openvinotoolkit/openvino: OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference - openvinotoolkit/openvino
BMW-InnovationLab/BMW-TensorFlow-Training-GUI
This repository allows you to get started with a gui based training a State-of-the-art Deep Learning model with little to no configuration needed! Training with TensorFlow has never been so easy.
Language: Python
#computer_vision #computervision #deep_learning #deep_neural_networks #deeplearning #detection_api #docker #gui #inference_api #neural_network #nvidia_docker #object_detection #objectdetection #resnet #rest_api #ssd #tensorboard #tensorflow #tensorflow_gui #tensorflow_training
Stars: 176 Issues: 0 Forks: 16
https://github.com/BMW-InnovationLab/BMW-TensorFlow-Training-GUI
This repository allows you to get started with a gui based training a State-of-the-art Deep Learning model with little to no configuration needed! Training with TensorFlow has never been so easy.
Language: Python
#computer_vision #computervision #deep_learning #deep_neural_networks #deeplearning #detection_api #docker #gui #inference_api #neural_network #nvidia_docker #object_detection #objectdetection #resnet #rest_api #ssd #tensorboard #tensorflow #tensorflow_gui #tensorflow_training
Stars: 176 Issues: 0 Forks: 16
https://github.com/BMW-InnovationLab/BMW-TensorFlow-Training-GUI
GitHub
GitHub - BMW-InnovationLab/BMW-TensorFlow-Training-GUI: This repository allows you to get started with a gui based training a State…
This repository allows you to get started with a gui based training a State-of-the-art Deep Learning model with little to no configuration needed! NoCode training with TensorFlow has never been so ...
BMW-InnovationLab/BMW-TensorFlow-Inference-API-GPU
This is a repository for an object detection inference API using the Tensorflow framework.
Language: Python
#api #computer_vision #deep_learning #deep_neural_networks #detection_inference_api #docker #dockerfile #gpu #inference #neural_network #nvidia #object_detection #rest_api #tensorflow #tensorflow_framework #tensorflow_models
Stars: 150 Issues: 2 Forks: 50
https://github.com/BMW-InnovationLab/BMW-TensorFlow-Inference-API-GPU
This is a repository for an object detection inference API using the Tensorflow framework.
Language: Python
#api #computer_vision #deep_learning #deep_neural_networks #detection_inference_api #docker #dockerfile #gpu #inference #neural_network #nvidia #object_detection #rest_api #tensorflow #tensorflow_framework #tensorflow_models
Stars: 150 Issues: 2 Forks: 50
https://github.com/BMW-InnovationLab/BMW-TensorFlow-Inference-API-GPU
GitHub
GitHub - BMW-InnovationLab/BMW-TensorFlow-Inference-API-GPU: This is a repository for an object detection inference API using the…
This is a repository for an object detection inference API using the Tensorflow framework. - BMW-InnovationLab/BMW-TensorFlow-Inference-API-GPU
vriad/zod
Typescript-first schema validation with static type inference
Language: TypeScript
#inference #schema_definitions #schema_validation #type_inference #typescript
Stars: 295 Issues: 8 Forks: 2
https://github.com/vriad/zod
Typescript-first schema validation with static type inference
Language: TypeScript
#inference #schema_definitions #schema_validation #type_inference #typescript
Stars: 295 Issues: 8 Forks: 2
https://github.com/vriad/zod
GitHub
GitHub - colinhacks/zod: TypeScript-first schema validation with static type inference
TypeScript-first schema validation with static type inference - colinhacks/zod
Tencent/Forward
a library for high performance deep learning inference on NVIDIA GPUs.
Language: C++
#cuda #deep_learning #forward #gpu #inference #inference_engine #keras #neural_network #pytorch #tensorflow #tensorrt
Stars: 102 Issues: 0 Forks: 8
https://github.com/Tencent/Forward
a library for high performance deep learning inference on NVIDIA GPUs.
Language: C++
#cuda #deep_learning #forward #gpu #inference #inference_engine #keras #neural_network #pytorch #tensorflow #tensorrt
Stars: 102 Issues: 0 Forks: 8
https://github.com/Tencent/Forward
GitHub
GitHub - Tencent/Forward: A library for high performance deep learning inference on NVIDIA GPUs.
A library for high performance deep learning inference on NVIDIA GPUs. - GitHub - Tencent/Forward: A library for high performance deep learning inference on NVIDIA GPUs.
tairov/llama2.mojo
Inference Llama 2 in one file of pure 🔥
#inference #llama #llama2 #modular #mojo #parallelize #performance #simd #tensor #vectorization
Stars: 200 Issues: 0 Forks: 7
https://github.com/tairov/llama2.mojo
Inference Llama 2 in one file of pure 🔥
#inference #llama #llama2 #modular #mojo #parallelize #performance #simd #tensor #vectorization
Stars: 200 Issues: 0 Forks: 7
https://github.com/tairov/llama2.mojo
GitHub
GitHub - tairov/llama2.mojo: Inference Llama 2 in one file of pure 🔥
Inference Llama 2 in one file of pure 🔥. Contribute to tairov/llama2.mojo development by creating an account on GitHub.
chengzeyi/stable-fast
An ultra lightweight inference performance optimization library for HuggingFace Diffusers on NVIDIA GPUs.
Language: Python
#cuda #deep_learning #deeplearning #diffusers #inference #inference_engine #performance_optimization #pytorch #stable_diffusion #triton
Stars: 134 Issues: 3 Forks: 5
https://github.com/chengzeyi/stable-fast
An ultra lightweight inference performance optimization library for HuggingFace Diffusers on NVIDIA GPUs.
Language: Python
#cuda #deep_learning #deeplearning #diffusers #inference #inference_engine #performance_optimization #pytorch #stable_diffusion #triton
Stars: 134 Issues: 3 Forks: 5
https://github.com/chengzeyi/stable-fast
GitHub
GitHub - chengzeyi/stable-fast: Best inference performance optimization framework for HuggingFace Diffusers on NVIDIA GPUs.
Best inference performance optimization framework for HuggingFace Diffusers on NVIDIA GPUs. - chengzeyi/stable-fast
Fuzzy-Search/realtime-bakllava
llama.cpp with BakLLaVA model describes what does it see
Language: Python
#bakllavva #cpp #demo_application #inference #llama #llamacpp #llm
Stars: 141 Issues: 1 Forks: 15
https://github.com/Fuzzy-Search/realtime-bakllava
llama.cpp with BakLLaVA model describes what does it see
Language: Python
#bakllavva #cpp #demo_application #inference #llama #llamacpp #llm
Stars: 141 Issues: 1 Forks: 15
https://github.com/Fuzzy-Search/realtime-bakllava
GitHub
GitHub - Fuzzy-Search/realtime-bakllava: llama.cpp with BakLLaVA model describes what does it see
llama.cpp with BakLLaVA model describes what does it see - Fuzzy-Search/realtime-bakllava
hpcaitech/SwiftInfer
Efficient AI Inference & Serving
Language: Python
#artificial_intelligence #deep_learning #gpt #inference #llama #llama2 #llm_inference #llm_serving
Stars: 299 Issues: 3 Forks: 14
https://github.com/hpcaitech/SwiftInfer
Efficient AI Inference & Serving
Language: Python
#artificial_intelligence #deep_learning #gpt #inference #llama #llama2 #llm_inference #llm_serving
Stars: 299 Issues: 3 Forks: 14
https://github.com/hpcaitech/SwiftInfer
GitHub
GitHub - hpcaitech/SwiftInfer: Efficient AI Inference & Serving
Efficient AI Inference & Serving. Contribute to hpcaitech/SwiftInfer development by creating an account on GitHub.
arc53/llm-price-compass
LLM provider price comparison, gpu benchmarks to price per token calculation, gpu benchmark table
Language: TypeScript
#benchmark #gpu #inference_comparison #llm #llm_comparison #llm_inference #llm_price
Stars: 138 Issues: 1 Forks: 5
https://github.com/arc53/llm-price-compass
LLM provider price comparison, gpu benchmarks to price per token calculation, gpu benchmark table
Language: TypeScript
#benchmark #gpu #inference_comparison #llm #llm_comparison #llm_inference #llm_price
Stars: 138 Issues: 1 Forks: 5
https://github.com/arc53/llm-price-compass
GitHub
GitHub - arc53/llm-price-compass: This project collects GPU benchmarks from various cloud providers and compares them to fixed…
This project collects GPU benchmarks from various cloud providers and compares them to fixed per token costs. Use our tool for efficient LLM GPU selections and cost-effective AI models. LLM provide...
zml/zml
High performance AI inference stack. Built for production. @ziglang / @openxla / MLIR / @bazelbuild
Language: Zig
#ai #bazel #hpc #inference #xla #zig
Stars: 691 Issues: 1 Forks: 19
https://github.com/zml/zml
High performance AI inference stack. Built for production. @ziglang / @openxla / MLIR / @bazelbuild
Language: Zig
#ai #bazel #hpc #inference #xla #zig
Stars: 691 Issues: 1 Forks: 19
https://github.com/zml/zml
GitHub
GitHub - zml/zml: High performance AI inference stack. Built for production. @ziglang / @openxla / MLIR / @bazelbuild
High performance AI inference stack. Built for production. @ziglang / @openxla / MLIR / @bazelbuild - zml/zml
AgibotTech/agibot_x1_infer
The inference module for AgiBot X1.
Language: C++
#inference #open_source #robotics
Stars: 455 Issues: 2 Forks: 152
https://github.com/AgibotTech/agibot_x1_infer
The inference module for AgiBot X1.
Language: C++
#inference #open_source #robotics
Stars: 455 Issues: 2 Forks: 152
https://github.com/AgibotTech/agibot_x1_infer
GitHub
GitHub - AgibotTech/agibot_x1_infer: The inference module for AgiBot X1.
The inference module for AgiBot X1. Contribute to AgibotTech/agibot_x1_infer development by creating an account on GitHub.
thu-ml/SageAttention
Quantized Attention that achieves speedups of 2.1x and 2.7x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models.
Language: Python
#attention #inference_acceleration #llm #quantization
Stars: 145 Issues: 6 Forks: 3
https://github.com/thu-ml/SageAttention
Quantized Attention that achieves speedups of 2.1x and 2.7x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models.
Language: Python
#attention #inference_acceleration #llm #quantization
Stars: 145 Issues: 6 Forks: 3
https://github.com/thu-ml/SageAttention
GitHub
GitHub - thu-ml/SageAttention: Quantized Attention that achieves speedups of 2.1-3.1x and 2.7-5.1x compared to FlashAttention2…
Quantized Attention that achieves speedups of 2.1-3.1x and 2.7-5.1x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models. - thu-ml/SageAt...