#python #gemini #gemini_ai #gemini_api #gemini_flash #gemini_pro #information_extration #large_language_models #llm #nlp #python #structured_data
**LangExtract** is a free Python library that uses AI models like Gemini to pull structured data—like names, emotions, or meds—from messy text such as reports or books. It links every fact to its exact spot in the original, creates interactive visuals for easy checks, handles huge files fast with chunking and parallel runs, and works with cloud or local models without fine-tuning. You benefit by quickly turning unstructured docs into reliable, organized data for analysis, saving time and boosting accuracy in fields like healthcare or research.
https://github.com/google/langextract
**LangExtract** is a free Python library that uses AI models like Gemini to pull structured data—like names, emotions, or meds—from messy text such as reports or books. It links every fact to its exact spot in the original, creates interactive visuals for easy checks, handles huge files fast with chunking and parallel runs, and works with cloud or local models without fine-tuning. You benefit by quickly turning unstructured docs into reliable, organized data for analysis, saving time and boosting accuracy in fields like healthcare or research.
https://github.com/google/langextract
GitHub
GitHub - google/langextract: A Python library for extracting structured information from unstructured text using LLMs with precise…
A Python library for extracting structured information from unstructured text using LLMs with precise source grounding and interactive visualization. - google/langextract
#python #ai_tool #darkweb #darkweb_osint #investigation_tool #llm_powered #osint #osint_tool
Robin is an AI tool that searches and scrapes the dark web, refines queries with large language models, filters results, and produces a concise investigation summary you can save or export, with Docker and CLI options and support for multiple LLMs (OpenAI, Anthropic, Gemini, local models) to fit your workflow. This helps you save hours of manual searching by automating multi-engine dark-web searches, scraping Onion sites via Tor, filtering noise with AI, and producing ready-to-use reports for faster, more focused OSINT investigations.
https://github.com/apurvsinghgautam/robin
Robin is an AI tool that searches and scrapes the dark web, refines queries with large language models, filters results, and produces a concise investigation summary you can save or export, with Docker and CLI options and support for multiple LLMs (OpenAI, Anthropic, Gemini, local models) to fit your workflow. This helps you save hours of manual searching by automating multi-engine dark-web searches, scraping Onion sites via Tor, filtering noise with AI, and producing ready-to-use reports for faster, more focused OSINT investigations.
https://github.com/apurvsinghgautam/robin
GitHub
GitHub - apurvsinghgautam/robin: AI-Powered Dark Web OSINT Tool
AI-Powered Dark Web OSINT Tool. Contribute to apurvsinghgautam/robin development by creating an account on GitHub.
#python #agent #agentic_ai #agentic_framework #agentic_workflow #ai #ai_agents #ai_companion #ai_roleplay #benchmark #framework #llm #mcp #memory #open_source #python #sandbox
MemU lets AI systems take in conversations, documents, and media, turn them into structured memories, and store them in a clear three-layer file system. It offers both fast embedding search and deeper LLM-based retrieval, works with many data types, and supports cloud or self-hosted setups with simple APIs. This helps you build AI agents that truly remember past interactions, retrieve the right context when needed, and improve over time, making your applications more accurate, personal, and efficient.
https://github.com/NevaMind-AI/memU
MemU lets AI systems take in conversations, documents, and media, turn them into structured memories, and store them in a clear three-layer file system. It offers both fast embedding search and deeper LLM-based retrieval, works with many data types, and supports cloud or self-hosted setups with simple APIs. This helps you build AI agents that truly remember past interactions, retrieve the right context when needed, and improve over time, making your applications more accurate, personal, and efficient.
https://github.com/NevaMind-AI/memU
GitHub
GitHub - NevaMind-AI/memU: Memory for 24/7 proactive agents like openclaw (moltbot, clawdbot).
Memory for 24/7 proactive agents like openclaw (moltbot, clawdbot). - NevaMind-AI/memU
#javascript #agent #agentic #agentic_ai #ai #ai_agents #automation #cursor #design #figma #generative_ai #llm #llms #mcp #model_context_protocol
Cursor Talk to Figma MCP lets Cursor AI read and edit your Figma designs directly, using tools like `get_selection` for info, `set_text_content` for bulk text changes, `create_rectangle` for shapes, and `set_instance_overrides` for components. Setup is quick: install Bun, run `bun setup` and `bun socket`, add the Figma plugin. This saves you hours by skipping context switches, automating repetitive tasks like text replacement or override propagation, speeding up design-to-code workflows, and keeping everything in sync for faster, precise builds.
https://github.com/grab/cursor-talk-to-figma-mcp
Cursor Talk to Figma MCP lets Cursor AI read and edit your Figma designs directly, using tools like `get_selection` for info, `set_text_content` for bulk text changes, `create_rectangle` for shapes, and `set_instance_overrides` for components. Setup is quick: install Bun, run `bun setup` and `bun socket`, add the Figma plugin. This saves you hours by skipping context switches, automating repetitive tasks like text replacement or override propagation, speeding up design-to-code workflows, and keeping everything in sync for faster, precise builds.
https://github.com/grab/cursor-talk-to-figma-mcp
GitHub
GitHub - grab/cursor-talk-to-figma-mcp: TalkToFigma: MCP integration between AI Agent (Cursor, Claude) and Figma, allowing Agentic…
TalkToFigma: MCP integration between AI Agent (Cursor, Claude) and Figma, allowing Agentic AI to communicate with Figma for reading designs and modifying them programmatically. - grab/cursor-talk-t...
#typescript #acp #ai #ai_agent #banana #chat #chatbot #claude_code #codex #cowork #excel #gemini #gemini_cli #gemini_pro #llm #multi_agent #nano_banana #office #qwen_code #skills #webui
AionUi is a free, open-source app that gives your CLI AI tools like Gemini CLI, Claude Code, and Qwen Code a simple graphical interface on macOS, Windows, or Linux. It auto-detects them for easy chatting, saves talks locally with multi-sessions, organizes files smartly, previews 9+ formats like PDF or code instantly, generates/editing images, and offers web access. You benefit by ditching complex commands for quick, secure AI help in office tasks, coding, or data work—saving time and boosting productivity without data leaving your device.
https://github.com/iOfficeAI/AionUi
AionUi is a free, open-source app that gives your CLI AI tools like Gemini CLI, Claude Code, and Qwen Code a simple graphical interface on macOS, Windows, or Linux. It auto-detects them for easy chatting, saves talks locally with multi-sessions, organizes files smartly, previews 9+ formats like PDF or code instantly, generates/editing images, and offers web access. You benefit by ditching complex commands for quick, secure AI help in office tasks, coding, or data work—saving time and boosting productivity without data leaving your device.
https://github.com/iOfficeAI/AionUi
GitHub
GitHub - iOfficeAI/AionUi: Free, local, open-source 24/7 Cowork app and OpenClaw for Gemini CLI, Claude Code, Codex, OpenCode,…
Free, local, open-source 24/7 Cowork app and OpenClaw for Gemini CLI, Claude Code, Codex, OpenCode, Qwen Code, Goose CLI, Auggie, and more | 🌟 Star if you like it! - iOfficeAI/AionUi
❤1
#jupyter_notebook #chinese_llm #chinese_nlp #finetune #generative_ai #instruct_gpt #instruction_set #llama #llm #lora #open_models #open_source #open_source_models #qlora
AirLLM is a tool that lets you run very large AI models on computers with limited memory by using a smart layer-by-layer loading technique instead of traditional compression methods. You can run a 70-billion-parameter model on just 4GB of GPU memory, or even a 405-billion-parameter model on 8GB, without losing model quality. The benefit is that you can use powerful AI models on affordable hardware without expensive upgrades, and the tool also offers optional compression features that can speed up performance by up to 3 times while maintaining accuracy.
https://github.com/lyogavin/airllm
AirLLM is a tool that lets you run very large AI models on computers with limited memory by using a smart layer-by-layer loading technique instead of traditional compression methods. You can run a 70-billion-parameter model on just 4GB of GPU memory, or even a 405-billion-parameter model on 8GB, without losing model quality. The benefit is that you can use powerful AI models on affordable hardware without expensive upgrades, and the tool also offers optional compression features that can speed up performance by up to 3 times while maintaining accuracy.
https://github.com/lyogavin/airllm
GitHub
GitHub - lyogavin/airllm: AirLLM 70B inference with single 4GB GPU
AirLLM 70B inference with single 4GB GPU. Contribute to lyogavin/airllm development by creating an account on GitHub.
#python #deepseek #demo #easy #embedding #flask #gpt #huggingface_transformers #llm #mcp #multimodal #openai #qwen #rag #sentence_transformers #ui #vllm #vlm
UltraRAG is a lightweight framework that makes building retrieval-augmented generation (RAG) systems simple and fast. It uses a low-code approach where you write just dozens of lines of YAML configuration instead of complex code to create sophisticated AI workflows with conditional logic and loops. The framework includes a visual development environment where you can drag-and-drop to build pipelines, adjust parameters in real-time, and instantly convert your logic into interactive chat applications. This means you can deploy powerful AI systems that ground answers in your own data—reducing hallucinations and improving accuracy—without needing extensive coding expertise or lengthy development cycles.
https://github.com/OpenBMB/UltraRAG
UltraRAG is a lightweight framework that makes building retrieval-augmented generation (RAG) systems simple and fast. It uses a low-code approach where you write just dozens of lines of YAML configuration instead of complex code to create sophisticated AI workflows with conditional logic and loops. The framework includes a visual development environment where you can drag-and-drop to build pipelines, adjust parameters in real-time, and instantly convert your logic into interactive chat applications. This means you can deploy powerful AI systems that ground answers in your own data—reducing hallucinations and improving accuracy—without needing extensive coding expertise or lengthy development cycles.
https://github.com/OpenBMB/UltraRAG
GitHub
GitHub - OpenBMB/UltraRAG: A Low-Code MCP Framework for Building Complex and Innovative RAG Pipelines
A Low-Code MCP Framework for Building Complex and Innovative RAG Pipelines - OpenBMB/UltraRAG
#python #abliteration #llm #transformer
Heretic is an automated tool that removes safety restrictions from AI language models while preserving their intelligence and capabilities. It uses advanced mathematical techniques called directional ablation to identify and disable the "refusal mechanisms" that prevent models from answering certain questions. The key benefit is that anyone can use it with a simple command—no technical expertise needed. Unlike manual methods that often damage model quality, Heretic achieves the same level of censorship removal with significantly better preservation of the original model's reasoning abilities, as measured by lower KL divergence scores. This means you get an uncensored model that still thinks clearly and produces high-quality responses.
https://github.com/p-e-w/heretic
Heretic is an automated tool that removes safety restrictions from AI language models while preserving their intelligence and capabilities. It uses advanced mathematical techniques called directional ablation to identify and disable the "refusal mechanisms" that prevent models from answering certain questions. The key benefit is that anyone can use it with a simple command—no technical expertise needed. Unlike manual methods that often damage model quality, Heretic achieves the same level of censorship removal with significantly better preservation of the original model's reasoning abilities, as measured by lower KL divergence scores. This means you get an uncensored model that still thinks clearly and produces high-quality responses.
https://github.com/p-e-w/heretic
GitHub
GitHub - p-e-w/heretic: Fully automatic censorship removal for language models
Fully automatic censorship removal for language models - p-e-w/heretic
#python #ai #claude #gemini #llama #llm #openai
You can access powerful AI language models for free or with trial credits through multiple legitimate platforms. Services like OpenRouter, Google AI Studio, Groq, and Mistral offer free tiers with varying request limits, while others like Fireworks, Baseten, and Inference.net provide trial credits ranging from $1 to $30. These platforms support diverse models including Llama, Gemma, Qwen, and DeepSeek, enabling you to build and test AI applications without upfront costs. The benefit is clear: you can prototype, develop, and deploy AI-powered features while managing your budget effectively, with options to scale up as your needs grow.
https://github.com/cheahjs/free-llm-api-resources
You can access powerful AI language models for free or with trial credits through multiple legitimate platforms. Services like OpenRouter, Google AI Studio, Groq, and Mistral offer free tiers with varying request limits, while others like Fireworks, Baseten, and Inference.net provide trial credits ranging from $1 to $30. These platforms support diverse models including Llama, Gemma, Qwen, and DeepSeek, enabling you to build and test AI applications without upfront costs. The benefit is clear: you can prototype, develop, and deploy AI-powered features while managing your budget effectively, with options to scale up as your needs grow.
https://github.com/cheahjs/free-llm-api-resources
GitHub
GitHub - cheahjs/free-llm-api-resources: A list of free LLM inference resources accessible via API.
A list of free LLM inference resources accessible via API. - cheahjs/free-llm-api-resources
#rust #ai #ai_ocr #attention_mechanism #gnn #gnn_model #gnns #graph #graph_neural_networks #llm_inference #low_latency #mincut #neo4j #ocr #onnx #rust #vector #wasm
RuVector is a free, open-source vector database that gets smarter with every query. Unlike static databases, it learns from usage via GNN layers, runs LLMs locally with no cloud costs, supports graph queries like Neo4j, scales freely across nodes, and deploys as a single self-booting file (125ms startup). Run with `npx ruvector`. You benefit from faster, more accurate AI search that improves automatically, zero operating costs, full offline/privacy control, and easy scaling—perfect for RAG, agents, or edge apps without vendor lock-in.
https://github.com/ruvnet/ruvector
RuVector is a free, open-source vector database that gets smarter with every query. Unlike static databases, it learns from usage via GNN layers, runs LLMs locally with no cloud costs, supports graph queries like Neo4j, scales freely across nodes, and deploys as a single self-booting file (125ms startup). Run with `npx ruvector`. You benefit from faster, more accurate AI search that improves automatically, zero operating costs, full offline/privacy control, and easy scaling—perfect for RAG, agents, or edge apps without vendor lock-in.
https://github.com/ruvnet/ruvector
GitHub
GitHub - ruvnet/ruvector: RuVector is a High Performance, Real-Time, Self-Learning, Vector Graph Neural Network, and Database built…
RuVector is a High Performance, Real-Time, Self-Learning, Vector Graph Neural Network, and Database built in Rust. - ruvnet/ruvector
👍2❤1
#rust #ai_gateway #ai_gateway_support #envoy #envoyproxy #gateway #generative_ai #llm_gateway #llm_inference #llm_proxy #llm_routing #llmops #llms #openai #prompt #proxy #proxy_server #routing
Plano is an AI-native proxy server that handles key tasks for agentic apps like routing between agents, smart LLM model selection, safety guardrails, and automatic traces for observability. Define agents in simple YAML, write basic HTTP code in any language, and start Plano to run multi-agent systems without custom plumbing or framework lock-in. You benefit by building and shipping reliable agents to production much faster, focusing on core logic while gaining safety, low latency, and easy scaling.
https://github.com/katanemo/plano
Plano is an AI-native proxy server that handles key tasks for agentic apps like routing between agents, smart LLM model selection, safety guardrails, and automatic traces for observability. Define agents in simple YAML, write basic HTTP code in any language, and start Plano to run multi-agent systems without custom plumbing or framework lock-in. You benefit by building and shipping reliable agents to production much faster, focusing on core logic while gaining safety, low latency, and easy scaling.
https://github.com/katanemo/plano
GitHub
GitHub - katanemo/plano: Delivery infrastructure for agentic apps - Plano is an AI-native proxy and data plane that offloads plumbing…
Delivery infrastructure for agentic apps - Plano is an AI-native proxy and data plane that offloads plumbing work, so you stay focused on your agent's core logic (via any AI framework). - k...
#javascript #ai #algorithm #artificial_intelligence #chatgpt #claude #cursor #deep_learning #deepseek #gemini #generative_ai #gpt #llm #mcp #openai #python #rag #vibe_coding #vibecoding #vue #vuepress
鱼皮的 AI知识库 offers a free Vibe Coding tutorial for beginners, teaching AI-powered programming with natural language prompts to build and monetize apps fast—no coding skills needed. It covers tools, projects, tips, and paths like making your first work in 10 minutes, plus AI guides on DeepSeek, Cursor, and more. You benefit by quickly creating profitable products, breaking tech barriers, and enjoying AI perks to improve life and work. Start at ai.codefather.cn/vibe.
https://github.com/liyupi/ai-guide
鱼皮的 AI知识库 offers a free Vibe Coding tutorial for beginners, teaching AI-powered programming with natural language prompts to build and monetize apps fast—no coding skills needed. It covers tools, projects, tips, and paths like making your first work in 10 minutes, plus AI guides on DeepSeek, Cursor, and more. You benefit by quickly creating profitable products, breaking tech barriers, and enjoying AI perks to improve life and work. Start at ai.codefather.cn/vibe.
https://github.com/liyupi/ai-guide
鱼皮AI导航
🌟 AI 编程零基础入门教程 Vibe Coding - 鱼皮的 AI 知识库(免费) - 鱼皮AI导航
# 鱼皮 Vibe Coding 零基础入门教程
大家好,我是程序员鱼皮。
如今 Vibe Coding 已经火遍全网,不仅是程序员,连设计师、产品运营、甚至完全不懂技术的人都开始用 Vibe C。鱼皮AI导航收录全球AI工具网站应用,专业学习资源资讯知识库,AI学习与交流社区。
大家好,我是程序员鱼皮。
如今 Vibe Coding 已经火遍全网,不仅是程序员,连设计师、产品运营、甚至完全不懂技术的人都开始用 Vibe C。鱼皮AI导航收录全球AI工具网站应用,专业学习资源资讯知识库,AI学习与交流社区。
#typescript #agent #agentic #agentic_framework #agentic_workflow #ai #ai_agents #bytedance #deep_research #harness #langchain #langgraph #langmanus #llm #multi_agent #nodejs #podcast #python #superagent #typescript
DeerFlow 2.0 is an open-source super agent harness that orchestrates multiple sub-agents, memory systems, and sandboxed execution environments to accomplish complex tasks. Built on LangGraph and LangChain, it combines research, coding, and content creation capabilities with extensible skills and tools. The platform features isolated Docker containers for safe execution, long-term memory that learns your preferences, and the ability to spawn sub-agents that work in parallel on different task angles. You benefit from dramatically reduced research and automation time—tasks that typically take hours complete in minutes—while maintaining full transparency and control over agent decisions through human-in-the-loop collaboration. Whether you need deep research reports, data analysis, slide decks, or custom workflows, DeerFlow handles multi-step complexity without requiring extensive coding knowledge.
https://github.com/bytedance/deer-flow
DeerFlow 2.0 is an open-source super agent harness that orchestrates multiple sub-agents, memory systems, and sandboxed execution environments to accomplish complex tasks. Built on LangGraph and LangChain, it combines research, coding, and content creation capabilities with extensible skills and tools. The platform features isolated Docker containers for safe execution, long-term memory that learns your preferences, and the ability to spawn sub-agents that work in parallel on different task angles. You benefit from dramatically reduced research and automation time—tasks that typically take hours complete in minutes—while maintaining full transparency and control over agent decisions through human-in-the-loop collaboration. Whether you need deep research reports, data analysis, slide decks, or custom workflows, DeerFlow handles multi-step complexity without requiring extensive coding knowledge.
https://github.com/bytedance/deer-flow
GitHub
GitHub - bytedance/deer-flow: An open-source SuperAgent harness that researches, codes, and creates. With the help of sandboxes…
An open-source SuperAgent harness that researches, codes, and creates. With the help of sandboxes, memories, tools, skills and subagents, it handles different levels of tasks that could take minute...
#typescript #agent #agent_development #ai_agent #claude #claude_code #educational #llm #python #teaching #tutorial
Claude Code is an AI agent framework that uses a simple loop: send messages to Claude, check if it needs tools, execute those tools, and repeat. The benefit is that you can build powerful autonomous agents by layering one feature at a time—from basic tool use to multi-agent teams—without rewriting the core loop. This modular approach lets you start simple with bash commands and scale to complex workflows with planning, skill loading, background tasks, and team coordination, making it easier to automate development work and delegate entire projects to AI agents.
https://github.com/shareAI-lab/learn-claude-code
Claude Code is an AI agent framework that uses a simple loop: send messages to Claude, check if it needs tools, execute those tools, and repeat. The benefit is that you can build powerful autonomous agents by layering one feature at a time—from basic tool use to multi-agent teams—without rewriting the core loop. This modular approach lets you start simple with bash commands and scale to complex workflows with planning, skill loading, background tasks, and team coordination, making it easier to automate development work and delegate entire projects to AI agents.
https://github.com/shareAI-lab/learn-claude-code
GitHub
GitHub - shareAI-lab/learn-claude-code: Bash is all you need - A nano Claude Code–like agent, built from 0 to 1
Bash is all you need - A nano Claude Code–like agent, built from 0 to 1 - shareAI-lab/learn-claude-code
#typescript #agentic_ai #ai_agents #claude_code #cli #codex #coding_agents #cursor_agent #desktop_app #developer_tools #electron #git_worktree #llm #mcp #opencode #orchestration #parallel_agents #terminal #tui #vibe_coding #worktrees
Superset is a turbocharged macOS terminal for running 10+ CLI coding agents like Claude Code, Cursor, and GitHub Copilot in parallel. It isolates tasks in separate Git worktrees to avoid interference, lets you monitor progress from one dashboard, review changes with a built-in diff viewer, and switch contexts quickly. You benefit by coding 10x faster, shipping more without context-switching delays or conflicts, saving time on development workflows.
https://github.com/superset-sh/superset
Superset is a turbocharged macOS terminal for running 10+ CLI coding agents like Claude Code, Cursor, and GitHub Copilot in parallel. It isolates tasks in separate Git worktrees to avoid interference, lets you monitor progress from one dashboard, review changes with a built-in diff viewer, and switch contexts quickly. You benefit by coding 10x faster, shipping more without context-switching delays or conflicts, saving time on development workflows.
https://github.com/superset-sh/superset
GitHub
GitHub - superset-sh/superset: IDE for the AI Agents Era - Run an army of Claude Code, Codex, etc. on your machine
IDE for the AI Agents Era - Run an army of Claude Code, Codex, etc. on your machine - superset-sh/superset
❤1