Open Source Tools for Developers

Explore top LinkedIn content from expert professionals.

  • View profile for Alex Wang
    Alex Wang Alex Wang is an Influencer

    Learn AI Together - I share my learning journey into AI & Data Science here, 90% buzzword-free. Follow me and let's grow together!

    1,134,060 followers

    The best open-source data science agent I’ve tried so far: 𝗗𝗮𝘁𝗮 𝗖𝗼𝗽𝗶𝗹𝗼𝘁 — it can build an entire notebook workflow from a single prompt. If you’ve worked in data science, you know how most AI coding tools fall short when it comes to Jupyter Notebooks. They don’t handle the notebook structure well — no context, no new cells, no real understanding of the data flow. Data Copilot changes that. It feels like Cursor, but built for data scientists. I just drop it into my Jupyter environment, and it picks up the context of my files and datasets automatically. *It's open source — install it in seconds: 𝗽𝗶𝗽 𝗶𝗻𝘀𝘁𝗮𝗹𝗹 𝗺𝗶𝘁𝗼-𝗮𝗶 𝗺𝗶𝘁𝗼𝘀𝗵𝗲𝗲𝘁 Here’s what I’ve seen it do: 🔹 From a single prompt, build a full machine learning notebook, including data importing, data cleaning, model training and testing 🔹 Take a notebook and swap all of the Matplotlib code for Plotly code 🔹 Automatically catch errors and debug them We’ve seen great AI tools for software developers. Data Copilot is one of the first tools that is excellent for Data Science workflows. Key features: 🔹 An AI agent for full notebook creation and editing 🔹 An AI Chat for editing specific cells 🔹 Automatic error debugging from the AI 🔹 Visual edits for DataFrames and Charts 📍Docs here: https://lnkd.in/gSJEMshP #productivity #datascience #machinelearning #aitools #opensource

  • View profile for Sahar Mor

    I help researchers and builders make sense of AI | ex-Stripe | aitidbits.ai | Angel Investor

    41,674 followers

    The open-source AI agent ecosystem is exploding, but most market maps and guides cater to VCs rather than builders. As someone in the trenches of agent development, I've found this frustrating. That's why I've created a comprehensive list of the open-source tools I've personally found effective in production. The overview includes 38 packages across: -> Agent orchestration frameworks that go beyond basic LLM wrappers: CrewAI for role-playing agents, AutoGPT for autonomous workflows, Superagent for quick prototyping -> Tools for computer control and browser automation: Open Interpreter for local machine control, Self-Operating Computer for visual automation, LaVague for web agents -> Voice interaction capabilities beyond basic speech-to-text: Ultravox for real-time voice, Whisper for transcription, Vocode for voice-based agents -> Memory systems that enable truly personalized experiences: Mem0 for self-improving memory, Letta for long-term context, LangChain's memory components -> Testing and monitoring solutions for production-grade agents: AgentOps for benchmarking, openllmetry for observability, Voice Lab for evaluation With the holiday season here, it's the perfect time to start building. Post https://lnkd.in/gCySSuS3

  • View profile for Aishwarya Srinivasan
    Aishwarya Srinivasan Aishwarya Srinivasan is an Influencer
    621,544 followers

    The Future of AI is Open-Source! 10 years ago when I started in ML, building out end-to-end ML applications would take you months, to say the least, but in 2025, going from idea to MVP to production happens in weeks, if not days. One of the biggest changes I am observing is "free access to the best tech", which is making the ML application development faster. You don't need to be working in the best-tech company to have access to these, now it is available to everyone, thanks to the open-source community!   I love this visual of the open-source AI stack by ByteByteGo. It lays out the tools/frameworks you can use (for free) and build these AI applications right on your laptop. If you are an AI engineer getting started, checkout the following tools: ↳ Frontend Technologies : Next.js, Vercel, Streamlit ↳ Embeddings and RAG Libraries : Nomic, Jina AI, Cognito, and LLMAware ↳ Backend and Model Access : FastAPI, LangChain, Netflix Metaflow, Ollama, Hugging Face ↳ Data and Retrieval : Postgres, Milvus, Weaviate, PGvector, FAISS ↳ Large Language Models: llama models, Qwen models, Gemma models, Phi models, DeepSeek models, Falcon models ↳ Vision Language Models: VisionLLM v2, Falcon 2 VLM, Qwen-VL Series, PaliGemma ↳ Speech-to-text & Text-to-speech models: OpenAI Whisper, Wav2Vec, DeepSpeech, Tacotron 2, Kokoro TTS, Spark-TTS, Fish Speech v1.5, StyleTTS (I added more models missing in the infographic) Plus, I would recommend checking out the following tools as well: ↳ Agent frameworks: CrewAI, AutoGen, SuperAGI, LangGraph ↳ Model Optimization & Deployment: vLLM, TensorRT, and LoRA methods for model fine-tuning PS: I had shared some ideas about portfolio projects you can build, in an earlier post, so if you are curious about that, check out my past post. Happy Learning 🚀  There is nothing stopping you to start building on your idea! ----------- If you found this useful, please do share it with your network ♻️ Follow me (Aishwarya Srinivasan) for more AI educational content and insights to help you stay up-to-date in the AI space :)

  • View profile for Shreya Khandelwal

    Data Scientist @ Bain | Microsoft AI MVP | Ex-IBMer | LinkedIn Top Voices | GenAI | LLMs | AI & Analytics | 10 x Multi- Hyperscale-Cloud Certified

    29,130 followers

    𝐁𝐮𝐢𝐥𝐝𝐢𝐧𝐠 𝐋𝐋𝐌 𝐀𝐩𝐩𝐬 𝐖𝐢𝐭𝐡𝐨𝐮𝐭 𝐂𝐨𝐝𝐞? 𝐘𝐞𝐬, 𝐢𝐭'𝐬 𝐩𝐨𝐬𝐬𝐢𝐛𝐥𝐞 Whether you're designing RAG pipelines, deploying AI agents, or fine-tuning LLMs, you no longer need to write thousands of lines of code. Here are 6 powerful open-source tools that let you build and deploy LLMs, Agents, and RAG workflows — no-code required 1️⃣ 𝑹𝑨𝑮𝑭𝒍𝒐𝒘: - A visual framework to design Retrieval-Augmented Generation (RAG) pipelines. - Combines document retrieval + LLMs - Great for building QA systems and enterprise knowledge assistants - 𝑮𝒊𝒕𝑯𝒖𝒃 𝒓𝒆𝒑𝒐: https://lnkd.in/gPrPxRVj 2️⃣ 𝒙𝒑𝒂𝒏𝒅𝒆𝒓.𝒂𝒊: - A backend for your AI agents, designed to work across multiple agent stacks - Works with any agent framework (AutoGen, CrewAI, etc.) - Handles memory, vector search, tools, APIs - Ideal for startups building backend agent infra. - 𝑮𝒊𝒕𝑯𝒖𝒃 𝒓𝒆𝒑𝒐: https://lnkd.in/g_bH5cqC 3️⃣ 𝑳𝑳𝒂𝑴𝑨-𝑭𝒂𝒄𝒕𝒐𝒓𝒚: - Fine-tune 100+ LLMs (like LLaMA, Mistral, Falcon) with a zero-code interface. - Preconfigured training templates - Great for data scientists who want model customization without touching training loops. - 𝑮𝒊𝒕𝑯𝒖𝒃 𝒓𝒆𝒑𝒐: https://lnkd.in/ghRb8jgs 4️⃣ ��𝒓𝒂𝒏𝒔𝒇𝒐𝒓𝒎𝒆𝒓 𝑳𝒂𝒃: - An all-in-one desktop app to run and experiment with LLMs locally. - Load open-source models - No setup needed — works out of the box - Perfect for beginners exploring LLM internals - 𝑮𝒊𝒕𝑯𝒖𝒃 𝒓𝒆𝒑𝒐: https://lnkd.in/gY2dUVU9 5️⃣ 𝑳𝒂𝒏𝒈𝒇𝒍𝒐𝒘: - Drag-and-drop interface to launch multi-agent apps with vector DBs and tool support - Graph-based design of chains and agents - Loved by developers building fast MVPs with agentic workflows. - 𝑮𝒊𝒕𝑯𝒖𝒃 𝒓𝒆𝒑𝒐: https://lnkd.in/gAzqK82F 6️⃣ 𝑨𝒖𝒕𝒐𝑨𝒈𝒆𝒏𝒕: - A fully autonomous and zero-code LLM agent framework. - Runs through natural language commands - Best for creating self-healing, goal-driven agent systems. - 𝑮𝒊𝒕𝑯𝒖𝒃 𝒓𝒆𝒑𝒐: https://lnkd.in/gEM5hjdp All of them are open-source, easy to deploy, and great for rapid prototyping. ☑️ 𝐏𝐞𝐫𝐟𝐞𝐜𝐭 𝐟𝐨𝐫: AI researchers testing agentic workflows Builders exploring LLMOps without deep infra setup Product teams needing fast experimentation 𝑾𝒂𝒏𝒕 𝒕𝒐 𝒄𝒐𝒏𝒏𝒆𝒄𝒕 𝒘𝒊𝒕𝒉 𝒎𝒆? 𝘍𝒊𝒏𝒅 𝒎𝒆 𝒉𝒆𝒓𝒆 --> https://lnkd.in/dTK-FtG3 Follow Shreya Khandelwal for more such content. ************************************************************************ #LargeLanguageModels #ArtificialIntelligence #GenerativeAI #LLM #MachineLearning #AI #DataScience #RAG #GenAI #AIagents #AgenticAI #OpenSource #RAGFlow #MLOps #VectorDB #PromptEngineering

  • View profile for Shubham Saboo

    Senior AI Product Manager @ Google | Awesome LLM Apps (#1 AI Agents GitHub repo with 104k+ stars) | 3x AI Author | Community of 350k+ AI developers | Views are my Own

    87,656 followers

    4 opensource LLM fine-tuning libraries you need to know about as an AI engineer. Fine-tuning used to require enterprise budgets and PhD-level expertise. Not anymore. Here are 4 libraries that makes LLM fine-tuning accessible for all: 1. Unsloth AI • 2x faster training with 80% VRAM reduction • Custom Triton kernels with manual backprop engine • Supports 4-bit to 16-bit quantization • Works on NVIDIA GPUs from V100 onwards (CUDA 7.0+) 2. HuggingFace TRL (Transformer Reinforcement Learning) • Specialized trainers: SFTTrainer, DPOTrainer, RewardTrainer • Full integration with Transformers and PEFT ecosystems • Scalable via Accelerate (single GPU to multi-node) • CLI interface for quick experimentation 3. Axolotl • Single YAML config for entire pipeline (preprocess → train → inference) • Advanced optimizations: Flash Attention, multipacking, sequence parallelism • Multi-GPU/multi-node support (FSDP, DeepSpeed, Ray) • Flexible data loading (local, HuggingFace, cloud storage) 4. LlamaFactory • Supports 100+ models including LLaMA, Mistral, Qwen, DeepSeek • Multiple training methods: full fine-tuning, LoRA, QLoRA (2/4/8-bit), DPO, PPO • Web UI and CLI interface - no coding required • Built-in experiment tracking (TensorBoard, Wandb, MLflow) These tools have removed the barriers. The question isn't whether you can fine-tune - it's which approach fits your use case. The best part? They're all 100% opensource. Which library are you planning to try first? Link to the GitHub Repos in the comments.

  • View profile for Paolo Perrone

    No BS AI/ML Content | ML Engineer with a Plot Twist 🥷100M+ Views 📝

    125,253 followers

    10 Open Source AI Tools Every Engineer Should Know After 3 months of testing, here's what survived my workflow: 1️⃣ Talkd.ai — JSON to AI Agent in Minutes Forget complex backends. Define agent behavior in YAML. Built a PDF analyzer agent during lunch break. Perfect for "I need this working by EOD" situations. 🔗 https://talkd.ai 2️⃣ Marimo — Python Notebooks That Don't Suck Reactive cells. Built-in versioning. No more kernel panic at 3am. Finally, notebooks I can push to production without shame. My data science team switched in a week. 🔗 https://lnkd.in/gxwrtBJc 3️⃣ Unsloth AI — Fine-tune LLMs on Your Gaming GPU Llama 3 fine-tuning on a single 3090. No cloud bills. 2x faster than standard methods. Your GPU won't melt. Democratizing model customization for real. 🔗 https://lnkd.in/gJZtH4Y4 4️⃣ HackingBuddyGPT — Ethical Hacking Assistant Fully offline. Generates payloads. Runs recon scripts. Because your pentesting data shouldn't touch the cloud. Red teamers, this one's for you. 🔗 https://lnkd.in/gRrJ-Zwh 5️⃣ Giskard — Unit Tests for AI Models Catch hallucinations before users do. Test for bias, toxicity, and edge cases systematically. Saved me from shipping a model that thought all CEOs were male. 🔗 https://lnkd.in/g3QhG9FB 6️⃣ OpenWebUI — Self-Hosted ChatGPT Runs Llama, Mistral, or Claude locally. Zero API costs. Tool calling, memory, custom personas included. Privacy-first teams love this one. 🔗 https://lnkd.in/gk3t65RG 7️⃣ Axolotl — YAML-Driven Fine-Tuning One config file. Multiple training strategies. QLORA, PEFT, LORA - pick your poison. Fine-tuning without the PhD in configuration. 🔗 https://lnkd.in/gu6pJxWk 8️⃣ FastRAG — RAG in 5 Minutes Flat No Pinecone. No LangChain bloat. Just local RAG that works. Point it at PDFs or websites. Start querying. Built for prototypes that become production. 🔗 https://lnkd.in/gNrG6HyE 9️⃣ Nav2 — Robot Navigation That Actually Ships ROS 2 based. Real-time obstacle avoidance. Multi-robot coordination out of the box. If you're building robots, you need this. 🔗 https://lnkd.in/gYiqsiTJ 🔟 MindsDB — ML Inside Your Database Train models with SQL: `SELECT predict(sales) FROM data` No export/import dance. No separate ML pipeline. Your DBA will either love or hate you. 🔗 https://lnkd.in/gYiqsiTJ My Quick Match Guide: Need fast prototypes? → Talkd.ai + FastRAG Building data apps? → Marimo + MindsDB Shipping to production? → Giskard + Axolotl Privacy critical? → OpenWebUI + HackingBuddyGPT The best part? Clone → Install → Ship. No waitlists. No API keys. No surprises. Open source AI isn't just catching up. It's setting the pace. What open source AI tool saved your project this week? ♻️ Repost to help a developer discover their next favorite tool

  • View profile for Karan Chandra Dey

    UI/UX & Creative Technology Designer | AI Prototyping, Implementation & Healthcare Innovation

    2,288 followers

    Excited to announce my new (free!) white paper: “Self-Improving LLM Architectures with Open Source” – the definitive guide to building AI systems that continuously learn and adapt. If you’re curious how Large Language Models can critique, refine, and upgrade themselves in real-time using fully open source tools, this is the resource you’ve been waiting for. I’ve put together a comprehensive deep dive on: Foundation Models (Llama 3, Mistral, Google Gemma, Falcon, MPT, etc.): How to pick the right LLM as your base and unlock reliable instruction-following and reasoning capabilities. Orchestration & Workflow (LangChain, LangGraph, AutoGen): Turn your model into a self-improving machine with step-by-step self-critiques and automated revisions. Knowledge Storage (ChromaDB, Qdrant, Weaviate, Neo4j): Seamlessly integrate vector and graph databases to store semantic memories and advanced knowledge relationships. Self-Critique & Reasoning (Chain-of-Thought, Reflexion, Constitutional AI): Empower LLMs to identify errors, refine outputs, and tackle complex reasoning by exploring multiple solution paths. Evaluation & Feedback (LangSmith Evals, RAGAS, W&B): Monitor and measure performance continuously to guide the next cycle of improvements. ML Algorithms & Fine-Tuning (PPO, DPO, LoRA, QLoRA): Transform feedback into targeted model updates for faster, more efficient improvements—without catastrophic forgetting. Bias Amplification: Discover open source strategies for preventing unwanted biases from creeping in as your model continues to adapt. In this white paper, you’ll learn how to: Architect a complete self-improvement workflow, from data ingestion to iterative fine-tuning. Deploy at scale with optimized serving (vLLM, Triton, TGI) to handle real-world production needs. Maintain alignment with human values and ensure continuous oversight to avoid rogue outputs. Ready to build the next generation of AI? Download the white paper for free and see how these open source frameworks come together to power unstoppable, ever-learning LLMs. Drop a comment below or send me a DM for the link! Let’s shape the future of AI—together. #AI #LLM #OpenSource #SelfImproving #MachineLearning #LangChain #Orchestration #VectorDatabases #GraphDatabases #SelfCritique #BiasMitigation #Innovation #aiagents

  • View profile for Avthar Sewrathan 🤖

    AI Product Leader & Educator | Obsessed with AI Coding

    4,380 followers

    Stop paying the OpenAI tax. The best AI developer tools are actually open-source, free to use, and give you full control over your data and privacy. While proprietary AI dominated early headlines, the true revolution is happening in open source - where a flourishing ecosystem of smarter models and easy to use developer tools is making advanced AI accessible to everyone. After speaking with hundreds of developers, my Timescale colleague Matvey Arye and I have curated the 'Easy Mode' Open Source AI stack - the most developer-friendly tools that work seamlessly together to help you build AI apps: 🦙 LLMs: Open source models like Llama 3 from AI at Meta and Qwen 2.5 are matching Claude and GPT’s performance on many benchmarks come with more data privacy guarantees. ↗️ Embeddings: Modern embedding models like Jina AI, BGE from BAII, and Nomic AI power help devs power accurate search and RAG without paying per token or dependence on third party APIs. 🦙 Model access and deployment: Ollama enables developers to access and deploy dozens of state of the art open-source models with just one command – no team of PhDs required. 🐘 Data and retrieval: PostgreSQL -- The world's most trusted database now handles AI workloads better than specialized vector DBs, thanks to extensions like pgvector and pgai. ⚡ Backend: FastAPI is the fastest way to build production-ready AI backends that actually scale. 🔺 Frontend: NextJS enables devs to build beautiful AI UIs with the framework that handles streaming, caching, and real-time updates out of the box. How’s your experience been with these tools? What did I miss? Let me know in the comments. We talk more about open-source AI and this easy mode stack for devs to build AI apps in the blog in the comments. #opensource #llama3 #ollama #nextjs #postgresql #opensourceai

  • View profile for Jiachen (Amber) Liu

    Build AI Co-Scientist for Everyone | Meta MSL | CS PhD @ UMich - Systems for LLM | SJTU

    7,462 followers

    Can AI agent run AI research experiments? Not yet, we found LLMs lack the practical knowledge of the research engineering layer — how to configure Megatron for distributed training, how to run RLHF with TRL, how to quantize models without breaking them.. This library documents skills for every ML framework, every tool, and open-sourced all of them. Now it's 83 skills across 20 categories: → Distributed training (DeepSpeed.ai, FSDP, Megatron-Core)  → Inference & optimization (vLLM, TensorRT-LLM, SGLang)  → Post-training & RLHF (VeRL , OpenRLHF, TRL) → Agents & RAG (LangChain, LlamaIndex ,FAISS , Qdrant)  → Writing AI research papers (LaTeX templates, citation verification) One command installs them into any coding agent — Claude Code, Codex, Gemini CLI, Cursor: 𝚗̲𝚙̲𝚡̲ ̲@̲𝚘̲𝚛̲𝚌̲𝚑̲𝚎̲𝚜̲𝚝̲𝚛̲𝚊̲–̲𝚛̲𝚎̲𝚜̲𝚎̲𝚊̲𝚛̲𝚌̲𝚑̲/̲𝚊̲𝚒̲–̲𝚛̲𝚎̲𝚜̲𝚎̲𝚊̲𝚛̲𝚌̲𝚑̲–̲𝚜̲𝚔̲𝚒̲𝚕̲𝚕̲𝚜̲ Do AI research by prompting with your hypothesis, not debugging infrastructure. https://lnkd.in/grXey8Fg

  • View profile for Ravi Shankar

    Engineering Manager, ML

    33,117 followers

    As AI evolves — from classical ML to deep learning to LLMs — teams are converging on a powerful open-source compute stack: ✅ Kubernetes + Ray + PyTorch + vLLM Each layer plays a distinct role: 💻 Kubernetes provisions and manages compute. 💻 Ray handles distributed scheduling, autoscaling, and failure recovery. 💻 PyTorch powers model training and inference. 💻 vLLM accelerates transformer-based inference with advanced optimizations (e.g., paged attention, speculative decoding). The blog mentions the following use cases: 💻 Pinterest cut dataset iteration time 6× and GPU costs by 25%. 💻 Uber 2–3× training throughput for LLaMA-2-70B. 💻 Roblox achieved 58% cost reduction on large-scale inference. 💻 Even post-training RLHF stacks (e.g., VeRL, OpenRLHF) consistently rely on this combination, deployed across Kubernetes and SLURM. This stack isn't just a trend — it's quickly becoming the standard for scalable, future-proof AI infrastructure. Full post: https://lnkd.in/gByKA3GP

Explore categories