Latest Tutorials

Learn about the latest technologies from fellow newline community members!

  • React
  • Angular
  • Vue
  • Svelte
  • NextJS
  • Redux
  • Apollo
  • Storybook
  • D3
  • Testing Library
  • JavaScript
  • TypeScript
  • Node.js
  • Deno
  • Rust
  • Python
  • GraphQL
  • React
  • Angular
  • Vue
  • Svelte
  • NextJS
  • Redux
  • Apollo
  • Storybook
  • D3
  • Testing Library
  • JavaScript
  • TypeScript
  • Node.js
  • Deno
  • Rust
  • Python
  • GraphQL
    NEW

    Claude Skills and Subagents Reduce Prompt Bloat

    Watch: How I Built an AI Council with Claude Code Subagents by Mark Kashef Claude Skills and Subagents offer a structured approach to reducing prompt bloat by enabling reusable, context-aware instructions that optimize token usage and improve context management. This section breaks down their advantages, implementation metrics, and real-world applications to help developers evaluate their suitability for different workflows. Claude Skills and Subagents stand out from traditional prompt reduction methods like static templates or function calls by offering dynamic, modular execution . Skills act as lightweight, reusable components that load only when needed, reducing token overhead by up to 40% in code-generation tasks. Subagents, on the other hand, handle complex workflows by delegating tasks to specialized agents, avoiding context bloat through isolated memory management. A comparison with older methods reveals:
    Thumbnail Image of Tutorial Claude Skills and Subagents Reduce Prompt Bloat
      NEW

      Using process rewards to train LLMs for better search reasoning

      Training large language models (LLMs) to improve search reasoning often involves process rewards -a technique that evaluates and reinforces step-by-step reasoning rather than just final answers. This approach enhances accuracy in complex tasks like math problems, logical deductions, and multi-step queries. Below is a structured overview of key techniques, their benefits, and implementation considerations. For foundational details on how process rewards differ from outcome-based methods, see the Why Process Rewards Matter section. ReST-MCTS stands out for combining Monte Carlo Tree Search (MCTS) with process rewards, enabling LLMs to explore reasoning paths more effectively. This method excels in tasks requiring iterative problem-solving, such as algebraic proofs or code debugging. For implementation guidelines on frameworks like RAG-Gym and ReST-MCTS , refer to the Practical Implementation Checklist section. Time and effort estimates vary: Basic implementations (e.g., Best-of-N) require minimal setup but offer limited gains. Advanced methods like ReST-MCTS* demand more engineering but yield significant improvements. Difficulty ratings reflect the complexity of integrating tree search algorithms and reward modeling.
      Thumbnail Image of Tutorial Using process rewards to train LLMs for better search reasoning

      I got a job offer, thanks in a big part to your teaching. They sent a test as part of the interview process, and this was a huge help to implement my own Node server.

      This has been a really good investment!

      Advance your career with newline Pro.

      Only $40 per month for unlimited access to over 60+ books, guides and courses!

      Learn More
        NEW

        Mitigating bias in LLM‑based scoring of English language learners

        Mitigating bias in LLM-based scoring for English language learners (ELLs) requires a structured approach to ensure fairness and accuracy. Below is a summary of key strategies, challenges, and outcomes based on recent research. Different LLMs employ varied bias mitigation methods. For example, GPT-4 uses data augmentation to diversify training samples, while BERT relies on bias-aware training to adjust scoring for linguistic diversity. Advanced frameworks like BRIDGE (LLM-based data augmentation) and AutoSCORE (multi-agent scoring systems) show promise in reducing subgroup bias. A comparison of these models reveals: See the Techniques for Mitigating Bias in LLM-Based Scoring section for more details on these frameworks and their implementation.
        Thumbnail Image of Tutorial Mitigating bias in LLM‑based scoring of English language learners
          NEW

          What Is Prompt Chaining and How to Use It

          Prompt chaining is a method where complex tasks are broken into sequential subtasks, each handled by a distinct prompt. This approach ensures context is preserved between steps and allows for structured problem-solving. Below is a breakdown of key aspects, techniques, and applications. Benefits : Challenges :
          Thumbnail Image of Tutorial What Is Prompt Chaining and How to Use It
            NEW

            How to Chain Prompts for Better LLM Flow

            Watch: Let The LLM Write The Prompt 2025 | Design Perfect Prompts for AI Agent | Prompt Mistakes (PART 1/7) by Amine DALY Prompt chaining enhances large language model (LLM) workflows by linking prompts sequentially or in parallel to solve complex tasks. This section breaks down techniques, metrics, and real-world use cases to help you design efficient chains. Prompt chaining methods vary in complexity and use cases. Serial chaining executes prompts one after another, ideal for tasks requiring step-by-step reasoning (e.g., data extraction followed by analysis). Parallel chaining splits tasks into simultaneous prompts, useful for multi-branch decisions or data aggregation. Hybrid approaches combine both for tasks like customer service workflows, where initial triage (parallel) triggers specialized follow-ups (serial).
            Thumbnail Image of Tutorial How to Chain Prompts for Better LLM Flow