LABORE KODJO AGBETSIASSIstarlabman.hashnode.dev·Dec 4, 2024My 30-60-90 Day Plan as a Developer Advocate at Arbitrum FoundationJoining Arbitrum Foundation as a Developer Advocate is an exciting opportunity to drive the adoption of Stylus, a key technology in the Arbitrum ecosystem, while building a dynamic and engaged developer community. Here’s my structured plan for the fi...Discuss·1 likeEducate developers
Tom X NguyenforDwarves Foundation's Team Blogdwarvesf.hashnode.dev·Oct 16, 2024Evaluation Guidelines for LLM ApplicationsOverview Evaluation is a hard part of building an RAG system, especially for application-integrated LLM solving your business problem. This guide outlines a clear, step-by-step approach to effectively evaluating and optimizing the integration of a th...DiscussEvaluation
Tom X NguyenforDwarves Foundation's Team Blogdwarvesf.hashnode.dev·Oct 16, 2024LLM as a JudgeWith the robust growth of LLM models currently , there is a new method is used to evaluate the performance of large language models (LLMs): LLM-as-a-Judge, also known as LLM-evaluators. This approach take adavantages of other advanced language models...DiscussEvaluation
Kavyaportkey-llm-elo-rating.hashnode.dev·Oct 14, 2024⭐️ Decoding OpenAI EvalsLearn how to use the eval framework to evaluate models & prompts to optimise LLM systems for the best outputs. Rohit Agarwal May 10, 2023 ∙ 9 min Conversation on Twitter There's been a lot of buzz around model evaluations since OpenAI open-sourced t...Discussopenai
Manish Singh PariharforFutureSmart AI Blogblog.futuresmart.ai·Oct 11, 2024A Beginner's Guide to Evaluating RAG Systems with LangSmithRetrieval-Augmented Generation (RAG) has become a cornerstone technique in leveraging large language models (LLMs) with proprietary or domain-specific data. This blog post will guide you through the process of integrating Langsmith evaluations into y...Discuss·2 likes·332 readsAI
Muhammad Fahad Bashirmfahadbashir.hashnode.dev·Sep 21, 2024Evaluating Classification Models: A Comprehensive Guide to Key MetricsEvaluating the performance of a model is a crucial step in any machine-learning project. Evaluation metrics help us assess how well our model performs, whether we can rely on its predictions, and how it will behave with new, unseen data. Without prop...DiscussMachine Learning
arman takmazyanarman-dev-blog.hashnode.dev·Sep 4, 2024Arithmetic Expressions Evaluator: Part 1Previous Posts: Creating TuffScript: Exploring Esoteric Programming Languages and Their Foundations Mastering Formal Grammars: An Introduction to the Chomsky Hierarchy Precedence and Associativity in Grammar Rules Mastering Lexical and Syntax Analysi...Discuss·35 readslexer
Pablo Salvador Lopezpabloaicorner.hashnode.dev·Aug 24, 2024Key Technical Challenges while Transitioning GenAI Applications to ProductionWhether you're leading the development of Generative AI apps in a startup or working in a Fortune 500 company, you'll encounter these questions at certain points in your production journey... "Can my application scale to accommodate a 1000-fold incre...Discuss·63 readsGenAI in Production 🧠genai
Amit Tyagiwww.cognitive-quest.com·Jul 21, 2024Red Teaming for GenAI ApplicationsWhat is Red Teaming? In today’s rapidly evolving digital landscape, ensuring the safety and security of generative applications has become a paramount concern. Traditionally, red teaming involves a group of security professionals, known as the red te...Discuss·49 readsResponsible AIredteaming
Yashraj Poddaryashrajp.hashnode.dev·Jan 28, 2024Measuring LLM capabilities?How LLMs are compared? There are a few main ways researchers benchmark the performance of huge language models (LLMs) in natural language processing. A common tactic is testing the models on standard datasets used by lots of NLP folks. These dataset...Discussllm