-
AI model development
Encord is the complete data engine for AI model development. It provides tools and workflows for advanced computer vision teams to streamline their labeling and workflow management, clean and curate data, validate label quality, and evaluate model performance.05 Apr 2024Readmore -
LLM playground
10 Mar 2024Readmore -
AI evaluation
Atla provides frontier AI evaluation models to evaluate generative AI, find and fix AI mistakes at scale, and build more reliable GenAI applications. It offers an LLM-as-a-Judge to test and evaluate prompts and model versions. Atla's Selene models provide precise judgments on AI app performance, running evals with accurate LLM Judges. They offer solutions optimized for speed and industry-leading accuracy, customizable to specific use cases with accurate scores and actionable critiques.11 Mar 2025Readmore -
Visual AI
Voxel51 is a company focused on making visual AI a reality. They provide tools and resources, like FiftyOne, to help visual AI builders curate better datasets and build better models quickly and efficiently. Their platform enables users to analyze, curate, and evaluate multimodal datasets to improve model performance, identify failure modes, biases, and data gaps.27 Jul 2024Readmore -
Multimodal models
Non finito is a platform designed for evaluating multimodal models, with a focus on making it easy to run and share evaluations. It aims to provide tools specifically tailored for multimodal models, which are often overlooked by tools primarily focused on language models (LLMs). The platform emphasizes easy comparison of models and public sharing of evaluations.27 Apr 2024Readmore -
LLM comparison
Model Royale is a website that allows users to compare different Large Language Models (LLMs) by using the same prompt and analyzing their time to respond, tokens used, and the quality of the response. It helps users pick the LLM that best fits their specific needs.07 Dec 2024Readmore