Topic: evaluation-framework Goto Github
Some thing interesting about evaluation-framework
Some thing interesting about evaluation-framework
evaluation-framework,The Core Reinforcement Learning library is intended to enable scalable deep reinforcement learning experimentation in a manner extensible to new simulations and new ways for the learning agents to interact with them. The hope is that this makes RL research easier by removing lock-in to particular simulations.The work is released under the follow APRS approval. Initial release of CoRL - Part #1 -Approved on 2022-05-2024 12:08:51 - PA Approval # [AFRL-2022-2455]" Documentation https://act3-ace.github.io/CoRL/
Organization: act3-ace
Home Page: https://www.act3-ace.com/
evaluation-framework,Entity linking evaluation and analysis tool
Organization: ad-freiburg
Home Page: https://elevant.cs.uni-freiburg.de/
evaluation-framework,Evaluation suite for large-scale language models.
Organization: ai21labs
evaluation-framework,This repository contains the implementation of evaluation metrics for recommendation systems. We have compared similarity, candidate generation, rating, ranking metrics performance on 5 different datasets - MovieLens 100k, MovieLens 1m, MovieLens 10m, Amazon Electronics Dataset and Amazon Movies and TV Dataset.
User: aryan-jadon
evaluation-framework,OD-test: A Less Biased Evaluation of Out-of-Distribution (Outlier) Detectors (PyTorch)
User: ashafaei
evaluation-framework,Python SDK for running evaluations on LLM generated responses
Organization: athina-ai
Home Page: https://docs.athina.ai
evaluation-framework,Expressive is a cross-platform expression parsing and evaluation framework. The cross-platform nature is achieved through compiling for .NET Standard so it will run on practically any platform.
User: bijington
evaluation-framework,This repository allows you to evaluate a trained computer vision model and get general information and evaluation metrics with little configuration.
Organization: bmw-innovationlab
evaluation-framework,BIRL: Benchmark on Image Registration methods with Landmark validations
User: borda
Home Page: http://borda.github.io/BIRL
evaluation-framework,Industrial-level evaluation benchmarks for Coding LLMs in the full life-cycle of AI native software developing.企业级代码大模型评测体系,持续开放中
Organization: codefuse-ai
evaluation-framework,The LLM Evaluation Framework
Organization: confident-ai
Home Page: https://docs.confident-ai.com/
evaluation-framework,Framework for Interactive Evaluation of Recommender Systems
User: cowjen01
evaluation-framework,🌾 Universal, customizable and deployable fine-grained evaluation for text generation.
User: davidheineman
Home Page: https://thresh.tools
evaluation-framework,A research library for automating experiments on Deep Graph Networks
User: diningphil
Home Page: https://pydgn.readthedocs.io
evaluation-framework,A suite of experiments for evaluating open-source binary taint trackers.
Organization: dynamic-rabbits
evaluation-framework,A framework for few-shot evaluation of language models.
Organization: eleutherai
Home Page: https://www.eleuther.ai
evaluation-framework,Test and evaluate LLMs, prompts and other configuration, across all the scenarios that matter for your application
Organization: empirical-run
Home Page: https://docs.empirical.run
evaluation-framework,Scalable Meta-Evaluation of LLMs as Evaluators
Organization: gair-nlp
evaluation-framework,LiDAR SLAM comparison and evaluation framework
User: haeyeoni
evaluation-framework,Official repository of RankEval: An Evaluation and Analysis Framework for Learning-to-Rank Solutions.
Organization: hpclab
Home Page: http://rankeval.isti.cnr.it/
evaluation-framework,LightEval is a lightweight LLM evaluation suite that Hugging Face has been using internally with the recently released LLM data processing library datatrove and LLM training library nanotron.
Organization: huggingface
evaluation-framework,Evaluation framework for oncology foundation models (FMs)
Organization: kaiko-ai
Home Page: https://kaiko-ai.github.io/eva/
evaluation-framework,Python client for Kolena's machine learning testing platform
Organization: kolenaio
Home Page: https://docs.kolena.io
evaluation-framework,An Evaluation Framework for Temporal Information Extraction Systems
Organization: liaad
evaluation-framework,An easy-to-use tool for evaluating tracking algorithms on many different benchmarks like OTB and Temple-Color
User: lukaswals
evaluation-framework,Evaluate your biometric verification models literally in seconds.
User: ma7555
evaluation-framework,This is the repository of our article published in RecSys 2019 "Are We Really Making Much Progress? A Worrying Analysis of Recent Neural Recommendation Approaches" and of several follow-up studies.
User: mauriziofd
evaluation-framework,Moonshot - A simple and modular tool to evaluate and red-team any LLM application.
User: moonshot-admin
evaluation-framework,ETUDE (Evaluation Tool for Unstructured Data and Extractions) is a Python-based tool that provides consistent evaluation options across a range of annotation schemata and corpus formats
Organization: musc-tbic
evaluation-framework,Multilingual Large Language Models Evaluation Benchmark
User: nlp-uoregon
evaluation-framework,A high-level Python framework to evaluate the skill of geospatial datasets by comparing candidates to benchmark maps producing agreement maps and metrics.
Organization: noaa-owp
Home Page: https://noaa-owp.github.io/gval/
evaluation-framework,The implementation of the paper "Evaluating Coherence in Dialogue Systems using Entailment"
User: nouhadziri
Home Page: https://arxiv.org/abs/1904.03371
evaluation-framework,The official implementation of the paper "To Generate or Not? Safety-Driven Unlearned Diffusion Models Are Still Easy To Generate Unsafe Images ... For Now". This work introduces one fast and effective attack method to evaluate the harmful-content generation ability of safety-driven unlearned diffusion models.
Organization: optml-group
evaluation-framework,Vectory provides a collection of tools to track and compare embedding versions.
Organization: pentoai
evaluation-framework,Power Flows DMN - Powerful decisions and rules engine
Organization: powerflows
evaluation-framework,Test your prompts, models, and RAGs. Catch regressions and improve prompt quality. LLM evals for OpenAI, Azure, Anthropic, Gemini, Mistral, Llama, Bedrock, Ollama, and other local & private models with CI/CD integration.
Organization: promptfoo
Home Page: https://www.promptfoo.dev/
evaluation-framework,Open-Source Evaluation for GenAI Application Pipelines
Organization: relari-ai
Home Page: https://docs.relari.ai/
evaluation-framework,Implementation of Common Image Evaluation Metrics by Sayed Nadim (sayednadim.github.io). The repo is built based on full reference image quality metrics such as L1, L2, PSNR, SSIM, LPIPS. and feature-level quality metrics such as FID, IS. It can be used for evaluating image denoising, colorization, inpainting, deraining, dehazing etc. where we have access to ground truth.
User: sayednadim
evaluation-framework,Simulator for training and evaluation of Recommender Systems
Organization: sb-ai-lab
Home Page: https://sb-ai-lab.github.io/Sim4Rec/
evaluation-framework,Python-based tools for pre-, post-processing, validating, and curating spike sorting datasets.
Organization: spikeinterface
Home Page: https://spikeinterface.readthedocs.io/en/latest/
evaluation-framework,Evaluation Framework for Dependency Analysis (EFDA)
Organization: srcclr
evaluation-framework,DevQualityEval: An evaluation benchmark 📈 and framework to compare and evolve the quality of code generation of LLMs.
Organization: symflower
evaluation-framework,Train, evaluate, and optimize implicit feedback-based recommender systems.
User: tohtsky
Home Page: https://irspack.readthedocs.io/
evaluation-framework,Metrics to evaluate the quality of responses of your Retrieval Augmented Generation (RAG) applications.
Organization: tonicai
Home Page: https://docs.tonic.ai/validate/
evaluation-framework,Optical Flow Dataset and Benchmark for Visual Crowd Analysis
User: tsenst
evaluation-framework,quica is a tool to run inter coder agreement pipelines in an easy and effective ways. Multiple measures are run and results are collected in a single table than can be easily exported in Latex
User: vinid
evaluation-framework,This is a machine learning framework that enables developers to iterate fast over different ML architecture designs.
User: vishal-keshav
evaluation-framework,Framework to evaluate Trajectory Classification Algorithms
Organization: yupidevs
evaluation-framework,AI Data Management & Evaluation Platform
Organization: zeno-ml
Home Page: https://zenoml.com
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.