2023 June 1: Follow further news from our MLCommons task force on automation and reproducibility
here.
2023 Feb 16: New alpha CK2 GUI to visualize all MLPerf results is available here.
2023 Jan 30: New alpha CK2 GUI to run MLPerf inference is available here.
2022 November: We are very excited to see that our new CK2 automation meta-framework (CM)
was successfully used at the Student Cluster Competition'22
to make it easier to prepare and run the MLPerf inference benchmark just under 1 hour.
If you have 20 minutes, please check this tutorial
to reproduce results yourself ;) !
2022 September: We have released MLCommons CM v1.0.1 -
the next generation of the MLCommons Collective Knowledge framework being developed
by the public workgroup.
We are very glad to see that more than 80% of all performance results and more than 95% of all power results
were automated by the MLCommons CK v2.6.1 in the latest MLPerf inference round thanks to submissions from Qualcomm, Krai, Dell, HPE and Lenovo!
2021 April: We are very excited to join forces with OctoML! Contact Grigori Fursin for more details!
2021 April: We have released CK v2.x under Apache 2.0 license based on the feedback from our users.
2021 March: See our ACM TechTalk about "reproducing 150 Research Papers and Testing Them in the Real World"
on the ACM YouTube channel.
2021 March: The report from the "Workflows Community Summit: Bringing the Scientific Workflows Community Together"
is available in ArXiv.
2021 March: We have published overview of the CK techology in the Philosophical Transactions A, the world's longest-running journal where Newton published: DOI, ArXiv.
2021 February 11: See the announcement for our ACM TechTalk
about reproducing 150 ML and Systems papers and introducing the CK technology.
2020 December: The cTuning foundation is honored to join MLCommons
as a founding member to accelerate machine learning innovation along with 50+ leading companies and universities.
The Collective Knowledge framework (CK)
helps to organize any software project as a database of reusable components
(algorithms, datasets, models, frameworks, scripts, experimental results, papers, etc)
with common automation actions and extensible meta descriptions
based on FAIR principles (findability, accessibility, interoperability and reusability).
The goal is to make it easier for researchers, practitioners and students to reproduce, compare
and build upon techniques from published papers shared in the common CK format,
adopt them in production and reuse best R&D practices.
See how the CK technology helps to automate benchmarking, optimization and design space exploration
of ML Systems and accelerate AI, ML and System innovation: journal article,
ACM tech talk,
cKnowledge.io portal,
and some real-world use cases
from MLPerf, General Motors, Arm, IBM, Amazon, Qualcomm, DELL, the Raspberry Pi foundation and ACM.
Presentations about CK (ACM, General Motors and FOSDEM)
Developing innovative technology (AI, ML, quantum, IoT),
and deploying it in the real world is a very painful, ad-hoc, time consuming and expensive process
due to continuously evolving software, hardware, models, data sets and research techniques.
After struggling with these problems for many years, we started developing the Collective Knowledge framework (CK)
to decompose complex research projects into reusable automation actions
and components with unified APIs, CLI and JSON meta descriptions.
These CK components help to abstract software, hardware, models, data sets, scripts, results
and can be connected into portable CK workflows
while applying powerful DevOps principles.
Such workflows use a portable meta package manager
with a list all dependencies
to automatically adapt to a given platform (detect hardware, software and environment, install missing packages and build/optimize code).
We have spent the last few years testing CK with our great academic and industrial partners
as a playground to implement and share reusable automation actions
and components typical in AI, ML and systems R&D
while agreeing on common APIs and JSON meta description.
We used such components to assemble portable workflows
from reproduced research papers
during the artifact evaluation process
that we have helped to arrange at different ML&systems conferences including ASPLOS, CGO, PPoPP and MLSys.
We then demonstrated that it was possible to use such portable workflows
to automate design space exploration of AI/ML/SW/HW stacks,
automate MLPerf benchmark submissions,
and simplify deployment of ML Systems in production
in the most efficient way (speed, accuracy, energy, costs)
across diverse platforms from data centers to edge devices.
CK framework is basically a common playground that connects researchers and practitioners
to learn how to collaboratively design, benchmark, optimize and validate innovative computational technology
including self-optimizing and bio-inspired computing systems.
Our mission is to organize and systematize all our AI, ML and systems knowledge
in the form of portable workflow, automation actions and reusable artifacts
using our open CK platform
with reproducible papers
and live SOTA scoreboards for crowd-benchmarking.
We continue using CK to support related initiatives including MLPerf, PapersWithCode,
ACM artifact review and badging,
and artfact evaluation.