Skip to content

January 25, 2023

January 25, 2023

Present

Wesley Brewer, Gregor von Laszewski, Gregg Barrett, Geoffrey Fox, Juri Papay, Murali Emani, Tom Gibbs, Mallikarjun Shankar, Aristeidis Tsaris,

Apologies

Jeyan Thiyagalingam, Tony Hey

Tentative Agenda

  • Any new members
  • Increasing academic involvement in MLCommons: (Continued)
  • FAIR metadata
  • Discussion of new Benchmarks (Continued)
  • AOB

Current Benchmarks

  • Gregor reported a bug in Cloudmask and that OSMI-Bench was running at Virginia.
  • There was discussion of current benchmarks on Cerebras (Argonne and Edinburgh) and NVIDIA including Hopper; Tom Gibbs noted the relevance of combining Grace CPU with Hopper
  • We discussed SDSC data repository which is set up like an s3 object store
  • Limited number of people can upload; Juri from RAL and Christine plus Kevin Coakley, kcoakley@sdsc.edu from SDSC
  • Later we discussed status of benchmarks with David Kanter and he arranged that Peter Mattson ask governing board of MLC to ap[prove our release
  • We discussed a paper withitems such as
  • Misdesign of disk systems
  • Computer systems AI read

FAIR metadata

Draft White Paper from Christine

MLCommons Science Concept Paper

  1. Opportunities to Improve MLCommons Benchmark Data
  2. Low Effort
    1. Create website/github pages for each dataset with basic information
      1) Add license designations (CC) to datasets
      2) Register MLCommons Science benchmark datasets in other repos, e.g. OpenML
      3) Create a DOI for each dataset and provide a suggested citation
    2. Setup collection for the datasets at one of our institutions
      1) Register MLCommons Science benchmarks as a data collection in R3Data
  3. Medium Effort (to be supported under existing grants?)
    1. Add Schema.org descriptors to MLCommons Science dataset pages
    2. Initial work to identify gaps in schema.org
    3. Assess opportunities for capturing power consumption in results
    4. Provide a notebook that shows each dataset being used
  4. Requires Additional Funding
    1. Improve benchmark data metadata and other data attributes. Compare benchmarking results pre and post changes
    2. Document gaps in schema.org as potential extension to schema.org for Computer Science
    3. Modify benchmarks to track for reproducibility variance
  5. Making MLCommons Data AI ready
  6. Making Benchmarks FAIR
  7. Potential Best Practices
    1. At dataset intake obtain metadata and provenance.
    2. Document vocabularies or ontologies used in the creation of the dataset.
    3. Work with provider to obtain a DOI (ideally through their home institution).
    4. Determine repos/resources where the dataset will be listed.
    5. Determine and include the CC license type/
  8. Checklist
  9. Model Cards
  10. Steps We’ve Taken to Improve MLCommons Science Benchmarks
  11. Opportunities for MLCommons Writ Large to Adopt as Community Practices