-
Notifications
You must be signed in to change notification settings - Fork 38
WIP: Add initial information for Agent #114
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Open
nickchase
wants to merge
3
commits into
lfai:main
Choose a base branch
from
nickchase:main
base: main
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Open
Changes from 2 commits
Commits
Show all changes
3 commits
Select commit
Hold shift + click to select a range
File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,92 @@ | ||
| # Introduction and instructions | ||
|
|
||
| *Current version: 0.1 (Please increment when making changes)* | ||
|
|
||
| We are creating an agent that will update the Model Openness Tool. | ||
| We're starting by running the prompt manually to make sure that it works. | ||
|
|
||
| To test, run these prompts in order: | ||
|
|
||
| 1. Setup prompt | ||
| 2. Instructions prompt | ||
| 3. Model request prompt | ||
|
|
||
| ## Setup prompt | ||
|
|
||
| ```yaml | ||
| I'm trying to update information about various open source LLMs for the Model Openness Tool. Read this page, including the tooltips: https://mot-dev.isitopen.ai/model/evaluate | ||
|
|
||
| This page ultimately populates a YAML document, such as: | ||
|
|
||
| framework: | ||
| name: 'Model Openness Framework' | ||
| version: '1.0' | ||
| date: '2024-12-15' | ||
| release: | ||
| name: DeepSeek-R1 | ||
| version: 671B | ||
| date: '2025-01-20' | ||
| license: { } | ||
| type: language | ||
| architecture: '' | ||
| origin: '' | ||
| producer: 'DeepSeek AI' | ||
| contact: '' | ||
| repository: 'https://github.com/deepseek-ai/DeepSeek-R1' | ||
| huggingface: 'https://huggingface.co/deepseek-ai/DeepSeek-R1' | ||
| components: | ||
| - | ||
| name: 'Model architecture' | ||
| description: "Well commented code for the model's architecture" | ||
| license: MIT | ||
| - | ||
| name: 'Inference code' | ||
| description: 'Code used for running the model to make predictions' | ||
| license: MIT | ||
| - | ||
| name: 'Supporting libraries and tools' | ||
| description: "Libraries and tools used in the model's development" | ||
| license: MIT | ||
| - | ||
| name: 'Model parameters (Final)' | ||
| description: 'Trained model parameters, weights and biases' | ||
| license: MIT | ||
| - | ||
| name: 'Evaluation data' | ||
| description: 'Data used for evaluating the model' | ||
| license: unlicensed | ||
| - | ||
| name: 'Model metadata' | ||
| description: 'Any model metadata including training configuration and optimizer states' | ||
| license: MIT | ||
| - | ||
| name: 'Model card' | ||
| description: 'Model details including performance metrics, intended use, and limitations' | ||
| license: unlicensed | ||
| - | ||
| name: 'Technical report' | ||
| description: 'Technical report detailing capabilities and usage instructions for the model' | ||
| license: unlicensed | ||
| - | ||
| name: 'Research paper' | ||
| description: 'Research paper detailing the development and capabilities of the model' | ||
| license: 'arXiv.org perpetual non-exclusive license 1.0' | ||
| - | ||
| name: 'Evaluation results' | ||
| description: 'The results from evaluating the model' | ||
| license: unlicensed | ||
|
|
||
| Analyze the form and correlate the YAML properties with the form fields. | ||
| ``` | ||
|
|
||
| ## Instructions prompt | ||
|
|
||
| ```yaml | ||
| Excellent. Now I want you to analyze an open source model. Do any research you have to answer all of these questions about the model and fill out the YAML document. If you don't know the answer, look it up online. DO NOT MAKE ANYTHING UP. Output the full YAML document, then follow up with an explanation of where you found each piece of information. This task is very important to my career. | ||
|
|
||
| Do you understand? | ||
| ``` | ||
|
|
||
| ## Model request prompt | ||
|
|
||
| Here just give the LLM the name of a model, such as `DeepSeek-R1` and it will output the YAML document. | ||
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,104 @@ | ||
| run-number: 1 | ||
| tester: Nick Chase | ||
| prompt-version: 0.1 | ||
| model-used: OpenAI ChatGPT 04-mini | ||
| model-analyzed: DeepSeek-V3 | ||
| results-md: | | ||
|
|
||
| ```yaml | ||
| framework: | ||
| name: "Model Openness Framework" | ||
| version: "1.0" | ||
| date: "2025-05-25" | ||
|
|
||
| release: | ||
| name: "DeepSeek-V3" | ||
| version: "671B" | ||
| date: "2024-12-27" | ||
| license: "MIT" | ||
| type: "language" | ||
| architecture: "Mixture-of-Experts (MoE) with Multi-head Latent Attention (MLA) and DeepSeekMoE" | ||
| origin: "research" | ||
| producer: "DeepSeek AI" | ||
| contact: "service@deepseek.com" | ||
| repository: "https://github.com/deepseek-ai/DeepSeek-V3" | ||
| huggingface: "https://huggingface.co/deepseek-ai/DeepSeek-V3" | ||
| components: | ||
| - | ||
| name: "Model architecture" | ||
| description: "Well commented code for the model’s architecture including MLA and DeepSeekMoE modules" | ||
| license: "MIT" | ||
| - | ||
| name: "Inference code" | ||
| description: "Code used for running the model to make predictions" | ||
| license: "MIT" | ||
| - | ||
| name: "Supporting libraries and tools" | ||
| description: "Libraries and tools used for inference and deployment (e.g., SGLang, LMDeploy, TensorRT-LLM, vLLM, LightLLM)" | ||
| license: "MIT" | ||
| - | ||
| name: "Model parameters (Final)" | ||
| description: "Trained model parameters (weights and biases) for DeepSeek-V3" | ||
| license: "LICENSE-MODEL" | ||
| - | ||
| name: "Evaluation data" | ||
| description: "Data used for evaluating the model across benchmarks (e.g., MMLU, BBH, etc.)" | ||
| license: "unlicensed" | ||
| - | ||
| name: "Model metadata" | ||
| description: "Training configuration, optimizer state, and multi-token prediction module weights" | ||
| license: "MIT" | ||
| - | ||
| name: "Model card" | ||
| description: "High-level details including capabilities, limitations, and intended uses" | ||
| license: "unlicensed" | ||
| - | ||
| name: "Technical report" | ||
| description: "DeepSeek-V3 technical report published on arXiv" | ||
| license: "arXiv.org perpetual non-exclusive license 1.0" | ||
| - | ||
| name: "Research paper" | ||
| description: "Publication detailing the development of DeepSeek-V3" | ||
| license: "arXiv.org perpetual non-exclusive license 1.0" | ||
| - | ||
| name: "Evaluation results" | ||
| description: "Benchmark results and performance metrics" | ||
| license: "unlicensed" | ||
| ``` | ||
|
|
||
| **Sources and Mapping:** | ||
|
|
||
| - **framework.date** (the date of filling the form): set to today’s date, May 25 2025. | ||
|
|
||
| - **release.name** “DeepSeek-V3” and **version** “671B” come from the GitHub model summary, which states 671B total parameters [GitHub](https://github.com/deepseek-ai/DeepSeek-V3). | ||
|
|
||
| - **release.date** “2024-12-27” is the publication date of the DeepSeek-V3 technical report on arXiv [arXiv](https://arxiv.org/abs/2412.19437?utm_source=chatgpt.com). | ||
|
|
||
| - **release.license** “MIT” is pulled from the repository’s code license listing on GitHub [GitHub](https://github.com/deepseek-ai/DeepSeek-V3). | ||
|
|
||
| - **release.type** “language” is the model’s category as an LLM. | ||
|
|
||
| - **release.architecture** is described in the repo introduction, specifying MoE with MLA and DeepSeekMoE [GitHub](https://github.com/deepseek-ai/DeepSeek-V3). | ||
|
|
||
| - **release.origin** is marked “research,” reflecting that DeepSeek-V3 was developed by DeepSeek AI’s research team. | ||
|
|
||
| - **release.producer** “DeepSeek AI” is the GitHub organization and company behind the model [GitHub](https://github.com/deepseek-ai/DeepSeek-V3). | ||
|
|
||
| - **release.contact** “service@deepseek.com” is cited from the Hugging Face model page’s contact section for DeepSeek-V3-0324 [Hugging Face](https://huggingface.co/deepseek-ai/DeepSeek-V3-0324?utm_source=chatgpt.com). | ||
|
|
||
| - **release.repository** is the GitHub URL for DeepSeek-V3 [GitHub](https://github.com/deepseek-ai/DeepSeek-V3). | ||
|
|
||
| - **release.huggingface** is taken from the GitHub downloads table linking to Hugging Face [GitHub](https://github.com/deepseek-ai/DeepSeek-V3). | ||
|
|
||
| - **Components & licenses**: | ||
|
|
||
| - Code components (architecture, inference code, supporting libraries, metadata) carry the MIT license as shown in the repo’s LICENSE-CODE file [GitHub](https://github.com/deepseek-ai/DeepSeek-V3). | ||
|
|
||
| - **Model parameters** use the custom “LICENSE-MODEL” defined in the repository’s LICENSE-MODEL file [Hugging Face](https://huggingface.co/deepseek-ai/DeepSeek-V3/blob/main/LICENSE-MODEL?utm_source=chatgpt.com). | ||
|
|
||
| - **Evaluation data**, **model card**, and **evaluation results** are marked “unlicensed” following the example pattern when no explicit license is provided. | ||
|
|
||
| - **Technical report** and **research paper** both fall under ArXiv’s perpetual non-exclusive license, as indicated on the arXiv page [arXiv](https://arxiv.org/abs/2412.19437?utm_source=chatgpt.com). | ||
|
|
||
| comments: | | ||
| This is reasonably close but doesn't include a Global license. It also mistakes the license file name for the license name. |
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't think you really need that. This is what github does for us. :-)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
That's true, but I an looking for something that can be referenced in the results document. So we can reference a commit id if you want, but we need something.