Selecting a Model Configuration:
- Navigate to the "Generative AI Model" section.
- Choose a predefined model from the "Generative AI Model" dropdown. This will automatically populate the model parameters for you and enable the "Training Dataset Size" dropdown.
- If you prefer, you can also manually adjust or enter the model parameters by clicking on "(show details)" next to the "Generative AI Model" header.
Choosing a Training Dataset Size:
- After selecting your model example and/or editing the model parameters, set the training dataset size in the "Training Dataset Size" field. This represents the number of words/word pieces (AKA tokens) in your training dataset.
Setting Batch Size or Opting for Inference Only:
- Decide if you are training the model or only performing inferences.
- If training, select the desired "Batch Size". The sizer will recommend a number of micro-batches accordingly.
- If performing inferences only, choose the "Inference Only" option.
Selecting a Server Configuration:
- Go to the "Server Configuration" section.
- Choose a predefined GPU configuration from the dropdown.
- For more advanced settings, click on "(show details)" next to the "Server Configuration" header and you can configure your own GPU. If your GPU does not have Tensor cores set the TFLOPs fields to the same value.
Reviewing Model Sizing Results:
- The sizer will calculate and display the resource requirements in the "Model Sizing" section.
- Details include the number of parameters, maximum GPU memory required when training or recommended GPU memory required when inferencing, and quantity of servers required depending on optimization techniques used and the time to complete the Training Dataset Size with each..
- Refer to the chart below the results for a visual representation.
Tool Tips:
- Hover over question mark icons "(?)" for more information or clarifications on specific fields and values.