Skip to content

Nghiauet/Optimize_Transformer_Moder_Experiment

Repository files navigation

Optimize_Transformer_Moder_Experiment

Experiment difference optimization method image The model experiment is the intents predictions use BERT base model fine tune tin CLINC150 dataset ( have 150 intents labels in 10 domains) Techniques to speed up the predictions and reduce the memory footprint include

  • Quantization
  • knowledge distillation
  • quantization training
  • pruning
  • graph optimization ( with ONNX and ORT)

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published