Why Machine Learning Models Should be Smaller in Size?
Recently, analysts at NVIDIA declared MegatronLM, a huge transformer model with 8.3 billion parameters (multiple times bigger than BERT) that accomplished cutting-edge performance on a variety of language tasks. There are numerous instances of monstrous
2020-12-22
- READ MORE