c5c350cf9f9b576cc9de939e4dc308404eb48852,catalyst/dl/experiments/experiment.py,ConfigExperiment,get_optimizer_and_model,#ConfigExperiment#Any#Any#,272
 
Before Change
        if fp16:
            utils.assert_fp16_available()
            from apex import amp
            if fp16_opt_level not in {"O1", "O2", "O3"}:
                raise ValueError("fp16 mode must be one of O1, O2, O3")
            model, optimizer = amp.initialize(
                model, optimizer, opt_level=fp16_opt_level)
        elif torch.cuda.device_count() > 1:
            model = torch.nn.DataParallel(model)
After Change
        optimizer_params = \
            self.stages_config[stage].get("optimizer_params", {})
        distributed_params = \
            self.stages_config[stage].get("distributed_params", {})
        optimizer = self._get_optimizer(
            model_params=model_params, **optimizer_params)

In pattern: SUPERPATTERN
Frequency: 3
Non-data size: 4
Instances
 Project Name: catalyst-team/catalyst
 Commit Name: c5c350cf9f9b576cc9de939e4dc308404eb48852
 Time: 2019-05-28
 Author: scitator@gmail.com
 File Name: catalyst/dl/experiments/experiment.py
 Class Name: ConfigExperiment
 Method Name: get_optimizer_and_model
 Project Name: asyml/texar
 Commit Name: f1b7e12c77298fdce190e7270af61c1a734fc8e1
 Time: 2018-03-30
 Author: zhitinghu@gmail.com
 File Name: texar/core/optimization.py
 Class Name: 
 Method Name: get_train_op
 Project Name: tensorlayer/tensorlayer
 Commit Name: c7e19808f4956666ae23f69364642108a4fe2513
 Time: 2019-04-04
 Author: ivbensekin@gmail.com
 File Name: tensorlayer/layers/core.py
 Class Name: Layer
 Method Name: __init__