@@ -59,9 +59,10 @@ class RangeError {};
59
59
60
60
// / Not support Error, such as access GPU memory directly, etc.
61
61
class UnsupportError : public std ::runtime_error {
62
- public:
63
- UnsupportError () : std::runtime_error(" " ){};
64
- UnsupportError (const std::string& message) : std::runtime_error(message){};
62
+ public:
63
+ UnsupportError () : std::runtime_error(" " ) {}
64
+ explicit UnsupportError (const std::string& message)
65
+ : std::runtime_error(message) {}
65
66
};
66
67
67
68
// / This type will map to python's list of float.
@@ -105,7 +106,7 @@ class Matrix {
105
106
DISABLE_COPY (Matrix);
106
107
static Matrix* createByPaddleMatrixPtr (void * sharedPtr);
107
108
108
- public:
109
+ public:
109
110
virtual ~Matrix ();
110
111
111
112
/* *
@@ -231,7 +232,7 @@ class Matrix {
231
232
232
233
bool isGpu () const ;
233
234
234
- private:
235
+ private:
235
236
void * getSharedPtr () const ;
236
237
237
238
MatrixPrivate* m;
@@ -248,7 +249,7 @@ class Vector {
248
249
249
250
void * getSharedPtr ();
250
251
251
- public:
252
+ public:
252
253
~Vector ();
253
254
254
255
// / Create Vector filled with zero.
@@ -310,10 +311,10 @@ class Vector {
310
311
// / __len__ in python
311
312
size_t getSize () const ;
312
313
313
- private:
314
+ private:
314
315
VectorPrivate* m;
315
316
316
- private:
317
+ private:
317
318
friend class Parameter ;
318
319
friend class ParameterOptimizer ;
319
320
friend struct ParameterTraverseCallbackPrivate ;
@@ -325,7 +326,7 @@ class IVector {
325
326
DISABLE_COPY (IVector);
326
327
static IVector* createByPaddleVectorPtr (void * ptr);
327
328
328
- public:
329
+ public:
329
330
// / Create IVector filled with zero
330
331
static IVector* createZero (size_t sz, bool useGpu = isUsingGpu());
331
332
@@ -389,7 +390,7 @@ class IVector {
389
390
// / This method will map to python __len__();
390
391
size_t getSize () const ;
391
392
392
- private:
393
+ private:
393
394
void * getSharedPtr () const ;
394
395
395
396
friend class Arguments ;
@@ -400,11 +401,11 @@ struct ArgumentsPrivate;
400
401
401
402
// / The Arguments is actual a std::vector<paddle::Argument> in paddle.
402
403
class Arguments {
403
- private:
404
+ private:
404
405
Arguments (); // Internal Create.
405
406
DISABLE_COPY (Arguments);
406
407
407
- public:
408
+ public:
408
409
/* *
409
410
* Create a arguments with size.
410
411
* Note that it can be zero.
@@ -475,12 +476,12 @@ class Arguments {
475
476
476
477
float sum () const ;
477
478
478
- private:
479
+ private:
479
480
static Arguments* createByPaddleArgumentVector (void * ptr);
480
481
static Arguments* createByPaddleArgument (const void * ptr);
481
482
void * getInternalArgumentsPtr () const ;
482
483
483
- private:
484
+ private:
484
485
ArgumentsPrivate* m;
485
486
friend class Trainer ;
486
487
friend class GradientMachine ;
@@ -507,18 +508,18 @@ class ParameterConfig {
507
508
static ParameterConfig* createParameterConfigFromParameterPtr (void * ptr);
508
509
void * getRawPtr ();
509
510
510
- public:
511
+ public:
511
512
~ParameterConfig ();
512
513
513
514
/* *
514
515
* return proto buf string.
515
516
*/
516
517
std::string toProtoString () const ;
517
518
518
- private:
519
+ private:
519
520
ParameterConfigPrivate* m;
520
521
521
- private:
522
+ private:
522
523
friend class Parameter ;
523
524
friend class ParameterOptimizer ;
524
525
friend struct ParameterTraverseCallbackPrivate ;
@@ -529,7 +530,7 @@ class OptimizationConfig {
529
530
DISABLE_COPY (OptimizationConfig);
530
531
OptimizationConfig ();
531
532
532
- public:
533
+ public:
533
534
static OptimizationConfig* createFromProtoString (const std::string& str);
534
535
~OptimizationConfig ();
535
536
@@ -538,7 +539,7 @@ class OptimizationConfig {
538
539
*/
539
540
std::string toProtoString ();
540
541
541
- private:
542
+ private:
542
543
OptimizationConfigPrivate* m;
543
544
544
545
friend class TrainerConfig ;
@@ -549,11 +550,11 @@ class OptimizationConfig {
549
550
550
551
struct ParameterPrivate ;
551
552
class Parameter {
552
- private:
553
+ private:
553
554
Parameter ();
554
555
DISABLE_COPY (Parameter);
555
556
556
- public:
557
+ public:
557
558
virtual ~Parameter ();
558
559
559
560
/* *
@@ -580,11 +581,11 @@ class Parameter {
580
581
581
582
size_t getSize () const ;
582
583
583
- private:
584
+ private:
584
585
static Parameter* createFromRawPtr (void * ptr);
585
586
static Parameter* createFromSharedPtr (void * ptr);
586
587
587
- private:
588
+ private:
588
589
ParameterPrivate* m;
589
590
friend class UpdateCallbackWrapper ;
590
591
friend class GradientMachine ;
@@ -598,14 +599,14 @@ struct ModelConfigPrivate;
598
599
* It is used by GradientMachine.
599
600
*/
600
601
class ModelConfig {
601
- private:
602
+ private:
602
603
ModelConfig ();
603
604
DISABLE_COPY (ModelConfig);
604
605
605
- public:
606
+ public:
606
607
virtual ~ModelConfig ();
607
608
608
- private:
609
+ private:
609
610
ModelConfigPrivate* m;
610
611
friend class TrainerConfig ;
611
612
friend struct TrainerConfigPrivate ;
@@ -619,11 +620,11 @@ struct TrainerConfigPrivate;
619
620
* It is used by GradientMachine.
620
621
*/
621
622
class TrainerConfig {
622
- private:
623
+ private:
623
624
TrainerConfig ();
624
625
DISABLE_COPY (TrainerConfig);
625
626
626
- public:
627
+ public:
627
628
virtual ~TrainerConfig ();
628
629
629
630
static TrainerConfig* createFromTrainerConfigFile (
@@ -634,7 +635,7 @@ class TrainerConfig {
634
635
635
636
OptimizationConfig* getOptimizationConfig () const ;
636
637
637
- private:
638
+ private:
638
639
TrainerConfigPrivate* m;
639
640
friend class Trainer ;
640
641
};
@@ -654,7 +655,7 @@ class TrainerConfig {
654
655
* @endcode
655
656
*/
656
657
class UpdateCallback {
657
- public:
658
+ public:
658
659
virtual ~UpdateCallback ();
659
660
virtual void apply (Parameter* p);
660
661
};
@@ -664,14 +665,14 @@ class ParameterTraverseCallback {
664
665
DISABLE_COPY (ParameterTraverseCallback);
665
666
ParameterTraverseCallback ();
666
667
667
- public:
668
+ public:
668
669
~ParameterTraverseCallback ();
669
670
670
671
void apply (const std::vector<Vector*>& vecs,
671
672
const ParameterConfig& config,
672
673
size_t sparseId);
673
674
674
- private:
675
+ private:
675
676
ParameterTraverseCallbackPrivate* m;
676
677
friend class ParameterOptimizer ;
677
678
};
@@ -686,7 +687,7 @@ class ParameterOptimizer {
686
687
DISABLE_COPY (ParameterOptimizer);
687
688
ParameterOptimizer ();
688
689
689
- public:
690
+ public:
690
691
static ParameterOptimizer* create (OptimizationConfig* config);
691
692
692
693
~ParameterOptimizer ();
@@ -710,19 +711,19 @@ class ParameterOptimizer {
710
711
ParameterTraverseCallback* needSpecialTraversal (
711
712
const ParameterConfig& config) const ;
712
713
713
- private:
714
+ private:
714
715
ParameterOptimizerPrivate* m;
715
716
};
716
717
717
718
class SequenceGenerator ;
718
719
class Evaluator ;
719
720
struct GradientMachinePrivate ;
720
721
class GradientMachine {
721
- private:
722
+ private:
722
723
GradientMachine ();
723
724
DISABLE_COPY (GradientMachine);
724
725
725
- public:
726
+ public:
726
727
virtual ~GradientMachine ();
727
728
728
729
/* *
@@ -817,7 +818,7 @@ class GradientMachine {
817
818
818
819
void eval (Evaluator* evaluator);
819
820
820
- private:
821
+ private:
821
822
GradientMachinePrivate* m;
822
823
823
824
static GradientMachine* createFromPaddleModelPtr (
@@ -833,10 +834,10 @@ class GradientMachine {
833
834
834
835
struct ParameterUpdaterPrivate ;
835
836
class ParameterUpdater {
836
- private:
837
+ private:
837
838
ParameterUpdater ();
838
839
839
- public:
840
+ public:
840
841
static ParameterUpdater* createLocalUpdater (OptimizationConfig* config);
841
842
static ParameterUpdater* createRemoteUpdater (OptimizationConfig* config,
842
843
int passCount,
@@ -911,17 +912,17 @@ class ParameterUpdater {
911
912
*/
912
913
void catchUpWith ();
913
914
914
- private:
915
+ private:
915
916
ParameterUpdaterPrivate* m;
916
917
};
917
918
918
919
struct EvaluatorPrivate ;
919
920
class Evaluator {
920
- private:
921
+ private:
921
922
Evaluator ();
922
923
DISABLE_COPY (Evaluator);
923
924
924
- public:
925
+ public:
925
926
~Evaluator ();
926
927
927
928
/* *
@@ -945,21 +946,21 @@ class Evaluator {
945
946
946
947
double getValue (const std::string name) const ;
947
948
948
- private:
949
+ private:
949
950
EvaluatorPrivate* m;
950
951
951
952
friend class GradientMachine ;
952
953
};
953
954
954
955
struct TrainerPrivate ;
955
956
class Trainer {
956
- private:
957
+ private:
957
958
TrainerPrivate* m;
958
959
Trainer ();
959
960
Trainer (TrainerConfig* optConfig, GradientMachine* gm);
960
961
DISABLE_COPY (Trainer);
961
962
962
- public:
963
+ public:
963
964
virtual ~Trainer ();
964
965
965
966
// / Create A Trainer By TrainerConfig. using paddle command line.
@@ -1002,7 +1003,7 @@ class Trainer {
1002
1003
1003
1004
// / the N-Best results generated from one input sequence.
1004
1005
class ISequenceResults {
1005
- public:
1006
+ public:
1006
1007
virtual ~ISequenceResults ();
1007
1008
1008
1009
// / Number of result.
@@ -1026,7 +1027,7 @@ class SequenceGenerator {
1026
1027
DISABLE_COPY (SequenceGenerator);
1027
1028
SequenceGenerator ();
1028
1029
1029
- public:
1030
+ public:
1030
1031
virtual ~SequenceGenerator ();
1031
1032
1032
1033
/* *
@@ -1044,10 +1045,10 @@ class SequenceGenerator {
1044
1045
void setMaxLength (size_t maxlength);
1045
1046
void setBeamSize (size_t beamSize);
1046
1047
1047
- private:
1048
+ private:
1048
1049
static SequenceGenerator* createByGradientMachineSharedPtr (void * ptr);
1049
1050
friend class GradientMachine ;
1050
1051
1051
- private:
1052
+ private:
1052
1053
SequenceGeneratorPrivate* m;
1053
1054
};
0 commit comments