Skip to content

waldekmaciejko/ML-DL-snieppets

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Folder DSP

Contains some implementations of fundamental algorithms use in Digital Signal Processing

Folder kaggle_competitions

Solutions of some kaggle competitions. Results base upon those notebooks were submited to kaggle webservice. In folder have not been located datasets, due to licence limits. To properly test solutions you have to approve licences on each kaggle competition, next you will be able to download datasets.

Folder NLP

Contains implementations of fundemantals of NLP

Folder pytorch_snipets

Practical use of basic tool for machine learning using PyTorch lib.


recapAllInOne.py


  • example of binary classification cats vs dogs
  • to start tensorboard ...$ tensorboard --logdir=logs
  • reads data from one, local source (no split into traning and validaion dataset) using lista and torch.utils.data.Dataset
  • uses torchvision to transform data
  • uses custom function in torchvision.transform pipline
  • estimate accuracy (sigmoid) (no nn.BCEWithLogitsLoss)
  • added the adaptive LR
  • added early stopping and dropout layer
  • added argument parser

autoregresive


  • contains several easy solutions of common problems of predictions time series using autoregresive method
  • framing function
  • Vanilla RNN

Folder utils

Various scripts for machine learning

extract_mfcc_libri_pools.py

function for Libri speech audio pcm data base that creates dict: keys - speaker label, value - path to flac file then creates callable class; during calling instance estimate MFCC using librosa lib and does some other minor preprocessing steps like: convert to mono, resampling to 16 kHz, trmming then using multiprocessing lib start process-based parallelism

pca_definition.py

function creates PCA algorithm using theoretical description

ensamble machine learning algorithms (ensambleML)

Jupyter notobook shows how to use ensamble methods like Bagging (BaggingClassifier), Boosting (GradientBoostingClassifier, AdaBoostClassifier, XGBoost), Stacking (used all above and Logistic Regression or SVM as a meta models) and Blending (used all above and Logistic Regression as a meta model). In example data from Kaggle were used.

About

Collection of short pieces of code as examples of use basic tools from Machine Learning and Deep Learning libraries

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors