Neural Networks

Different results on different batch sizes? Culprit : Batch Normalization

We can train out model with or without using Batch Norm. Basically Batch Norm normalizes the activations (if used after each layer then it will normalize the activations after every layer). For this purpose it will learn two additional parameters in each of its use, mean and standard deviation. This extra flexibility helps to represent identity transformation and preserve the…

Read More

Neural Networks

Solution: optimizer got an empty parameter list

Recently while building a custom Pytorch model i got a puzzling error “ValueError: optimizer got an empty parameter list” which got some hours of troublshooting, only to find that it just needs a small trick. Reproducing the error: It usually happens when you have a model with your own defined parameters, e.g. consider following example with only one parameter: While…

Read More

Neural Networks, NLP

NLTK vs Spacy: Performance Comparison of NLP Libraries in Text Tokenization

Natural Language Processing Tokenization with Spacy and NLTK

Tokenized Text Computers do not actually understand human language because the only language they understand is binary (0s & 1s). This is also the case because we have so many languages and dialects. In order to make computers understand the natural language (the language humans write and speak) we need to convert it into a computers understandable format. While working…

Read More

C++/Algorithems

Addition without using any arthmetic operator

Addition without Arithmatic Operator

Often is this puzzle brought up to test someone’s ‘bits’ knowledge but it is rather straight forward to add two numbers (whole numbers). It works by same principle as we do it by Hand, i.e. add two numbers, Keep the carry and add it into the sum. In programming we can do it achieve the simple ‘addition’ via XOR operator i.e. ^…

Read More

DWH

SSAS : Save hours by automating the cube deployments!

Analysis Services cube

Every now and then it is quite an overhead in large development environments to deploy the SSAS cubes at a large scale. Automating this process particularly helps when doing lots of changes in the SSDT (SQL Server data tools) in the development environment and when finished the changes then deploying the cubes in the UAT environment. Then after completion of…

Read More

DWH

Junk Dimensions?!

Junk dimensions in Data warehouses

Quite often SSAS/DWH designers face with the situation with several if not dozens of small what could be called small dimensions, e.g. Yes/No flags, status etc. To make each of them a separate dimension (say 20 different Yes/No flags dimensions) would simply clutter the data mart and eventually the SSAS cube. The convenient way in my opinion is to rather…

Read More