Sunday, 4 February 2018

Multiple Plots into One Figure in Python MatplotLib

Lets learn, How to plot multiple charts into one Figure. Embedding a jupyter notebook here with the examples.

Like the below page to get update  
Facebook Page      Facebook Group      Twitter Feed      Google+ Feed      Telegram Group     

Saturday, 3 February 2018

Jupyter Lab Env - Quick Start Script

Every Data Analyst, who is working in Python, is very well aware of Jupyter or Jupyter Lab. 

Sometime it seems little annoying, to lazy person like me, to start the Command Prompt, go to your working code directory and type jupyter command to start the notebook.

To avoid this burden I have wrote a small batch script which will do the task for me in one click. Sharing here the same, Lazy Programmers, please share this post if you like it :-)

For Better View, Click on "View Raw"

Like the below page to get update
Join Telegram DataStage4You Group

Tuesday, 16 January 2018

Python Pickle - to save your efforts

This tutorial is for Python beginners who just started getting dirty in Python :-) Lots of time when we are working on some data set and completed lots of cleaning and pre-processing steps, It's advisable from GURUs to save that intermediary dataset to avoid the re-do all the steps if something unexpected happen with Python, Jupyter notebook or your system for rebooted without your permission, weird though :-/

So, let's learn how to save any dataset or variable with Pickle library -

To Import:

To Save: 

To Read:

Like the below page to get update

Sunday, 14 January 2018

Mongo DB - Installation and Configuration

MongoDB  is an open-source document database, and the leading NoSQL database. Written in C++.
MongoDB features:
    Document-Oriented Storage
    Full Index Support
    Replication & High Availability
    Fast In-Place Updates

Reduce cost, accelerate time to market, and mitigate risk with proactive support and enterprise-grade capabilities.

Today, We will see how to install and run the MongoDB.

MongoDB Installation on Linux

1. DOWNLOAD the stable version of MongoDB. It will a tar file
2. Extract the tar file to some directory.
$ tar -xvf mongodb.tar -C /learn/mongodb

3.  change the permisson of folder to user who run the db here-  In my case User - hduser and Group - hadoop
$ chown -R hduser:hadoop /learn/mongodb

4. Add the env var in .bashrc
export MONGO_HOME=/learn/mongodb

5. Create the default DB directory for Mongo
$ mkdir -R /data/db
$ chown -R hduser:hadoop /data/db

This is by default, you can specify ur db path when starting the mongo db

$ mongod --dbpath /app/mongodata
this command will start the mongodb. in other terminal you can start work on db. "--dbpath /app/mongodata" is totally optional

If you just use just $ mongod , it will start n use the default db which we have defined in step 5.

Please don't close the current terminal, It can be kill the mongodb process.

6. Start working on MongoDB
$ mongo

Like the below page to get update

Tuesday, 2 January 2018

Conda Commands

Sharing some CONDA commands which fasten the administration of different python environment. All these commands are available on Anaconda Site 

Like the below page to get update

Wednesday, 20 December 2017

How to export the datastage job in individual dsx

This script will export the datastage jobs into individual dsx files as sometimes, we need individual dsx to maintain the code version or syncing it with SVN.

Manually export of each datastage job is quite hectic and boring hence this script :-)

This script requires little bit information from you and set you free for your other chores

Prerequisites for this script -

1. Job List to Export
2. DataStage Client Installation Location
3. DataStage Credential

Like the below page to get update

Sunday, 19 November 2017

Get Job Parameters for all DataStage Job

Feels awesome while writing this post after 7 week, In this post I am going to share a script which will fetch the job parameters for all the datastage job available in given project.

This week, one of my colleague ask to compare all the jobs parameter from one project (older) to another project (migrated),  So, wrote down this script which worked like charm for me and hoping will work for you as well :-)


Like the below page to get update