How to install go-daddy ssl certificate on amazon load balancer.

Written by Tarun Jangra on December 29th, 2012

I was struggling around to get this done. And finally i’ve made that. Heres the straight Forward Steps.

Requirements & Prerequisites:

  1. Linux having openssl and apache installed.
  2. Open shell terminal on your Linux Box.


openssl genrsa -des3 -out private.key 1024
openssl req -new -key private.key -out www.your-web-site.com.csr

It will ask you to add some basic information. Make sure you have added “Common Name” as your domain, like “www.xyz.com”

  1. Go to www.godaddy.com along your ssl management control panel
  2. Select your Certificate. And click on Re-Key button.
  3. Copy content of “www.your-web-site.com.csr” and paste the content in “CSR” field. And press Re-Key.
  4. It will prompt you to download the key. I found, Apache, Other, Nginx are same. So use anyone. By the way, i used “Other” to download my keys.
  5. Now unzip the downloaded content. It should have two *.crt files.

Now back to your terminal.

openssl rsa -in private.key -out private.pem

Now you will have following files in your current location.

  1. private.key
  2. private.pem
  3. www.web-site.com.csr
  4. sf_bundle.crt
  5. your-domain.com.crt

Now open your load balancer console and add https support. it will prompt you to add following values.

  1. Certificate Name:* -> Put any friendly name
  2. Private Key:* -> Paste content of private.pem
  3. Public Key Certificate:* -> Paste content of your-domain.com.crt.
  4. Certificate Chain: -> Paste content of sf_bundle.crt

Save this and you are done. It is quite easy. No?

Twitter Weekly Updates for 2012-06-24

Written by Tarun Jangra on June 24th, 2012

Continous Integration Experience.

Written by Tarun Jangra on February 9th, 2012

As a DevOps, It was my first experience to create a strategy where we could have automatic deployment stack. It was really a great experience and i get in to complete execution flow of Agile Methodology. The whole idea is just to work around the cleaner continuous deployment, Where i could have all level of tests like unittest, codecoverage, acceptance testing and so on….
First i tried to get implement with Hudson, It was really great tool but i found it little time consuming. Might be it was my first drive. But in the 2nd inning, i try to get in Jenkin and it really makes me so comfortable. I just fell in love with Jenkin. I found so many “Jenkin Vs Hudson”. Well at this level, i have so personal preference of my experience. It might be the reason of my second inning. But i would like to remain with Jenkin. Now going to create a whole stake of CI server at my local environment for one of our project. Actually client is concerned to Amazon’s Cost so it really make sense to avail local resources to generate quality output. And My Team is also feeling same what i am.
Would like to thanks to the great contribution of Jenkin Community. They really made an out standing product. Hats Off for such contributors to the community.

wow i am writting this from wordperss app for blackberry.

Written by Tarun Jangra on January 19th, 2012

It is really exciting to write this post from my black berry application for wordpress. I am really impressed.

Our development work flow with git.

Written by Tarun Jangra on January 19th, 2012

It is pretty worth able to share this experience. I used to git since 2009 but using git in production is no longer than 3 months. We were using SVN before that and now moved to git. It is really cool. We were using it with Master branch only where all developers suppose to push and code is suppose to move to development server and after testing, it is suppose to deploy on production server. Which is really quite unstable process. and as we are getting in the requirement of better development with less efforts we seriously need to think about branching and all such stuff.
Well, i created a quite simple but power full scenario. Master branch will be now our production ready branch. And Development branch will be our dev server branch. These two branches are suppose to be in the system for infinite time. And i introduced some short life branches like “Feature branches” and “Release branches” which will really play great role in the architecture we are working.
We are using “Pivotal Tracker” for our Agile methodology, So when we have new mile stone with the story id. It means, developer need to create new branche with the name “Feature-“. This branche is suppose to be cloned from development branch and suppose to be in the system till the completion of the feature. and than merge back to development branch. So in the whole release we are suppose to complete all pivotal stories by story ids.
I am looking for some automatic process where all stories get started when developer creates the Feature branch. And when he deliver the whole feature and merge the branch back to the development. It should automatically change the status of the story to be “Delivered”. QA team will test and either accept or reject the corresponding story.
Will share if i could figure this out. Overall, really challenging and quite interesting situations i’m facing and finding solutions for them.

Being a DevOps At YourSports

Written by Tarun Jangra on November 18th, 2011

WOW! This is what i was dreamed to be. Really exciting job. From the last couple of year, I did lots of work in social networking application, Cloud infrastructure management, Ruby. Since i joined YourSports, I am damn busy. Being a DevOps, And it is quite hard to get time to write something here. But i am trying again to get some time to write here. No doubt, This post is after so long. But i will try to keep this on. My job is to manage all project activities between engineers. Our Engineers are distributed team and i have to manage to communicate with them in different timezones. Yeah, somewhat difficult. But really enjoying. Cloud Infrastructure, This is something i am die fan of. It is really great feeling when we are involve with some thing which needs to invent and i am experiencing something like that. Working with Engineers from different communities and cultures also improved my management skills. What we are building is on elgg Which is itself makes me Proud over My Team (Karam (Elgg Developer) , Chetan Sharma (Elgg Developer), Narinder Kaur (Lucene Expert), Daniyal Nawaz(UI Engineer) ). Special Thanks to Elgg team for their incredible product. There is nothing you can not achieve with elgg even with EAV infrastructure, Which is elgg’s strongest point.

Round-robin at application level to Balance MySQL Database Load.

Written by Tarun Jangra on June 19th, 2011

Round robin technique facilitates you to distribute your task on number of available resources even at different location. Huge traffic sites like Facebook has to has such techniques working at the background to serve as fast as possible. I would like to discuss one of my personal implementation experience for such a large potential social networking site. Cloud computing is really help full but it also needs logical approach at programming level. Click here to continue reading →

How to create custom amazon AMI.

Written by Tarun Jangra on March 4th, 2011

Today, i am going to explain how you can create custom amazon ami to launch instance anytime later. This will have you clone of your server anytime you need. I am considering you are able to login your current running instance and you also have your private key and certificate downloaded on some location.

  1. Upload your private key and certificate on the running instance.
    scp -i path/of/yourkeypair.pem path/of/cert.pem /mnt
    scp -i path/of/yourkeypair.pem path/of/pk.pem /mnt
  2. Login to your instance and check if uploaded files are available in /mnt.
  3. ec2-bundle-vol -d /mnt -k /mnt/pk.pem -c /mnt/cert.pem -u 673491274719 -p name-of-ami

    This will take some time and create the desired ami to be uploaded in the bucket. So you can use that later anytime you need.

  4. Now upload your bundle to amazon s3 storage.
    ec2-upload-bundle -b <S3-bucket-name> -m /mnt/name-of-ami.manifest.xml -a <AWS-access-key-id> -s  <AWS-secret-access-key> --location US-EAST-1C

    Note: Remember to upload to an S3 bucket in correct region. Also: if the bucket does not exit, it will be created for you. (I’ve used a European bucket as an example.)

  5. Now we need to register AMI. Do following< br />
    ec2-register <bucket-name>/sampleimage.manifest.xml --region US-EAST-1C

    It will return the new AMI ID (like ami-).

That’s it you are done with your custom ami.

Guest eth0 is missing on guest machine (virtual box).

Written by Tarun Jangra on January 27th, 2011

My host machine is Kubuntu. And i have some old .vdi files. I normally added them in new virtualBox and start guest os (Squeeze). I see old network configurations are missing now. There were no ethier net (eth0).
I removed “/etc/udev/rules.d/70-presistent-net.rules” and poweroff the guest operating system. Selected “bridge adapter” from network configuration of the guest operating system. Start again and i get all network configurations back.
Now my host can ping guest operating system as it is available in the network.

Working on Qt Development

Written by Tarun Jangra on December 24th, 2010

It’s been too long to write back here. I was quite busy from last couple of months. Working on so many different tasks and getting no time to research here and there. But finally i got some time to get start with something very excited. This is learning of Qt Framework. I scheduled my next year to start working on Qt Framework applications. For the time being, i have not so much to write here. I will try to keep continue writing on my blog.