Skip to main content

Java Design Patterns: Simple Factory pattern

Design Patterns are much talked about yet much feared of, much explained and yet much confused topic. A lot of fundamentals used in Design Patterns are actually application of basic Object Oriented Design principles.

Take example of 'Simple Factory' (or just 'Factory') pattern, which is not actually a pattern but more of a good design practice. Think of a scenario where you need to create a different variant of some class depending on some condition.

E.g. You have a base class House and various sub classes extending House, say Villa, Flat and Penthouse.

Now depending on some user input, say price range, you need to use objects of Villa, Flat or Penthouse to perform some function.


public double getHouseArea(double higherPriceRange){
    House house;

    if(higherPriceRange < 4000000){
      house = new Flat();
    }else if(higherPriceRange < 10000000){
      house = new Villa();
    }else{
      house = new Penthouse();
    }
    
    return house.getArea();
}

Now assume there is another similar function and you need to write the House creation code (code in blue) again! And what if we decide to add another type of house to our system, say a Treehouse, we will have to go an alter code in ALL those methods.

Won't it be a better idea to move all this code to a separate place and call that code in all the methods, so that in case of any kind of change, we have to update code in just one place.

How about this:

public class HouseFactory{
   
  public static createHouse(double higherPriceRange){
    House house;
    if(higherPriceRange < 4000000){
      house = new Flat();
    }else if(higherPriceRange < 10000000){
      house = new Villa();
    }else{
      house = new Penthouse();
    }
    return house;
  }
  
}


public double getHouseArea(double higherPriceRange){
    House house = HouseFactory.createHouse(higherPriceRange);       
    return house.getArea();
}

Awesome right? Yeah. Elementary? Absolutely.

If you noticed we just followed the basic principle of 'encapsulate what varies'. Since the House creation code can vary if any new types of houses are added, we just rounded it off to a separate class. Also this avoids code duplication across various classes.

As I said, this is not a design pattern. It is just a simple design idiom. But whatever it is, no reason for us to not use it!


Comments

Popular posts from this blog

How to upload to Google Cloud Storage buckets using CURL

Signed URLs are pretty nifty feature given by Google Cloud Platform to let anyone access your cloud storage (bucket or any file in the bucket) without need to sign in. Official documentation gives step by step details as to how to read/write to the bucket using gsutil or through a program. This article will tell you how to upload a file to the bucket using curl so that any client which doesn't have cloud SDK installed can do this using a simple script. This command creates a signed PUT URL for your bucket. gsutil signurl -c 'text/plain' -m PUT serviceAccount.json gs://test_bucket_location Here is my URL: https://storage.googleapis.com/test_sl?GoogleAccessId=my-project-id@appspot.gserviceaccount.com&Expires=1490266627&Signature=UfKBNHWtjLKSBEcUQUKDeQtSQV6YCleE9hGG%2BCxVEjDOmkDxwkC%2BPtEg63pjDBHyKhVOnhspP1%2FAVSr%2B%2Fty8Ps7MSQ0lM2YHkbPeqjTiUcAfsbdcuXUMbe3p8FysRUFMe2dSikehBJWtbYtjb%2BNCw3L09c7fLFyAoJafIcnoIz7iJGP%2Br6gAUkSnZXgbVjr6wjN%2FIaudXIqA

Running Apache Beam pipeline using Spark Runner on a local standalone Spark Cluster

The best thing about Apache Beam ( B atch + Str eam ) is that multiple runners can be plugged in and same pipeline can be run using Spark, Flink or Google Cloud Dataflow. If you are a beginner like me and want to run a simple pipeline using Spark Runner then whole setup may be tad daunting. Start with Beam's WordCount examples  which help you quickstart with running pipelines using different types of runners. There are code snippets for running the same pipeline using different types of runners but here the code is running on your local system using Spark libraries which is good for testing and debugging pipeline. If you want to run the pipeline on a Spark cluster you need to do a little more work! Let's start by setting up a simple standalone single-node cluster on our local machine. Extending the cluster is as easy as running a command on another machine, which you want to add to cluster. Start with the obvious: install spark on your machine! (Remember to have Java a

java.lang.IllegalArgumentException: Malformed \uxxxx encoding

I was getting this exception during build while running ant. Googling didn't help much and I was flummoxed because the same code was running fine till now. My code reads a text file and does some operations on the basis of values read. It was only when I saw the text files I understood the error. I had copied the text in wordpad and saved it as .txt file. Wordpad had put lot of formatting information before and after the content. Also there was "\par" after every line, which was giving this error. So moral of the story: if you get this exception check your properties file (or any other file that your code might be reading.)