I have recently going through the Google MAPs API and learned about displaying the traffic information of a city or a place in the Google Map.
It’s very easy to integrate . If we want to show the traffic information in our site, then this API comes in handy for us.
Before using this one in your site, you should have to get the Google API Key as this one requires a valid API key.
I want to find out the traffic information for Columbus, Ohio. Hence updated the latitude and longitude.
Jsfiddle page with code and output:
In this post, I am going to show three things,
- How to Install the Apache solr 6.3.0 on Ubuntu
- How to load a CSV file which contains US baby names information
- How to query the data with the Solr API
To install Apache Solr, Follow the below steps
sudo wget http://apache.claz.org/lucene/solr/6.3.0/solr-6.3.0.tgz
sudo gunzip solr-6.3.0.tgz
sudo tar -xvf solr-6.3.0.tar
Go into the folder solr-6.3.0 and open up a terminal and then type below command to start the Solr server
Check the solr admin console with this link http://localhost:8983/solr
Then the next step is to create the collections and load the CSV data.
bin/solr create -c babynames
Once we create the collections, then we have to specify the field definitions in the schema file. The schema file is available under the /server/solr/babynames/conf/ folder. managed-schema is the schema file name. You can rename this to schema.xml. But i just keep it as it and add the below fields in that file
<field name="Count" type="int" indexed="true" stored="true"/>
<field name="Gender" type="string" indexed="true" stored="true"/>
<field name="Id" type="int" indexed="false" stored="false"/>
<field name="Name" type="text_general" indexed="true" stored="true"/>
<field name="Year" type="int" indexed="true" stored="true"/>
Then load the CSV file with the below command. I have used this file https://github.com/dkbalachandar/spark-scala-examples/blob/master/src/main/resources/NationalNames.csv for this exercise
bin/solr post -c babynames NationsNames.csv
Finally, I query the data with the Solr REST API.
To search with Name: http://localhost:8983/solr/babynames/select?q=Name:%22Mary%22
To search with Gender : http://localhost:8983/solr/babynames/select?q=Gender:%22M%22
To search with year range: http://localhost:8983/solr/babynames/select?q=*&fq=Year:%5B1880%20TO%201890%5D
Refer below the screenshots taken.
I have developed a MEAN stack application which shows the San Francisco Food inspections details.
Source: Food Inspection(Use Food Inspections – LIVES Standard)
I have used Spark, Scala, MongoDB, NodeJs, AngularJs to do this.
My spark job reads the input CSV data contains food inspection details and processes it and stores the data in MongoDB as collections. I have allFoodInspection and filterFoodInspection collections here. The first one has all the data and the second one has the business name, the unique risk category and number of risk’s committed.
My MEAN stack REST layer reads the data from Mongodb and processes and exposes the data and the Web Layer uses the data and display it and use the data for drawing a chart.
Let us see how we can execute this.
- Follow the steps given in this post to install scala, sbt and spark in your machine if you are using Ubuntu. Refer my another post to know how to install these. How to install Scala, SBT and Spark in Ubuntu
- Clone the git repository https://github.com/dkbalachandar/sf-food-inspection-spark.git and go inside of sf-inspection-spark folder and run ‘sbt assembly’ to create a far jar with all the dependencies.Here I have used spark 2.0.2 and scala 2.11.8 (Spark 2.0.2 version is compatible with scala 2.11.x version).
If you don’t use the compatible version then you will end up with lots of errors.
- Copy the ../sf-food-inspection-spark/target/scala-2.11/sf-food-inspection-spark-assembly-1.0.jar to /usr/local/spark folder
- Download Food_Inspections_-_LIVES_Standard.csv from https://data.sfgov.org/browse?q=food+inspection and move it to /usr/local/spark folder
- Install Mongodb with the below steps
sudo apt-key adv --keyserver hkp://keyserver.ubuntu.com:80 --recv 0C49F3730359A14518585931BC711F9BA15703C6
echo "deb http://repo.mongodb.org/apt/ubuntu trusty/mongodb-org/3.4 multiverse" | sudo tee /etc/apt/sources.list.d/mongodb-org-3.4.list
sudo apt-get update
sudo apt-get install -y mongodb-org
sudo service mongod start
Run the spark job with the below command
bin/spark-submit --class com.spark.SFFoodInspectionAnalysis --master local sf-food-inspection-spark-assembly-1.0.jar file:///usr/local/spark/Food_Inspections_-_LIVES_Standard.csv
- Then check the Mongo Db and check the collections and make sure that the data are getting inserted and availableOpen up a terminal window and type ‘mongo’ and enter. It will open a shell window. Then use the below commands to verify the data
- Clone the git repository https://github.com/dkbalachandar/sf-food-inspection-web.git and go inside of sf-food-inspection-web folder, then run below commands to build and run the application
- Open the http://localhost:8081 and check the page. I have used the data and created a table and display a chart with the details.
Please are the some of the screenshots taken from the application
Install Scala 2.11.8 and SBT
sudo apt-get remove scala-library scala
sudo wget www.scala-lang.org/files/archive/scala-2.11.8.deb
sudo dpkg -i scala-2.11.8.deb
echo "deb https://dl.bintray.com/sbt/debian /" | sudo tee -a /etc/apt/sources.list.d/sbt.list
sudo apt-key adv --keyserver hkp://keyserver.ubuntu.com:80 --recv 642AC823
sudo apt-get update
sudo apt-get install sbt
Install Spark 2.0.2
sudo wget http://d3kbcqa49mib13.cloudfront.net/spark-2.0.2-bin-hadoop2.7.tgz
sudo chmod -R 755 spark-2.0.2-bin-hadoop2.7.tgz
sudo gunzip spark-2.0.2-bin-hadoop2.7.tgz
sudo tar -xvf spark-2.0.2-bin-hadoop2.7.tar
sudo mv spark-2.0.2-bin-hadoop2.7 spark
sudo mv spark /usr/local/spark
Open bashrc file and add spark_home and update PATH
sudo vi ~/.bashrc
In this post, I am going to explain about my exposure to these cloud computing terms.
SaaS – Software as a service
– One of my client X uses a Financial Planning application that has been developed by a different company Y. The application is available and deployed on the client network. The main drawback is that company Y has to closely follow up with company X to make any changes on the application and also maintaining the latest version of this is hard for them. So Company Y wants to deploy that application on their network and provide the access to Company X users. So Company Y gives Software as a service to Company X
PaaS – Platform as a service
– Recently, I have participated in a Hackathon to develop a health application that should be deployed on the Red Hat Openshift Platform. The OpenShift provides numerous things to develop, host and scale LAMP applications. Here we use their cloud environment to host our application. So OpenShift provides Platform as a service to us for hosting the application
IaaS – Infrastructure as a service
– It is an instant computing infrastructure, provisioned and managed over the Internet. It helps us to quickly scale up and down with demand and pay only for what we use. We use Openstack for provisioning and managing the new virtual machines.