Program  

Courses
Location
Corporate
Student Success
Resources
Bootcamp Programs
Short Courses
Portfolio Courses
Bootcamp Programs

Launch your career in Data and AI through our bootcamp programs

  • Industry-leading curriculum
  • Real portfolio/industry projects
  • Career support program
  • Both Full-time & Part-time options.
Data Science Bootcamp

Become a data engineer by learning how to build end-to-end data pipelines

 

Become a data analyst through building hands-on data/business use cases

Become an AI/ML engineer by getting specialized in deep learning, computer vision, NLP, and MLOps

Become a DevOps Engineer by learning AWS, Docker, Kubernetes, IaaS, IaC (Terraform), and CI/CD

Short Courses

Improve your data & AI skills through self-paced and instructor-led courses

  • Industry-leading curriculum
  • Portfolio projects
  • Part-time flexible schedule
Portfolio Courses

Learn to build impressive data/AI portfolio projects that get you hired

  • Portfolio project workshops
  • Work on real industry data & AI project
  • Job readiness assessment
  • Career support & job referrals

Build data strategies and solve ML challenges for real clients

Help real clients build BI dashboard and tell data stories

Build end to end data pipelines in the cloud for real clients

Location

Choose to learn at your comfort home or at one of our campuses

Corporate Partners

We’ve partnered with many companies on corporate upskilling, branding events, talent acquisition, as well as consulting services.

AI/Data Transformations with our customized and proven curriculum

Do you need expert help on data strategies and project implementations? 

Hire Data, AI, and Engineering talents from WeCloudData

Student Success

Meet our amazing alumni working in the Data industry

Read our students’ stories on how WeCloudData have transformed their career

Resources

Check out our events and blog posts to learn and connect with like-minded professionals working in the industry

Read blogs and updates from our community and alumni

Explore different Data Science career paths and how to get started

Our free courses and workshops gives you the skills and knowledge needed to transform your career in tech

Blog

Student Blog

Data Visualisation in Einstein Analytics using Stack Over Flow data from Redshift.

November 9, 2020

The blog is posted by WeCloudData’s student Sneha Mehrin.

This Article Outlines the Key Steps in Creating a Highly Interactive Dashboard in Einstein Analytics by Connecting to Redshift.

dashboard for redshift
image from https://www.searchenginejournal.com/

This article is a part of the series and continuation from the previous article where we build a data warehouse in Redshift to store the streamed and processed data.

Let’s briefly review our pipeline and the schedule of the batch jobs that we developed.

Data Pipeline with Schedule

We already covered the streaming and batch jobs in aws, now let’s move on to Einstein Analytics.

  • Einstein Analytics is a cloud based analytics solution offered by Salesforce.
  • The data is stored on its own platform, but because it is a Salesforce Product, it is well integrated with the Salesforce platform.
  • The key advantage of Einstein Analytics is its own suite of data prep tools DataFlow and Recipes , which provides a lot of flexibility to the analytics developer.

  • Einstein uses Data Sync to connect to external data sources and its own native Salesforce local objects.
  • We can schedule the Data Sync to extract the data at a specified time.
  • The data can be further cleaned, prepped and transformed using DataFlows and registered in the form of datasets which further gets used in Dashboards.

In our technical design, we are connecting EA to a a Redshift cluster using the Redshift Connector provided by EA and registering into a dataset using DataFlow.

Key Steps In this Process

Please note that this article is written keeping in mind that readers already have some idea of how to get started with Einstein.

  1. Configure the Redshift cluster details in EA.

2. Once the Connection is successful, select the table in Redshift cluster you created and the columns.

Image for post
Table in Redshift

Step 3 :After this is done , go to the connect tab, your connections will show up. Click on Run Data Sync to sync the data.

Step 4 :Create a simple DataFlow using SFDC Digest as the amazon connection and use SFDC register to register the dataset.

Since, I already deleted the redshift cluster to avoid any costs, the nodes are showed with warning. However, the process remains the same.

Dashboard has three main components

  1. Key KPI Metrics : This is controlled by the Date filter and users can toggle between different date filters.
  2. Metrics Over Time : This is a static visualisation showing the key KPI metrics over time in the current month.
  3. Comparison of Metrics Over time : This is also a static visualisation, however this gives a comparison of the metrics compared to last year.

Key Components of the Dashboard

  1. Date Filter
  • This is a static filter with the below values

Image for post

2. Key KPI Metrics

These are compact number widgets with the static date filter binded to it in the filter section.

Logic behind this is quite simple; only additional configuration these visualisations have are the bindings.

New Questions : Unique(question_id)

Average Answer count : avg(answer_count)

Average View Count : avg(view_count)

Average Score : avg(view_score)

Here is the snippet of the bindings in the filter section encoded in the underlying json of the dashboard.

3. Line Charts

These visualisations do not have any bindings, but it has an additional metric such as % of growth in the tooltip.

Here is a quick video on how I configured these charts.

Since the visualisation can have only measure, the key tip is to create % of Growth using compare table and then hide it, so that the query projects it ,but it doesn’t render in the visualisation. You can then use the hidden measure in the tooltip

4. Line Charts Comparison

Since these charts are essentially a new query, I used pages to provide a seamless experience of navigating between the two views.

I used SAQL for the underlying query of these charts.

Here is a quick video on how I configured this visualisation.

That’s all for this series!!

It was a true learning experience to build this pipeline. If you have any questions, please feel free to shoot me a message!!

To find out more about the courses our students have taken to complete these projects and what you can learn from WeCloudData, view the learning path. To read more posts from Sneha, check out her Medium posts here.
SPEAK TO OUR ADVISOR
Join our programs and advance your career in Cloud EngineeringData EngineeringData Science

"*" indicates required fields

Name*
This field is for validation purposes and should be left unchanged.
Other blogs you might like
Student Blog
The blog is posted by WeCloudData’s Big Data course student Abhilash Mohapatra. This story represents an easy path to…
by Student WeCloudData
April 30, 2020
Career Guide, Guest Blog, Learning Guide
The newest article by Taylor Nichols on switchup shows that the move to online was more popular than people…
by Student WeCloudData
October 2, 2020
Student Blog
This blog series is posted by WeCloudData’s Data Science Immersive Bootcamp student Bob Huang (Linkedin) OVERVIEW: The digital marketing…
by Student WeCloudData
October 28, 2019
Previous
Next

Kick start your career transformation