R is a tool which has been used over a long period of time and still continues to be used by both students and industries in the fields of statistics and data science. R continues to be important because of its versatality in these fields, especially in data exploration.

Xaltius conducted an R Bootcamp for the students of Yale-NUS to prepare them for their internal hackathon, Datathon. More than a 150 students were trained on R over the span of three workshops.

Through this R bootcamp, the students learnt how to do the following:

  • Webscraping using various R libraries and cleaning the scraped data.
  • Relational data cleaning and exploration through Exploratory data analysis.
  • Data Visualization using ggplots and plotly.
  • How business questions should be asked and given data how to approach the solution?

The students learnt the above through intensive hands-on during the sessions and self-practice. We received tremendous positive feedback and response from Yale-NUS for taking up these sessions.

If you are interested conduct such workshops and talks for your institution or be part of one, please get in touch with us.


Data Science and Python, though not a lot of people may realize it, go hand in hand with each other. Businesses today, especially the higher level management, require to see accurate and efficient depictions of various data science solutions and projects. Knowing both bridges that gap considerably.

Xaltius took to imparting the fundamentals of both these areas to over 150 students at NUS Business School over an 8 hour, hands-on intensive seminar and workshop.

Through the workshop the keys takeaways for the students were:

  • Understand the fundamentals of Python and working with small datasets.
  • How to create basic data visualizations through seaborn in python.
  • How to tell a story about your data, which is one of the most important lessons.
  • Basics of HTML, CSS and JavaScript which would help users learn about how to create basic web pages.

The end of the workshop ended by doing a small hack with the students where they were given data and had to tell a story around it. They were given an opportunity to present their findings and many of did amazingly well in such a short period!

If you are interested conduct such workshops and talks for your institution or be part of one, please get in touch with us.


Deep learning is an aspect of artificial intelligence (AI) that is concerned with emulating the learning approach that human beings use to gain certain types of knowledge. It is fast becoming a technology to reckon with and has been quite in vogue.

Xaltius held a deep learning session at Informatics Academy in Singapore, to help students and corporate professionals understand the fundaments of deep learning through an intensive hands-on session on tensorflow and keras

The key takeaways for the participants from this session was to get an understanding of the subtle differences between machine and deep learning, to understand how to build neural network models using tensorflow and keras on particular use cases, the parameters involved and how to monitor and understand how the models work.

If you are interested to be part of such workshops and talks for your institution, please get in touch with us.


With most industries relying on data today for their business growth, especially data intensive industries like banking, insurance, retail, telecoms among others, managing data to be error-free becomes important. It is known that one way of achieving maximum efficiency is to reduce all kinds of data errors and inconsistencies. If the company aims to optimize its working and increase their profits by using data, then data quality is of utmost importance. Old and inaccurate data can have an impact on results. Data quality problems can occur anywhere in information systems.

These problems can be solved by using various data cleaning techniques. Data cleaning is a process used to determine inaccurate, incomplete or unreasonable data and then improve the quality through correcting of detected errors and omissions.

What are the benefits?

Since data is a major asset in many companies, inaccurate data can be dangerous. Incorrect data can reduce the marketing effectiveness, thereby bringing down the sales and efficiency. If the organization had clean data, then falling into such situations can be avoided. And data cleaning is the way to go. It removes major errors and inconsistencies that are inevitable when multiple sources of data are getting pulled into one dataset. Using tools to cleanup data will make everyone more efficient. Fewer errors mean happier customers and fewer frustrated employees. Increased productivity and better decisions are other benefits of using data cleaning.

What are some common errors that could happen while dealing with data?

Some of the most common mistakes that occur in structured data are missing fields. Such errors can be fixed using tools like Google’s Structured Data Testing tool. This tool gives you a list detailing all of the errors, along with detailed information on the structured data Google currently detects on your website. Omissions, data that is duplicate, inaccurate or incorrect data can create expensive interruptions.If it is believed that any event does not represent a normal outcome, it needs to be filtered out from the analysis. Comparing different sets of population, segment or cluster can also result in data inconsistencies. So does drawing inferences on thin data. Another mistake that can happen is when wrong applications of the inferences are accepted. Data cleaning is an important aspect of data management which cannot be ignored. Once the data cleaning process is completed, the company can confidently move forward and use the data for deep, operational insights.

How to go about the process of data cleaning?

The manual part of the process is what can make data cleaning an overwhelming task. While much of data cleaning can be done by software, it must be monitored and inconsistencies reviewed.

Some general guidelines that all companies can follow to data clean include forming a data quality plan. By standardizing the data process, one will ensure a good point of entry and reduce the risk of duplication. Monitoring errors and fixing the data at the source can save both time and resources. Investing in tools that measure data accuracy is another wise way that can be adopted.  A reliable third-party source can capture information directly from first-party sites. It would then clean and compile the data to provide more complete information for business intelligence and analytics.

Certain data cleaning tools helps in keeping the data clean and consistent to let you analyse data to make informed decision visually and statistically. Few of such tools are free, while others may be priced with free trial available on their website.  OpenRefine, formerly known as Google Refine is a free and open source data cleansing tool. It cleans inaccurate data and transforms it. It can also transform data from one format to another, letting you explore big data sets with ease, reconcile and match data, clean and transform at a faster pace. Trifacta Wrangler is another free tool that cleans and transforms data. It takes less time formatting and focuses on analyzing data. It’s machine learning algorithms help in preparing data by suggesting common transformations and aggregations. Other tools include Drake, TIBCO Clarity, Winpure, Data Ladder and Cloudingo, among others.

This blog was written by our Content Writing Intern – Rona Sara George. Click on the name to view her LinkedIn profile.

Author: Xaltius (Rona Sara George)

This content is not for distribution. Any use of the content without intimation to its owner will be considered as violation.