Image for post
Image for post
Photo by Mohamed Nohassi on Unsplash

Daily Self Examination

Just as we go to school, attend lectures/ classes and take tests/ examinations, we also have to do same with our life style. Life is one big school filled with lessons and tests but do we evaluate ourselves?

Normally before you leave your house, you check the mirror to see if there’s any flaw in your dressing or makeup, self–examination or self–evaluation can be likened to this. It is basically asking yourself questions, giving accurate answers and grading yourself. Seems a lot?, …


Image for post
Image for post
Photo by Richy Great on Unsplash

This article will guide you on what Open Source is and how you can start contributing to open source projects.

What is Open Source?

Open Source refers to something people can modify and share because its design is publicly accessible — Open Source

How to Get Started

Requirements; A PC, Internet Connection, Git (if you don’t have it installed, you can read this article on how to do so), The project that you want to contribute to, and a Github Account.

In this article, the GitHub Uniform Resource Locator (URL) for the project is https://github.com/Team-thedatatribune/IPL-Analysis and the URL for the forked repository is https://github.com/KemmieKemy/IPL-Analysis.git, …


Image for post
Image for post

Requirements; Anaconda, A PC with Windows 10 OS, Internet connection and at least 1GB of data.

1. Run Anaconda

If you do not have Anaconda installed, you can download it from here and install then search for Anaconda in the Windows search bar. Run the Navigator and Anaconda prompt.


Have you ever wondered how you can get data off a website that does not have an Application Programming Interface (API)? Well, continue reading this article and you will learn about how you can do that.

This article will cover the following sub topics;

  • Definition of Web Scraping
  • Web Scraping Workflow
  • Web Scraping Requirements
  • How to get the best website URLs for Web Scraping
  • Hands-On experience

Definition of Web Scraping

Web Scraping is the process of extracting data from a website.

Copying and Pasting information from a website can also be termed as web scraping but on a manual level. …


Image from https://blogsbrasil.com.br/uploads/blogs/Bonequinha_png_1450118925571.png
Image from https://blogsbrasil.com.br/uploads/blogs/Bonequinha_png_1450118925571.png
Image from blogsbrasil.com.br

The last month of the She Code Africa Mentoring Program was the most thrilling because I learnt new things while building my final project but my laptop had to be formatted due to some issues. Despite everything, I’m happy that I was able to learn and also submit my project.

This article is a summary of what I learnt in the last month of the program.

Learning Path:

  • Machine Learning
  • Software Engineering
  • Data Science Project

Machine Learning

Machine learning algorithms are classified as supervised or unsupervised. I learnt two machine learning algorithms; K-Nearest Neighbors and Linear Regression algorithm. …


https://www.freepik.com/free-vector/charts-concept-illustration_5202672.htm
https://www.freepik.com/free-vector/charts-concept-illustration_5202672.htm

Building a Linear Regression Machine Learning Model

Here is a step by step process on how to do this;

  • Import libraries
  • Data Collection
  • Exploratory Data Analysis
  • Data Cleaning
  • Build model
  • Evaluate metrics

Import Libraries

I used Jupyter Notebook as the Integrated Development Environment (IDE). The libraries required are; numpy, pandas, matplotlib, pickle or joblib and scikit-learn. These are pre-installed in the latest version of Anaconda. If you don’t have any of these libraries you can pip install them or update conda.

Data Collection

The dataset used for this model is the Pima Indians Diabetes dataset which consists of several medical predictor variables and one target variable, Outcome. Predictor variables include the number of pregnancies the patient has had, their Body Mass Index, insulin level, glucose level, diabetes pedigree function, blood pressure, skin thickness and age. …


Image from www.clipart-gratis.com — A lady sitting on a desk using a laptop
Image from www.clipart-gratis.com — A lady sitting on a desk using a laptop

To understand this article better, I would advice that you read the piece above if you have not already done that.

This article is a summary of what I learnt in the second month of the program.

Learning Path:

· Python Libraries for Data Analysis

· Data Wrangling

· Mathematics for Data Science

· Analysis of a Dataset

Python Libraries for Data Analysis

I learnt the rudiments of NumPy and Pandas libraries for data analysis.

NumPy stands for Numerical Python. It is a library consisting of multidimensional array objects and a collection of routines for processing these arrays. …


Image for post
Image for post
1000-downloads-dx-plugin-base

This analysis was performed on Google Play Store App dataset from Kaggle using Python . In this dataset, there are 10,842 applications and each application has the following details: Application name, Category, Rating (on a scale of 5), Number of Reviews, Size of the app, Number of installs, Application Type — Free or Paid, Price, Content Rating, Genres, Last updated On, Current Version, Required Android Version.

In order to get 1,000+ downloads on your Android Application, you need to know the current statistics of the android market.

This article will display the results of the following questions;

1. What category of applications have the highest number of installs? …


Image from esty.com
Image from esty.com
Image from esty.com

Last month was quite exciting for me as an aspiring Data Scientist; I was selected to participate as a Mentee in the 1st Cohort of the She Code Africa Mentoring Program, Data Science track. I was assigned a Mentor and a Learning Path. Needless to say, I was filled with overwhelming joy to advance my Data Science journey with an experienced mind in the field solidly behind my endeavor.

This article is a summary of what I’ve learnt so far.

Learning Path:

· Basic SQL.

· Tableau for Data Visualization.

· Python

· Git

Each path had its assignments and resources.

Basic SQL

I learnt the fundamentals of SQL Joins, SQL Queries and SQL Aggregate functions. SQL Joins are used to combine rows or columns from two or more tables, SQL Aggregate functions are used to produce summarized data from a database, these functions are; COUNT, MAX, MIN, SUM, AVG…

About

Ekemini Okpongkpong

Aspiring Data Scientist | Tech Savvy

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store