Data Science Assignment

Assignment ~ Week 2

Arya Fharezi

Arya Fharezi

NIM : 52250008
Data Science Student Institut Teknologi Sains Bandung
Academic Information
Subject
:
Data Science Programming I
Lecturer
:
Bakti Siregar, M.Sc., CDS.
Active Student - Semester 2


1 Question 1

What is the Main Purpose of Our Study? (Data Sains Programming)

The primary goal of Data Science Programming is to empower practitioners with the technical expertise to transform raw data into actionable solutions. This encompasses mastering Python and R for data processing, developing statistical models, and crafting visualizations that drive data-driven innovation. Ultimately, the focus is on harnessing computational logic to convert unrefined data into meaningful insights.

Specifically, this study aims to:

Build Technical Data Literacy

Not only read data, but manipulate, clean, and process it using code (Python/R) that ordinary software cannot do.

Analysis Automation

Create efficient systems where massive statistical calculations and data processing run automatically without repetitive manual intervention.

Predictive Modeling

Learn to train algorithms to recognize hidden patterns. Shift from understanding “what happened” to predicting “what will happen”.

Objectivity in Decisions

Replace subjective intuition with empirical evidence. Every result has a strong methodological foundation and is scientifically accountable.

2 Question 2

Why Do We Learn About It?

Data Science Programming is essential in today’s data-driven world because it bridges the gap between raw data and meaningful decisions. Organizations across all industries need professionals who can not only analyze data but also build scalable systems to process and derive insights automatically.

Key reasons:

  1. High Industry Demand

Data professionals are highly sought after; mastering these skills ensures career resilience and competitiveness in the global market.

  1. Data-Driven Decisions

Transition from intuition to evidence-based strategy, using programming to uncover patterns that lead to superior outcomes.

  1. Automation & Efficiency

Eliminate manual redundancy by building autonomous systems capable of processing massive datasets 24/7.

  1. Competitive Advantage

Organizations with robust data capabilities consistently outperform rivals, making you an invaluable asset in any sector.

3 Question 3

What tools to have to expert about?

Essential Tools for Data Analysis

01

Programming Languages

R logo

R

Ideal for statistical modeling, hypothesis testing, and creating publication-ready visualizations.

Python logo

Python

The go-to language for building end-to-end data pipelines and deploying machine learning solutions.

02

Data Manipulation Libraries

Pandas logo

Pandas

Provides flexible data structures for cleaning, transforming, and analyzing structured data efficiently.

NumPy logo

NumPy

Enables high-performance mathematical operations on multidimensional arrays and matrices.

03

Machine Learning Frameworks

Scikit-learn logo

Scikit-learn

Offers consistent interfaces for classification, regression, clustering, and model evaluation with extensive documentation.

04

Database & Version Control

MySQL logo

SQL

Essential for extracting, filtering, and aggregating data directly from relational database systems.

Git logo

Git

Tracks changes in code, facilitates collaboration, and ensures reproducibility across team members.

05

Documentation & Reproducibility

Jupyter logo

Jupyter Notebook

Combines live code, equations, visualizations, and narrative text for transparent and reproducible analysis workflows.

4 Quastion 4

Give me your Interest Domain Knowledge in Data Science?

I have a strong interest in Data Science within the Economics and Finance sector, particularly in Predictive Modeling. For me, economic data is more than just numbers it reflects market behavior and financial stability that can be anticipated through predictive models.

I am passionate about exploring algorithms such as Time Series Forecasting, Regression Analysis, and Gradient Boosting to predict economic indicators like inflation, exchange rates, and credit risk. To achieve this, I leverage SQL, Python, and Tableau as the technical foundation for building efficient analytical pipelines.

My long-term goal is to become a Financial Data Scientist who can transform complex economic data into strategic predictive insights that drive real business impact.