Chapter 1 Tupad 2
Chapter 1 Tupad 2
Chapter 1 Tupad 2
BSIS 4B
Members:
Erguiza, Jeric E.
1.1 Introduction
workers for various tasks, TUPAD improves the physical environment of communities,
making them cleaner, safer, and more livable. It also supports workers in earning an
income through community-based projects, enhances their skills for future employment,
and aids in their livelihood recovery. The increasing adoption of machine learning
models and advanced algorithms for predictive analytics, coupled with a growing
efficient resource allocation, timely support, and community development have become
technology and data analytics aims to enhance social welfare outcomes, addressing
disasters or economic downturns. the TUPAD program faces several challenges for both
beneficiaries and management. Beneficiaries often struggle with limited job opportunities
that don't match their skills, leading to financial instability due to the program's
services, such as training, further complicate their experiences (Luh Putu Saraswati Devia
Jayanti, 2022). For management, effective resource allocation and accurate data tracking
are significant challenges that require robust systems (Murat Levent Demircan & Kaan
Aksaç, 2022). Coordination with various stakeholders can be complex, and public
Employment, 2023). This study aims to enhance the predictive accuracy of employee
model utilizing an expanded dataset. Previous studies have employed Random Forest and
Naive Bayes classification algorithms, achieving accuracy levels of 70-80% (Luh Putu
71.19% accuracy rate through five-fold validation, while Support Vector Machine
(SVM), K-Nearest Neighbors (KNN), Decision Trees (DT), and Multi-Layer Perceptron
Testing the developed model on a fresh dataset yielded a 71.67% accuracy (Murat Levent
Demircan & Kaan Aksaç, 2022). This investigation seeks to address limitations by
incorporating a larger and more diverse dataset, additional relevant features, and
model, the TUPAD program can improve beneficiary selection accuracy, reduce errors,
machine learning model for predicting TUPAD beneficiaries, achieving the highest
accuracy to date. By leveraging this model, TUPAD can facilitate fast and accurate
enables efficient beneficiary selection, reducing processing time and resources while
enhancing community impact by targeting those most in need. Ultimately, the findings
welfare initiatives.
Machine learning techniques can improve the selection process for qualified
methods are prone to biases and inefficiencies, while machine learning models like
decision trees and random forests can analyze complex data to predict eligibility more
effectively. This approach ensures faster, more accurate beneficiary selection, helping
human resource management (Ikram et al., 2019). The use of machine learning models,
such as decision trees and random forests, has demonstrated improved accuracy in
identifying high-performing employees (Son & Kim, 2019). These models analyze
(Kim, 2019).
models offering predictive insights to mitigate turnover rates (Ikram et al., 2024). By
predictive models like random forests and neural networks can forecast turnover risk,
enabling organizations to take proactive measures (Perkins & Neumayer, 2024). This
In the study by Nandy and Kamila (2022), machine learning techniques are
for specific roles. By utilizing models such as logistic regression, support vector
machines, and random forests, the analysis identifies key features like skills, experience,
and qualifications that contribute to successful job placement. This approach not only
improves the efficiency of recruitment but also reduces bias and helps organizations
gained significant attention, enhancing the ability to predict candidate success and
streamline hiring (Khan et al., 2022). By leveraging various algorithms, organizations can
analyze large datasets to identify patterns and factors that contribute to successful job
placements (Shah et al., 2022). Machine learning models enable recruiters to assess
candidate qualifications, predict turnover risks, and improve overall hiring efficiency (Ali
et al., 2022). These predictive insights not only help in making informed hiring decisions
but also foster a more data-driven recruitment strategy that aligns with organizational
goals.
Predictive Analysis of Job Recruitment Using Machine Learning
This study aims to develop a predictor model that proposes suitable job positions
This study seeks to explore methods for predicting beneficiaries of TUPAD or the
identify individuals who qualify for assistance by utilizing various machine learning
tools. the study will help in the selection of qualified beneficiaries ensuring that support
reaches those who need it most. The study can suggest that social welfare programs use
technology to identify the right individuals who need assistance, resulting in more
effective outcomes.
Future Researchers: This research can serve as a valuable resource for future
researchers, providing them with insights and guidance for their own investigations. The
findings of this study can serve as a reference point and inform their future work.
Social Workers and Practitioners: The research may benefit to social workers
beneficiaries' needs, enabling them to provide more targeted support and interventions.
as the University of Caloocan City, to understand how machine learning can be integrated
into social policies, this research will support the development of an innovative
organization like NGOs they can use the study’s findings to advocate for better support
systems for working individuals, enhancing their programs to meet community needs
effectively.
Beneficiaries: The study may have benefits for the beneficiaries as well because
it may lead to better selection procedures that will help those who are poor and families
receive right on time assistance, which may improve their quality of life.
1.4 Theoretical/Conceptual Framework
Figure 1.
table 1 Shows the project aligns with business goals by first defining objectives
and success criteria, such as improving candidate selection. Next, it assesses resources,
risks, and conducts a cost-benefit analysis. Data mining goals are then set to determine
technical success metrics, like predictive accuracy. Finally, a detailed project plan
outlines tools, technologies, and timelines to ensure the process meets both business and
technical requirements. table 2 Shows the process starts with collecting initial data,
ensuring all necessary datasets are acquired and loaded into the analysis tools. Next, the
data is described by examining its structure, such as format, number of records, and field
identities. Afterward, the data is explored more deeply through queries, visualizations,
and relationship analysis. Finally, data quality is verified by identifying and documenting
any issues or inconsistencies to ensure it is suitable for further processing. table 3 the first
step is to select relevant datasets and document the reasons for their inclusion or
exclusion. Next, data cleaning is performed, often the most time-consuming task, to
correct, impute, or remove errors, ensuring the quality of the inputs. The process also
involves constructing new attributes, such as calculating a body mass index from height
and weight. Data integration follows, where multiple sources are combined to create
strings to numeric values, to ensure compatibility with further analysis and modeling.
table 4 Shows the process starts by selecting appropriate modeling techniques, such as
regression or neural networks. A test design is created by splitting the data into training,
testing, and validation sets. The model is then built, often through straightforward code
knowledge, predefined success criteria, and the test design. All the criteria and figures
discussed thus far lead to *table 5*, which focuses on the *Evaluation* phase of the
CRISP-DM model. In this phase, the aim is to assess the effectiveness and capability of
the models developed during the previous phases. By examining performance metrics and
ensuring alignment with the defined business objectives and success criteria, this
evaluation provides critical insights into the reliability and accuracy of the data mining
process. table 6 Will illustrate the Deployment phase, where the successful models are
implemented into a real-world environment. This phase ensures that the insights gained
from the evaluation are translated into actionable strategies, allowing businesses to apply
the findings effectively. Overall, both figures emphasize the importance of evaluating the
models' capabilities and ultimately deploying them to achieve the desired business
outcomes. We chose the CRISP-DM model for its structured yet flexible approach to data
The main concern of the study was to make a Machine learning that could predict
TUPAD members being qualified or unqualified Specifically, the study sought to answer
member
2. How accurately can the machine learning model predict whether a TUPAD
3. What machine learning algorithms are most effective for predicting the
4. How does the performance of the machine learning model compare to traditional
5. What implications does the model's accuracy have for improving the overall
6. How can the predictions from the machine learning model inform the decision-
1.6 Synthesis
The issues faced in our study revolve around the challenge of hiring qualified
beneficiaries for the TUPAD program. Based on our references, companies often
encounter significant costs and lengthy processes when recruiting suitable candidates
(Luh Putu Saraswati Devia Jayanti, 2022). To address this, the proposed study will utilize
classification algorithms such as J48, Random Forest, and Naive Bayes. These algorithms
productivity datasets. The evaluation of these models will focus on three key metrics:
accuracy, Mean Absolute Error (MAE), and Root Mean Square Error (RMSE). Results
from prior studies indicate that Random Forest outperforms other standard algorithms,
achieving an accuracy of 0.983 and an RMSE of 0.1423 (Ruba Obiedat, Sara Amjad
Taubasi, 2022). Furthermore, this study will implement the Cross Industry Standard
Process for Data Mining (CRISP-DM) model, which is well-suited for guiding our
research as `it provides a comprehensive framework for developing the data mining
Machine Learning: We use machine learning in our studies to automatically learn and
improve from experience without being explicitly programmed. For the purpose of this
study, machine learning involves using algorithms to analyze employee data and predict
future turnover rates.
CRISP-DM: This iterative framework helps ensure that projects align with business
objectives while effectively managing data throughout the process. We used the Crispdm
model to efficiently and accurately analyze what our proposed machine learning and to
be more data driven on decision making to prevent risks or errors that may encounter
during the working phase of our proposed study.
such as Weka and Akkiko Inc. The dataset utilized in this research will consist of
3,000 instances, sourced from verified government records and previous studies related to
socioeconomic factors affecting eligibility for TUPAD assistance. The dataset will
level, income, and geographic location. The treatment approach will emphasize retaining
relevant features that significantly impact beneficiary selection while excluding irrelevant
or redundant data points. This ensures that the model focuses on individuals who
For classification, this study will specifically utilize three algorithms: Random
Forest, Naive Bayes, and J48 (a decision tree algorithm). These algorithms were chosen
for their effectiveness in handling classification tasks and their ability to provide insights
into feature importance. The study will also include system figures that illustrate the data
processing workflow, model evaluation metrics, and comparative performance results of
The study has several limitations that could impact its findings and applicability.
Firstly, it relies heavily on the availability and quality of data from verified sources; any
gaps or inconsistencies in this data may adversely affect the performance of the model.
Additionally, the complexity of social factors influencing eligibility may not be fully
captured in the dataset, which could lead to biases in the predictions made by the model.
While the focus on specific algorithms such as Random Forest, Naive Bayes, and J48
allows for a targeted analysis, it also means that other potentially valuable algorithms are
excluded from consideration. Furthermore, although the developed model may perform
well on training data, it could encounter challenges when generalizing to new or unseen
The TUPAD program, run by the Department of Labor and Employment (DOLE),
emergency jobs. The program has several strengths. It not only helps individuals by
providing income and improving their skills, but it also benefits the community. TUPAD
works on local projects that make communities safer and better places to live.
However, there are also some weaknesses. Beneficiaries often struggle to find jobs that
match their skills, which can lead to financial problems since the program only offers
temporary work. Also, there are not enough support services like training programs that
could help beneficiaries find long-term work. From a management perspective, it’s
difficult to track and manage the large amount of data, making it harder to allocate
resources effectively.
Despite these challenges, there are opportunities to improve the program. Our research
will use machine learning to help select beneficiaries more accurately and efficiently. By
improving the process of choosing who should receive support, we can reduce mistakes
and ensure that help reaches those who need it most. Expanding the data and including
more useful information can help the program make better decisions. There is also the
chance to offer long-term solutions for beneficiaries, making the program’s impact last
longer than just temporary jobs. Collaborating with local governments, NGOs, and other